197 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			197 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			C++
		
	
	
	
//===- Win32/Memory.cpp - Win32 Memory Implementation -----------*- C++ -*-===//
 | 
						|
//
 | 
						|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
 | 
						|
// See https://llvm.org/LICENSE.txt for license information.
 | 
						|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
 | 
						|
//
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
//
 | 
						|
// This file provides the Win32 specific implementation of various Memory
 | 
						|
// management utilities
 | 
						|
//
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
 | 
						|
#include "llvm/Support/DataTypes.h"
 | 
						|
#include "llvm/Support/ErrorHandling.h"
 | 
						|
#include "llvm/Support/Process.h"
 | 
						|
#include "llvm/Support/WindowsError.h"
 | 
						|
 | 
						|
// The Windows.h header must be the last one included.
 | 
						|
#include "llvm/Support/Windows/WindowsSupport.h"
 | 
						|
 | 
						|
static DWORD getWindowsProtectionFlags(unsigned Flags) {
 | 
						|
  switch (Flags & llvm::sys::Memory::MF_RWE_MASK) {
 | 
						|
  // Contrary to what you might expect, the Windows page protection flags
 | 
						|
  // are not a bitwise combination of RWX values
 | 
						|
  case llvm::sys::Memory::MF_READ:
 | 
						|
    return PAGE_READONLY;
 | 
						|
  case llvm::sys::Memory::MF_WRITE:
 | 
						|
    // Note: PAGE_WRITE is not supported by VirtualProtect
 | 
						|
    return PAGE_READWRITE;
 | 
						|
  case llvm::sys::Memory::MF_READ|llvm::sys::Memory::MF_WRITE:
 | 
						|
    return PAGE_READWRITE;
 | 
						|
  case llvm::sys::Memory::MF_READ|llvm::sys::Memory::MF_EXEC:
 | 
						|
    return PAGE_EXECUTE_READ;
 | 
						|
  case llvm::sys::Memory::MF_READ |
 | 
						|
         llvm::sys::Memory::MF_WRITE |
 | 
						|
         llvm::sys::Memory::MF_EXEC:
 | 
						|
    return PAGE_EXECUTE_READWRITE;
 | 
						|
  case llvm::sys::Memory::MF_EXEC:
 | 
						|
    return PAGE_EXECUTE;
 | 
						|
  default:
 | 
						|
    llvm_unreachable("Illegal memory protection flag specified!");
 | 
						|
  }
 | 
						|
  // Provide a default return value as required by some compilers.
 | 
						|
  return PAGE_NOACCESS;
 | 
						|
}
 | 
						|
 | 
						|
// While we'd be happy to allocate single pages, the Windows allocation
 | 
						|
// granularity may be larger than a single page (in practice, it is 64K)
 | 
						|
// so mapping less than that will create an unreachable fragment of memory.
 | 
						|
static size_t getAllocationGranularity() {
 | 
						|
  SYSTEM_INFO  Info;
 | 
						|
  ::GetSystemInfo(&Info);
 | 
						|
  if (Info.dwPageSize > Info.dwAllocationGranularity)
 | 
						|
    return Info.dwPageSize;
 | 
						|
  else
 | 
						|
    return Info.dwAllocationGranularity;
 | 
						|
}
 | 
						|
 | 
						|
// Large/huge memory pages need explicit process permissions in order to be
 | 
						|
// used. See https://blogs.msdn.microsoft.com/oldnewthing/20110128-00/?p=11643
 | 
						|
// Also large pages need to be manually enabled on your OS. If all this is
 | 
						|
// sucessfull, we return the minimal large memory page size.
 | 
						|
static size_t enableProcessLargePages() {
 | 
						|
  HANDLE Token = 0;
 | 
						|
  size_t LargePageMin = GetLargePageMinimum();
 | 
						|
  if (LargePageMin)
 | 
						|
    OpenProcessToken(GetCurrentProcess(), TOKEN_ADJUST_PRIVILEGES | TOKEN_QUERY,
 | 
						|
                     &Token);
 | 
						|
  if (!Token)
 | 
						|
    return 0;
 | 
						|
  LUID Luid;
 | 
						|
  if (!LookupPrivilegeValue(0, SE_LOCK_MEMORY_NAME, &Luid)) {
 | 
						|
    CloseHandle(Token);
 | 
						|
    return 0;
 | 
						|
  }
 | 
						|
  TOKEN_PRIVILEGES TP{};
 | 
						|
  TP.PrivilegeCount = 1;
 | 
						|
  TP.Privileges[0].Luid = Luid;
 | 
						|
  TP.Privileges[0].Attributes = SE_PRIVILEGE_ENABLED;
 | 
						|
  if (!AdjustTokenPrivileges(Token, FALSE, &TP, 0, 0, 0)) {
 | 
						|
    CloseHandle(Token);
 | 
						|
    return 0;
 | 
						|
  }
 | 
						|
  DWORD E = GetLastError();
 | 
						|
  CloseHandle(Token);
 | 
						|
  if (E == ERROR_SUCCESS)
 | 
						|
    return LargePageMin;
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
namespace llvm {
 | 
						|
namespace sys {
 | 
						|
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
//=== WARNING: Implementation here must contain only Win32 specific code
 | 
						|
//===          and must not be UNIX code
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
 | 
						|
MemoryBlock Memory::allocateMappedMemory(size_t NumBytes,
 | 
						|
                                         const MemoryBlock *const NearBlock,
 | 
						|
                                         unsigned Flags,
 | 
						|
                                         std::error_code &EC) {
 | 
						|
  EC = std::error_code();
 | 
						|
  if (NumBytes == 0)
 | 
						|
    return MemoryBlock();
 | 
						|
 | 
						|
  static size_t DefaultGranularity = getAllocationGranularity();
 | 
						|
  static size_t LargePageGranularity = enableProcessLargePages();
 | 
						|
 | 
						|
  DWORD AllocType = MEM_RESERVE | MEM_COMMIT;
 | 
						|
  bool HugePages = false;
 | 
						|
  size_t Granularity = DefaultGranularity;
 | 
						|
 | 
						|
  if ((Flags & MF_HUGE_HINT) && LargePageGranularity > 0) {
 | 
						|
    AllocType |= MEM_LARGE_PAGES;
 | 
						|
    HugePages = true;
 | 
						|
    Granularity = LargePageGranularity;
 | 
						|
  }
 | 
						|
 | 
						|
  size_t NumBlocks = (NumBytes + Granularity - 1) / Granularity;
 | 
						|
 | 
						|
  uintptr_t Start = NearBlock ? reinterpret_cast<uintptr_t>(NearBlock->base()) +
 | 
						|
                                NearBlock->allocatedSize()
 | 
						|
                           : 0;
 | 
						|
 | 
						|
  // If the requested address is not aligned to the allocation granularity,
 | 
						|
  // round up to get beyond NearBlock. VirtualAlloc would have rounded down.
 | 
						|
  if (Start && Start % Granularity != 0)
 | 
						|
    Start += Granularity - Start % Granularity;
 | 
						|
 | 
						|
  DWORD Protect = getWindowsProtectionFlags(Flags);
 | 
						|
 | 
						|
  size_t AllocSize = NumBlocks * Granularity;
 | 
						|
  void *PA = ::VirtualAlloc(reinterpret_cast<void *>(Start),
 | 
						|
                            AllocSize, AllocType, Protect);
 | 
						|
  if (PA == NULL) {
 | 
						|
    if (NearBlock || HugePages) {
 | 
						|
      // Try again without the NearBlock hint and without large memory pages
 | 
						|
      return allocateMappedMemory(NumBytes, NULL, Flags & ~MF_HUGE_HINT, EC);
 | 
						|
    }
 | 
						|
    EC = mapWindowsError(::GetLastError());
 | 
						|
    return MemoryBlock();
 | 
						|
  }
 | 
						|
 | 
						|
  MemoryBlock Result;
 | 
						|
  Result.Address = PA;
 | 
						|
  Result.AllocatedSize = AllocSize;
 | 
						|
  Result.Flags = (Flags & ~MF_HUGE_HINT) | (HugePages ? MF_HUGE_HINT : 0);
 | 
						|
 | 
						|
  if (Flags & MF_EXEC)
 | 
						|
    Memory::InvalidateInstructionCache(Result.Address, AllocSize);
 | 
						|
 | 
						|
  return Result;
 | 
						|
}
 | 
						|
 | 
						|
  std::error_code Memory::releaseMappedMemory(MemoryBlock &M) {
 | 
						|
  if (M.Address == 0 || M.AllocatedSize == 0)
 | 
						|
    return std::error_code();
 | 
						|
 | 
						|
  if (!VirtualFree(M.Address, 0, MEM_RELEASE))
 | 
						|
    return mapWindowsError(::GetLastError());
 | 
						|
 | 
						|
  M.Address = 0;
 | 
						|
  M.AllocatedSize = 0;
 | 
						|
 | 
						|
  return std::error_code();
 | 
						|
}
 | 
						|
 | 
						|
  std::error_code Memory::protectMappedMemory(const MemoryBlock &M,
 | 
						|
                                       unsigned Flags) {
 | 
						|
  if (M.Address == 0 || M.AllocatedSize == 0)
 | 
						|
    return std::error_code();
 | 
						|
 | 
						|
  DWORD Protect = getWindowsProtectionFlags(Flags);
 | 
						|
 | 
						|
  DWORD OldFlags;
 | 
						|
  if (!VirtualProtect(M.Address, M.AllocatedSize, Protect, &OldFlags))
 | 
						|
    return mapWindowsError(::GetLastError());
 | 
						|
 | 
						|
  if (Flags & MF_EXEC)
 | 
						|
    Memory::InvalidateInstructionCache(M.Address, M.AllocatedSize);
 | 
						|
 | 
						|
  return std::error_code();
 | 
						|
}
 | 
						|
 | 
						|
/// InvalidateInstructionCache - Before the JIT can run a block of code
 | 
						|
/// that has been emitted it must invalidate the instruction cache on some
 | 
						|
/// platforms.
 | 
						|
void Memory::InvalidateInstructionCache(
 | 
						|
    const void *Addr, size_t Len) {
 | 
						|
  FlushInstructionCache(GetCurrentProcess(), Addr, Len);
 | 
						|
}
 | 
						|
 | 
						|
} // namespace sys
 | 
						|
} // namespace llvm
 |