16#if defined(LLVM_ON_UNIX) && !defined(__ANDROID__)
40 return PageSize.takeError();
41 return std::make_unique<InProcessMemoryMapper>(*PageSize);
54 std::lock_guard<std::mutex> Lock(
Mutex);
55 Reservations[MB.base()].Size = MB.allocatedSize();
63 return Addr.toPtr<
char *>();
74 auto Size = Segment.ContentSize + Segment.ZeroFillSize;
82 std::memset((
Base + Segment.ContentSize).toPtr<
void *>(), 0,
83 Segment.ZeroFillSize);
95 if (!DeinitializeActions)
96 return OnInitialized(DeinitializeActions.takeError());
99 std::lock_guard<std::mutex> Lock(
Mutex);
102 Allocations[MinAddr].Size = MaxAddr - MinAddr;
103 Allocations[MinAddr].DeinitializationActions =
104 std::move(*DeinitializeActions);
105 Reservations[AI.
MappingBase.
toPtr<
void *>()].Allocations.push_back(MinAddr);
108 OnInitialized(MinAddr);
117 std::lock_guard<std::mutex> Lock(
Mutex);
122 Allocations[
Base].DeinitializationActions)) {
123 AllErr =
joinErrors(std::move(AllErr), std::move(Err));
138 OnDeinitialized(std::move(AllErr));
145 for (
auto Base : Bases) {
146 std::vector<ExecutorAddr> AllocAddrs;
149 std::lock_guard<std::mutex> Lock(
Mutex);
150 auto &R = Reservations[
Base.toPtr<
void *>()];
152 AllocAddrs.swap(R.Allocations);
156 std::promise<MSVCPError>
P;
157 auto F =
P.get_future();
171 std::lock_guard<std::mutex> Lock(
Mutex);
172 Reservations.
erase(
Base.toPtr<
void *>());
175 OnReleased(std::move(Err));
179 std::vector<ExecutorAddr> ReservationAddrs;
181 std::lock_guard<std::mutex> Lock(
Mutex);
183 ReservationAddrs.reserve(Reservations.
size());
184 for (
const auto &R : Reservations) {
189 std::promise<MSVCPError>
P;
190 auto F =
P.get_future();
191 release(ReservationAddrs, [&](
Error Err) {
P.set_value(std::move(Err)); });
200#if (!defined(LLVM_ON_UNIX) || defined(__ANDROID__)) && !defined(_WIN32)
201 llvm_unreachable(
"SharedMemoryMapper is not supported on this platform yet");
207#if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32)
210 return PageSize.takeError();
212 return std::make_unique<SharedMemoryMapper>(EPC, SAs, *PageSize);
214 return make_error<StringError>(
215 "SharedMemoryMapper is not supported on this platform yet",
222#if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32)
227 [
this, NumBytes, OnReserved = std::move(OnReserved)](
228 Error SerializationErr,
230 if (SerializationErr) {
232 return OnReserved(std::move(SerializationErr));
236 return OnReserved(
Result.takeError());
239 std::string SharedMemoryName;
240 std::tie(RemoteAddr, SharedMemoryName) = std::move(*
Result);
242 void *LocalAddr =
nullptr;
244#if defined(LLVM_ON_UNIX)
248 reinterpret_cast<const uint8_t *
>(SharedMemoryName.c_str()),
249 SharedMemoryName.size());
250 auto HashedName = BLAKE3::hash<sizeof(key_t)>(
Data);
251 key_t Key = *
reinterpret_cast<key_t *
>(HashedName.data());
253 shmget(Key, NumBytes, IPC_CREAT | __IPC_SHAREAS | 0700);
254 if (SharedMemoryId < 0) {
256 std::error_code(errno, std::generic_category())));
258 LocalAddr = shmat(SharedMemoryId,
nullptr, 0);
259 if (LocalAddr ==
reinterpret_cast<void *
>(-1)) {
261 std::error_code(errno, std::generic_category())));
264 int SharedMemoryFile = shm_open(SharedMemoryName.c_str(), O_RDWR, 0700);
265 if (SharedMemoryFile < 0) {
270 shm_unlink(SharedMemoryName.c_str());
272 LocalAddr = mmap(
nullptr, NumBytes, PROT_READ | PROT_WRITE, MAP_SHARED,
273 SharedMemoryFile, 0);
274 if (LocalAddr == MAP_FAILED) {
278 close(SharedMemoryFile);
283 std::wstring WideSharedMemoryName(SharedMemoryName.begin(),
284 SharedMemoryName.end());
285 HANDLE SharedMemoryFile = OpenFileMappingW(
286 FILE_MAP_ALL_ACCESS, FALSE, WideSharedMemoryName.c_str());
287 if (!SharedMemoryFile)
291 MapViewOfFile(SharedMemoryFile, FILE_MAP_ALL_ACCESS, 0, 0, 0);
293 CloseHandle(SharedMemoryFile);
297 CloseHandle(SharedMemoryFile);
301 std::lock_guard<std::mutex> Lock(
Mutex);
302 Reservations.insert({RemoteAddr, {LocalAddr, NumBytes}});
310 OnReserved(make_error<StringError>(
311 "SharedMemoryMapper is not supported on this platform yet",
317 auto R = Reservations.upper_bound(
Addr);
318 assert(R != Reservations.begin() &&
"Attempt to prepare unreserved range");
323 return static_cast<char *
>(R->second.LocalAddr) +
Offset;
328 auto Reservation = Reservations.upper_bound(AI.
MappingBase);
329 assert(Reservation != Reservations.begin() &&
"Attempt to initialize unreserved range");
332 auto AllocationOffset = AI.
MappingBase - Reservation->first;
341 char *
Base =
static_cast<char *
>(Reservation->second.LocalAddr) +
342 AllocationOffset + Segment.Offset;
343 std::memset(
Base + Segment.ContentSize, 0, Segment.ZeroFillSize);
346 SegReq.
RAG = {Segment.AG.getMemProt(),
349 SegReq.
Size = Segment.ContentSize + Segment.ZeroFillSize;
357 [OnInitialized = std::move(OnInitialized)](
359 if (SerializationErr) {
361 return OnInitialized(std::move(SerializationErr));
364 OnInitialized(std::move(
Result));
366 SAs.
Instance, Reservation->first, std::move(FR));
375 [OnDeinitialized = std::move(OnDeinitialized)](
Error SerializationErr,
377 if (SerializationErr) {
379 return OnDeinitialized(std::move(SerializationErr));
382 OnDeinitialized(std::move(
Result));
389#if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32)
393 std::lock_guard<std::mutex> Lock(
Mutex);
395 for (
auto Base : Bases) {
397#if defined(LLVM_ON_UNIX)
400 if (shmdt(Reservations[
Base].LocalAddr) < 0)
403 if (munmap(Reservations[
Base].LocalAddr, Reservations[
Base].
Size) != 0)
409 if (!UnmapViewOfFile(Reservations[
Base].LocalAddr))
415 Reservations.erase(
Base);
422 [OnReleased = std::move(OnReleased),
424 if (SerializationErr) {
427 joinErrors(std::move(Err), std::move(SerializationErr)));
434 OnReleased(make_error<StringError>(
435 "SharedMemoryMapper is not supported on this platform yet",
441 std::lock_guard<std::mutex> Lock(
Mutex);
442 for (
const auto &R : Reservations) {
444#if defined(LLVM_ON_UNIX) && !defined(__ANDROID__)
447 shmdt(R.second.LocalAddr);
449 munmap(R.second.LocalAddr, R.second.Size);
454 UnmapViewOfFile(R.second.LocalAddr);
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static cl::opt< int > PageSize("imp-null-check-page-size", cl::desc("The page size of the target in bytes"), cl::init(4096), cl::Hidden)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
bool erase(const KeyT &Val)
Lightweight error class with error context and mandatory checking.
static ErrorSuccess success()
Create a success value.
Tagged union holding either a T or a Error.
Represents an address in the executor process.
static ExecutorAddr fromPtr(T *Ptr, UnwrapFn &&Unwrap=UnwrapFn())
Create an ExecutorAddr from the given pointer.
std::enable_if_t< std::is_pointer< T >::value, T > toPtr(WrapFn &&Wrap=WrapFn()) const
Cast this ExecutorAddr to a pointer of the given type.
ExecutorProcessControl supports interaction with a JIT target process.
void callSPSWrapperAsync(RunPolicyT &&Runner, ExecutorAddr WrapperFnAddr, SendResultT &&SendResult, const ArgTs &...Args)
Run a wrapper function using SPS to serialize the arguments and deserialize the results.
void initialize(AllocInfo &AI, OnInitializedFunction OnInitialized) override
Ensures executor memory is synchronized with working copy memory, sends functions to be called after ...
void reserve(size_t NumBytes, OnReservedFunction OnReserved) override
Reserves address space in executor process.
InProcessMemoryMapper(size_t PageSize)
~InProcessMemoryMapper() override
void deinitialize(ArrayRef< ExecutorAddr > Allocations, OnDeinitializedFunction OnDeInitialized) override
Runs previously specified deinitialization actions Executor addresses returned by initialize should b...
static Expected< std::unique_ptr< InProcessMemoryMapper > > Create()
char * prepare(ExecutorAddr Addr, size_t ContentSize) override
Provides working memory.
void release(ArrayRef< ExecutorAddr > Reservations, OnReleasedFunction OnRelease) override
Release address space acquired through reserve()
static Expected< std::unique_ptr< SharedMemoryMapper > > Create(ExecutorProcessControl &EPC, SymbolAddrs SAs)
void reserve(size_t NumBytes, OnReservedFunction OnReserved) override
Reserves address space in executor process.
void deinitialize(ArrayRef< ExecutorAddr > Allocations, OnDeinitializedFunction OnDeInitialized) override
Runs previously specified deinitialization actions Executor addresses returned by initialize should b...
~SharedMemoryMapper() override
void initialize(AllocInfo &AI, OnInitializedFunction OnInitialized) override
Ensures executor memory is synchronized with working copy memory, sends functions to be called after ...
char * prepare(ExecutorAddr Addr, size_t ContentSize) override
Provides working memory.
void release(ArrayRef< ExecutorAddr > Reservations, OnReleasedFunction OnRelease) override
Release address space acquired through reserve()
SharedMemoryMapper(ExecutorProcessControl &EPC, SymbolAddrs SAs, size_t PageSize)
This class encapsulates the notion of a memory block which has an address and a size.
static std::error_code releaseMappedMemory(MemoryBlock &Block)
This method releases a block of memory that was allocated with the allocateMappedMemory method.
static MemoryBlock allocateMappedMemory(size_t NumBytes, const MemoryBlock *const NearBlock, unsigned Flags, std::error_code &EC)
This method allocates a block of memory that is suitable for loading dynamically generated code (e....
static void InvalidateInstructionCache(const void *Addr, size_t Len)
InvalidateInstructionCache - Before the JIT can run a block of code that has been emitted it must inv...
static std::error_code protectMappedMemory(const MemoryBlock &Block, unsigned Flags)
This method sets the protection flags for a block of memory to the state specified by /p Flags.
static Expected< unsigned > getPageSize()
Get the process's page size.
unique_function is a type-erasing functor similar to std::function.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
shared::SPSExpected< shared::SPSExecutorAddr >(shared::SPSExecutorAddr, shared::SPSExecutorAddr, shared::SPSSharedMemoryFinalizeRequest) SPSExecutorSharedMemoryMapperServiceInitializeSignature
shared::SPSError(shared::SPSExecutorAddr, shared::SPSSequence< shared::SPSExecutorAddr >) SPSExecutorSharedMemoryMapperServiceReleaseSignature
shared::SPSExpected< shared::SPSTuple< shared::SPSExecutorAddr, shared::SPSString > >(shared::SPSExecutorAddr, uint64_t) SPSExecutorSharedMemoryMapperServiceReserveSignature
shared::SPSError(shared::SPSExecutorAddr, shared::SPSSequence< shared::SPSExecutorAddr >) SPSExecutorSharedMemoryMapperServiceDeinitializeSignature
Error runDeallocActions(ArrayRef< WrapperFunctionCall > DAs)
Run deallocation actions.
Expected< std::vector< WrapperFunctionCall > > runFinalizeActions(AllocActions &AAs)
Run finalize actions.
@ Finalize
Finalize memory should be allocated by the allocator, and then be overwritten and deallocated after a...
sys::Memory::ProtectionFlags toSysMemoryProtectionFlags(MemProt MP)
Convert a MemProt value to a corresponding sys::Memory::ProtectionFlags value.
This is an optimization pass for GlobalISel generic memory operations.
std::error_code inconvertibleErrorCode()
The value returned by this function can be returned from convertToErrorCode for Error values where no...
auto reverse(ContainerTy &&C)
Error joinErrors(Error E1, Error E2)
Concatenate errors.
void cantFail(Error Err, const char *Msg=nullptr)
Report a fatal error if Err is a failure value.
Error errorCodeToError(std::error_code EC)
Helper for converting an std::error_code to a Error.
std::error_code errnoAsErrorCode()
Helper to get errno as an std::error_code.
std::error_code mapWindowsError(unsigned EV)
Represents an address range in the exceutor process.
Represents a single allocation containing multiple segments and initialization and deinitialization a...
std::vector< SegInfo > Segments
shared::AllocActions Actions
ExecutorAddr Deinitialize
std::vector< SharedMemorySegFinalizeRequest > Segments
shared::AllocActions Actions