| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213 |
- //-------------------------------------------------------------------------------------------------------
- // Copyright (C) Microsoft. All rights reserved.
- // Licensed under the MIT license. See LICENSE.txt file in the project root for full license information.
- //-------------------------------------------------------------------------------------------------------
- #include "CommonMemoryPch.h"
- #if ENABLE_NATIVE_CODEGEN || DYNAMIC_INTERPRETER_THUNK
- #include "Memory/XDataAllocator.h"
- #if defined(_M_ARM)
- #include <wchar.h>
- #endif
- #include "CustomHeap.h"
- #if PDATA_ENABLED && defined(_WIN32)
- #include "Core/DelayLoadLibrary.h"
- #include <malloc.h>
- #endif
- namespace Memory
- {
- namespace CustomHeap
- {
- #pragma region "Constructor and Destructor"
- template<typename TAlloc, typename TPreReservedAlloc>
- Heap<TAlloc, TPreReservedAlloc>::Heap(ArenaAllocator * alloc, CodePageAllocators<TAlloc, TPreReservedAlloc> * codePageAllocators, HANDLE processHandle):
- auxiliaryAllocator(alloc),
- codePageAllocators(codePageAllocators),
- lastSecondaryAllocStateChangedCount(0),
- processHandle(processHandle)
- #if DBG_DUMP
- , freeObjectSize(0)
- , totalAllocationSize(0)
- , allocationsSinceLastCompact(0)
- , freesSinceLastCompact(0)
- #endif
- #if DBG
- , inDtor(false)
- #endif
- {
- for (int i = 0; i < NumBuckets; i++)
- {
- this->buckets[i].Reset();
- }
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- Heap<TAlloc, TPreReservedAlloc>::~Heap()
- {
- #if DBG
- inDtor = true;
- #endif
- this->FreeAll();
- }
- #pragma endregion
- #pragma region "Public routines"
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeAll()
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- FreeBuckets(false);
- FreeLargeObjects();
- FreeDecommittedBuckets();
- FreeDecommittedLargeObjects();
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::Free(__in Allocation* object)
- {
- Assert(object != nullptr);
- if (object == nullptr)
- {
- return;
- }
- BucketId bucket = (BucketId) GetBucketForSize(object->size);
- if (bucket == BucketId::LargeObjectList)
- {
- #if PDATA_ENABLED
- if(!object->xdata.IsFreed())
- {
- FreeXdata(&object->xdata, object->largeObjectAllocation.segment);
- }
- #endif
- if (!object->largeObjectAllocation.isDecommitted)
- {
- FreeLargeObject(object);
- }
- return;
- }
- #if PDATA_ENABLED
- if(!object->xdata.IsFreed())
- {
- FreeXdata(&object->xdata, object->page->segment);
- }
- #endif
- if (!object->page->isDecommitted)
- {
- FreeAllocation(object);
- }
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::DecommitAll()
- {
- // This function doesn't really touch the page allocator data structure.
- // DecommitPages is merely a wrapper for VirtualFree
- // So no need to take the critical section to synchronize
- DListBase<Allocation>::EditingIterator i(&this->largeObjectAllocations);
- while (i.Next())
- {
- Allocation& allocation = i.Data();
- Assert(!allocation.largeObjectAllocation.isDecommitted);
- this->codePageAllocators->DecommitPages(allocation.address, allocation.GetPageCount(), allocation.largeObjectAllocation.segment);
- i.MoveCurrentTo(&this->decommittedLargeObjects);
- allocation.largeObjectAllocation.isDecommitted = true;
- }
- for (int bucket = 0; bucket < BucketId::NumBuckets; bucket++)
- {
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, &(this->fullPages[bucket]), bucketIter1)
- {
- Assert(page.inFullList);
- this->codePageAllocators->DecommitPages(page.address, 1 /* pageCount */, page.segment);
- bucketIter1.MoveCurrentTo(&(this->decommittedPages));
- page.isDecommitted = true;
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, &(this->buckets[bucket]), bucketIter2)
- {
- Assert(!page.inFullList);
- this->codePageAllocators->DecommitPages(page.address, 1 /* pageCount */, page.segment);
- bucketIter2.MoveCurrentTo(&(this->decommittedPages));
- page.isDecommitted = true;
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::IsInHeap(DListBase<Page> const& bucket, __in void * address)
- {
- DListBase<Page>::Iterator i(&bucket);
- while (i.Next())
- {
- Page& page = i.Data();
- if (page.address <= address && address < page.address + AutoSystemInfo::PageSize)
- {
- return true;
- }
- }
- return false;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::IsInHeap(DListBase<Page> const buckets[NumBuckets], __in void * address)
- {
- for (uint i = 0; i < NumBuckets; i++)
- {
- if (this->IsInHeap(buckets[i], address))
- {
- return true;
- }
- }
- return false;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::IsInHeap(DListBase<Allocation> const& allocations, __in void *address)
- {
- DListBase<Allocation>::Iterator i(&allocations);
- while (i.Next())
- {
- Allocation& allocation = i.Data();
- if (allocation.address <= address && address < allocation.address + allocation.size)
- {
- return true;
- }
- }
- return false;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::IsInHeap(__in void* address)
- {
- return IsInHeap(buckets, address) || IsInHeap(fullPages, address) || IsInHeap(largeObjectAllocations, address);
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- Page * Heap<TAlloc, TPreReservedAlloc>::GetExistingPage(BucketId bucket, bool canAllocInPreReservedHeapPageSegment)
- {
- // TODO: this can get a non-prereserved page even if you want one
- if (!this->buckets[bucket].Empty())
- {
- Assert(!this->buckets[bucket].Head().inFullList);
- return &this->buckets[bucket].Head();
- }
- return FindPageToSplit(bucket, canAllocInPreReservedHeapPageSegment);
- }
- /*
- * Algorithm:
- * - Find bucket
- * - Check bucket pages - if it has enough free space, allocate that chunk
- * - Check pages in bigger buckets - if that has enough space, split that page and allocate from that chunk
- * - Allocate new page
- */
- template<typename TAlloc, typename TPreReservedAlloc>
- Allocation* Heap<TAlloc, TPreReservedAlloc>::Alloc(size_t bytes, ushort pdataCount, ushort xdataSize, bool canAllocInPreReservedHeapPageSegment, bool isAnyJittedCode, _Inout_ bool* isAllJITCodeInPreReservedRegion)
- {
- Assert(bytes > 0);
- Assert((codePageAllocators->AllocXdata() || pdataCount == 0) && (!codePageAllocators->AllocXdata() || pdataCount > 0));
- Assert(pdataCount > 0 || (pdataCount == 0 && xdataSize == 0));
- // Round up to power of two to allocate, and figure out which bucket to allocate in
- int _;
- size_t bytesToAllocate = PowerOf2Policy::GetSize(bytes, &_ /* modFunctionIndex */);
- BucketId bucket = (BucketId) GetBucketForSize(bytesToAllocate);
- if (bucket == BucketId::LargeObjectList)
- {
- return AllocLargeObject(bytes, pdataCount, xdataSize, canAllocInPreReservedHeapPageSegment, isAnyJittedCode, isAllJITCodeInPreReservedRegion);
- }
- VerboseHeapTrace(_u("Bucket is %d\n"), bucket);
- VerboseHeapTrace(_u("Requested: %d bytes. Allocated: %d bytes\n"), bytes, bytesToAllocate);
- do
- {
- Page* page = GetExistingPage(bucket, canAllocInPreReservedHeapPageSegment);
- if (page == nullptr && UpdateFullPages())
- {
- page = GetExistingPage(bucket, canAllocInPreReservedHeapPageSegment);
- }
- if (page == nullptr)
- {
- page = AllocNewPage(bucket, canAllocInPreReservedHeapPageSegment, isAnyJittedCode, isAllJITCodeInPreReservedRegion);
- }
- else if (!canAllocInPreReservedHeapPageSegment && isAnyJittedCode)
- {
- *isAllJITCodeInPreReservedRegion = false;
- }
- // Out of memory
- if (page == nullptr)
- {
- return nullptr;
- }
- #if defined(DBG)
- MEMORY_BASIC_INFORMATION memBasicInfo;
- size_t resultBytes = VirtualQueryEx(this->processHandle, page->address, &memBasicInfo, sizeof(memBasicInfo));
- if (resultBytes == 0)
- {
- MemoryOperationLastError::RecordLastError();
- }
- else
- {
- Assert(memBasicInfo.Protect == PAGE_EXECUTE_READ);
- }
- #endif
- Allocation* allocation = nullptr;
- if (AllocInPage(page, bytesToAllocate, pdataCount, xdataSize, &allocation))
- {
- return allocation;
- }
- } while (true);
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- BOOL Heap<TAlloc, TPreReservedAlloc>::ProtectAllocationWithExecuteReadWrite(Allocation *allocation, __in_opt char* addressInPage)
- {
- DWORD protectFlags = 0;
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- protectFlags = PAGE_EXECUTE_RW_TARGETS_NO_UPDATE;
- }
- else
- {
- protectFlags = PAGE_EXECUTE_READWRITE;
- }
- return this->ProtectAllocation(allocation, protectFlags, PAGE_EXECUTE_READ, addressInPage);
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- BOOL Heap<TAlloc, TPreReservedAlloc>::ProtectAllocationWithExecuteReadOnly(__in Allocation *allocation, __in_opt char* addressInPage)
- {
- DWORD protectFlags = 0;
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- protectFlags = PAGE_EXECUTE_RO_TARGETS_NO_UPDATE;
- }
- else
- {
- protectFlags = PAGE_EXECUTE_READ;
- }
- return this->ProtectAllocation(allocation, protectFlags, PAGE_EXECUTE_READWRITE, addressInPage);
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- BOOL Heap<TAlloc, TPreReservedAlloc>::ProtectAllocation(__in Allocation* allocation, DWORD dwVirtualProtectFlags, DWORD desiredOldProtectFlag, __in_opt char* addressInPage)
- {
- // Allocate at the page level so that our protections don't
- // transcend allocation page boundaries. Here, allocation->address is page
- // aligned if the object is a large object allocation. If it isn't, in the else
- // branch of the following if statement, we set it to the allocation's page's
- // address. This ensures that the address being protected is always page aligned
- Assert(allocation != nullptr);
- Assert(allocation->isAllocationUsed);
- Assert(addressInPage == nullptr || (addressInPage >= allocation->address && addressInPage < (allocation->address + allocation->size)));
- char* address = allocation->address;
- size_t pageCount;
- void * segment;
- if (allocation->IsLargeAllocation())
- {
- #if DBG_DUMP || defined(RECYCLER_TRACE)
- if (Js::Configuration::Global.flags.IsEnabled(Js::TraceProtectPagesFlag))
- {
- Output::Print(_u("Protecting large allocation\n"));
- }
- #endif
- segment = allocation->largeObjectAllocation.segment;
- if (addressInPage != nullptr)
- {
- if (addressInPage >= allocation->address + AutoSystemInfo::PageSize)
- {
- size_t page = (addressInPage - allocation->address) / AutoSystemInfo::PageSize;
- address = allocation->address + (page * AutoSystemInfo::PageSize);
- }
- pageCount = 1;
- }
- else
- {
- pageCount = allocation->GetPageCount();
- }
- VerboseHeapTrace(_u("Protecting 0x%p with 0x%x\n"), address, dwVirtualProtectFlags);
- return this->codePageAllocators->ProtectPages(address, pageCount, segment, dwVirtualProtectFlags, desiredOldProtectFlag);
- }
- else
- {
- #if DBG_DUMP || defined(RECYCLER_TRACE)
- if (Js::Configuration::Global.flags.IsEnabled(Js::TraceProtectPagesFlag))
- {
- Output::Print(_u("Protecting small allocation\n"));
- }
- #endif
- segment = allocation->page->segment;
- address = allocation->page->address;
- pageCount = 1;
- VerboseHeapTrace(_u("Protecting 0x%p with 0x%x\n"), address, dwVirtualProtectFlags);
- return this->codePageAllocators->ProtectPages(address, pageCount, segment, dwVirtualProtectFlags, desiredOldProtectFlag);
- }
- }
- #pragma endregion
- #pragma region "Large object methods"
- template<typename TAlloc, typename TPreReservedAlloc>
- Allocation* Heap<TAlloc, TPreReservedAlloc>::AllocLargeObject(size_t bytes, ushort pdataCount, ushort xdataSize, bool canAllocInPreReservedHeapPageSegment, bool isAnyJittedCode, _Inout_ bool* isAllJITCodeInPreReservedRegion)
- {
- size_t pages = GetNumPagesForSize(bytes);
- if (pages == 0)
- {
- return nullptr;
- }
- void * segment = nullptr;
- char* address = nullptr;
- #if PDATA_ENABLED
- XDataAllocation xdata;
- #endif
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- address = this->codePageAllocators->Alloc(&pages, &segment, canAllocInPreReservedHeapPageSegment, isAnyJittedCode, isAllJITCodeInPreReservedRegion);
- // Out of memory
- if (address == nullptr)
- {
- return nullptr;
- }
- char* localAddr = this->codePageAllocators->AllocLocal(address, pages*AutoSystemInfo::PageSize, segment);
- if (!localAddr)
- {
- return nullptr;
- }
- FillDebugBreak((BYTE*)localAddr, pages*AutoSystemInfo::PageSize);
- this->codePageAllocators->FreeLocal(localAddr, segment);
- if (this->processHandle == GetCurrentProcess())
- {
- DWORD protectFlags = 0;
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- protectFlags = PAGE_EXECUTE_RO_TARGETS_NO_UPDATE;
- }
- else
- {
- protectFlags = PAGE_EXECUTE_READ;
- }
- this->codePageAllocators->ProtectPages(address, pages, segment, protectFlags /*dwVirtualProtectFlags*/, PAGE_READWRITE /*desiredOldProtectFlags*/);
- }
- #if PDATA_ENABLED
- if(pdataCount > 0)
- {
- if (!this->codePageAllocators->AllocSecondary(segment, (ULONG_PTR) address, bytes, pdataCount, xdataSize, &xdata))
- {
- this->codePageAllocators->Release(address, pages, segment);
- return nullptr;
- }
- }
- #endif
- }
- #if defined(DBG)
- MEMORY_BASIC_INFORMATION memBasicInfo;
- size_t resultBytes = VirtualQueryEx(this->processHandle, address, &memBasicInfo, sizeof(memBasicInfo));
- if (resultBytes == 0)
- {
- MemoryOperationLastError::RecordLastError();
- }
- else
- {
- Assert(memBasicInfo.Protect == PAGE_EXECUTE_READ);
- }
- #endif
- Allocation* allocation = this->largeObjectAllocations.PrependNode(this->auxiliaryAllocator);
- if (allocation == nullptr)
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- this->codePageAllocators->Release(address, pages, segment);
- #if PDATA_ENABLED
- if(pdataCount > 0)
- {
- this->codePageAllocators->ReleaseSecondary(xdata, segment);
- }
- #endif
- return nullptr;
- }
- allocation->address = address;
- allocation->largeObjectAllocation.segment = segment;
- allocation->largeObjectAllocation.isDecommitted = false;
- allocation->size = pages * AutoSystemInfo::PageSize;
- allocation->thunkAddress = 0;
- #if PDATA_ENABLED
- allocation->xdata = xdata;
- #endif
- return allocation;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeDecommittedLargeObjects()
- {
- // CodePageAllocators is locked in FreeAll
- Assert(inDtor);
- FOREACH_DLISTBASE_ENTRY_EDITING(Allocation, allocation, &this->decommittedLargeObjects, largeObjectIter)
- {
- VerboseHeapTrace(_u("Decommitting large object at address 0x%p of size %u\n"), allocation.address, allocation.size);
- this->codePageAllocators->ReleaseDecommitted(allocation.address, allocation.GetPageCount(), allocation.largeObjectAllocation.segment);
- largeObjectIter.RemoveCurrent(this->auxiliaryAllocator);
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- //Called during Free (while shutting down)
- template<typename TAlloc, typename TPreReservedAlloc>
- DWORD Heap<TAlloc, TPreReservedAlloc>::EnsurePageWriteable(Page* page)
- {
- return EnsurePageReadWrite<PAGE_READWRITE>(page);
- }
- // this get called when freeing the whole page
- template<typename TAlloc, typename TPreReservedAlloc>
- DWORD Heap<TAlloc, TPreReservedAlloc>::EnsureAllocationWriteable(Allocation* allocation)
- {
- return EnsureAllocationReadWrite<PAGE_READWRITE>(allocation);
- }
- // this get called when only freeing a part in the page
- template<typename TAlloc, typename TPreReservedAlloc>
- DWORD Heap<TAlloc, TPreReservedAlloc>::EnsureAllocationExecuteWriteable(Allocation* allocation)
- {
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- return EnsureAllocationReadWrite<PAGE_EXECUTE_RW_TARGETS_NO_UPDATE>(allocation);
- }
- else
- {
- return EnsureAllocationReadWrite<PAGE_EXECUTE_READWRITE>(allocation);
- }
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeLargeObjects()
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- FOREACH_DLISTBASE_ENTRY_EDITING(Allocation, allocation, &this->largeObjectAllocations, largeObjectIter)
- {
- EnsureAllocationWriteable(&allocation);
- #if PDATA_ENABLED
- Assert(allocation.xdata.IsFreed());
- #endif
- this->codePageAllocators->Release(allocation.address, allocation.GetPageCount(), allocation.largeObjectAllocation.segment);
- largeObjectIter.RemoveCurrent(this->auxiliaryAllocator);
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeLargeObject(Allocation* allocation)
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- EnsureAllocationWriteable(allocation);
- #if PDATA_ENABLED
- Assert(allocation->xdata.IsFreed());
- #endif
- this->codePageAllocators->Release(allocation->address, allocation->GetPageCount(), allocation->largeObjectAllocation.segment);
- this->largeObjectAllocations.RemoveElement(this->auxiliaryAllocator, allocation);
- }
- #pragma endregion
- #pragma region "Page methods"
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::AllocInPage(Page* page, size_t bytes, ushort pdataCount, ushort xdataSize, Allocation ** allocationOut)
- {
- Allocation * allocation = AnewNoThrowStruct(this->auxiliaryAllocator, Allocation);
- if (allocation == nullptr)
- {
- return true;
- }
- Assert(Math::IsPow2((int32)bytes));
- uint length = GetChunkSizeForBytes(bytes);
- BVIndex index = GetFreeIndexForPage(page, bytes);
- if (index == BVInvalidIndex)
- {
- CustomHeap_BadPageState_unrecoverable_error((ULONG_PTR)this);
- return false;
- }
- char* address = page->address + Page::Alignment * index;
- #if PDATA_ENABLED
- XDataAllocation xdata;
- if(pdataCount > 0)
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- if (this->ShouldBeInFullList(page))
- {
- Adelete(this->auxiliaryAllocator, allocation);
- // If we run out of XData space with the segment, move the page to the full page list, and return false to try the next page.
- BucketId bucket = page->currentBucket;
- VerboseHeapTrace(_u("Moving page from bucket %d to full list\n"), bucket);
- Assert(!page->inFullList);
- this->buckets[bucket].MoveElementTo(page, &this->fullPages[bucket]);
- page->inFullList = true;
- return false;
- }
- if (!this->codePageAllocators->AllocSecondary(page->segment, (ULONG_PTR)address, bytes, pdataCount, xdataSize, &xdata))
- {
- Adelete(this->auxiliaryAllocator, allocation);
- return true;
- }
- }
- #endif
- #if DBG
- allocation->isAllocationUsed = false;
- allocation->isNotExecutableBecauseOOM = false;
- #endif
- allocation->page = page;
- allocation->size = bytes;
- allocation->address = address;
- allocation->thunkAddress = 0;
- #if DBG_DUMP
- this->allocationsSinceLastCompact += bytes;
- this->freeObjectSize -= bytes;
- #endif
- //Section of the Page should already be freed.
- if (!page->freeBitVector.TestRange(index, length))
- {
- CustomHeap_BadPageState_unrecoverable_error((ULONG_PTR)this);
- return false;
- }
- //Section of the Page should already be freed.
- if (!page->freeBitVector.TestRange(index, length))
- {
- CustomHeap_BadPageState_unrecoverable_error((ULONG_PTR)this);
- return false;
- }
- page->freeBitVector.ClearRange(index, length);
- VerboseHeapTrace(_u("ChunkSize: %d, Index: %d, Free bit vector in page: "), length, index);
- #if VERBOSE_HEAP
- page->freeBitVector.DumpWord();
- #endif
- VerboseHeapTrace(_u("\n"));
- if (this->ShouldBeInFullList(page))
- {
- BucketId bucket = page->currentBucket;
- VerboseHeapTrace(_u("Moving page from bucket %d to full list\n"), bucket);
- Assert(!page->inFullList);
- this->buckets[bucket].MoveElementTo(page, &this->fullPages[bucket]);
- page->inFullList = true;
- }
- #if PDATA_ENABLED
- allocation->xdata = xdata;
- #endif
- *allocationOut = allocation;
- return true;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- Page* Heap<TAlloc, TPreReservedAlloc>::AllocNewPage(BucketId bucket, bool canAllocInPreReservedHeapPageSegment, bool isAnyJittedCode, _Inout_ bool* isAllJITCodeInPreReservedRegion)
- {
- void* pageSegment = nullptr;
- char* address = nullptr;
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- address = this->codePageAllocators->AllocPages(1, &pageSegment, canAllocInPreReservedHeapPageSegment, isAnyJittedCode, isAllJITCodeInPreReservedRegion);
- if (address == nullptr)
- {
- return nullptr;
- }
- }
- char* localAddr = this->codePageAllocators->AllocLocal(address, AutoSystemInfo::PageSize, pageSegment);
- if (!localAddr)
- {
- return nullptr;
- }
- FillDebugBreak((BYTE*)localAddr, AutoSystemInfo::PageSize);
- this->codePageAllocators->FreeLocal(localAddr, pageSegment);
- DWORD protectFlags = 0;
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- protectFlags = PAGE_EXECUTE_RO_TARGETS_NO_UPDATE;
- }
- else
- {
- protectFlags = PAGE_EXECUTE_READ;
- }
- //Change the protection of the page to Read-Only Execute, before adding it to the bucket list.
- this->codePageAllocators->ProtectPages(address, 1, pageSegment, protectFlags, PAGE_READWRITE);
- // Switch to allocating on a list of pages so we can do leak tracking later
- VerboseHeapTrace(_u("Allocing new page in bucket %d\n"), bucket);
- Page* page = this->buckets[bucket].PrependNode(this->auxiliaryAllocator, address, pageSegment, bucket);
- if (page == nullptr)
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- this->codePageAllocators->ReleasePages(address, 1, pageSegment);
- return nullptr;
- }
- #if DBG_DUMP
- this->totalAllocationSize += AutoSystemInfo::PageSize;
- this->freeObjectSize += AutoSystemInfo::PageSize;
- #endif
- return page;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- Page* Heap<TAlloc, TPreReservedAlloc>::AddPageToBucket(Page* page, BucketId bucket, bool wasFull)
- {
- Assert(bucket > BucketId::InvalidBucket && bucket < BucketId::NumBuckets);
- BucketId oldBucket = page->currentBucket;
- page->currentBucket = bucket;
- if (wasFull)
- {
- #pragma prefast(suppress: __WARNING_UNCHECKED_LOWER_BOUND_FOR_ENUMINDEX, "targetBucket is always in range >= SmallObjectList, but an __in_range doesn't fix the warning.");
- Assert(page->inFullList);
- this->fullPages[oldBucket].MoveElementTo(page, &this->buckets[bucket]);
- page->inFullList = false;
- }
- else
- {
- Assert(!page->inFullList);
- #pragma prefast(suppress: __WARNING_UNCHECKED_LOWER_BOUND_FOR_ENUMINDEX, "targetBucket is always in range >= SmallObjectList, but an __in_range doesn't fix the warning.");
- this->buckets[oldBucket].MoveElementTo(page, &this->buckets[bucket]);
- }
- return page;
- }
- /*
- * This method goes through the buckets greater than the target bucket
- * and if the higher bucket has a page with enough free space to allocate
- * something in the smaller bucket, then we bring the page to the smaller
- * bucket.
- * Note that if we allocate something from a page in the given bucket,
- * and then that page is split into a lower bucket, freeing is still not
- * a problem since the larger allocation is a multiple of the smaller one.
- * This gets more complicated if we can coalesce buckets. In that case,
- * we need to make sure that if a page was coalesced, and an allocation
- * pre-coalescing was freed, the page would need to get split upon free
- * to ensure correctness. For now, we've skipped implementing coalescing.
- * findPreReservedHeapPages - true, if we need to find pages only belonging to PreReservedHeapSegment
- */
- template<typename TAlloc, typename TPreReservedAlloc>
- Page* Heap<TAlloc, TPreReservedAlloc>::FindPageToSplit(BucketId targetBucket, bool findPreReservedHeapPages)
- {
- for (BucketId b = (BucketId)(targetBucket + 1); b < BucketId::NumBuckets; b = (BucketId) (b + 1))
- {
- #pragma prefast(suppress: __WARNING_UNCHECKED_LOWER_BOUND_FOR_ENUMINDEX, "targetBucket is always in range >= SmallObjectList, but an __in_range doesn't fix the warning.");
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, pageInBucket, &this->buckets[b], bucketIter)
- {
- Assert(!pageInBucket.inFullList);
- if (findPreReservedHeapPages && !this->codePageAllocators->IsPreReservedSegment(pageInBucket.segment))
- {
- //Find only pages that are pre-reserved using preReservedHeapPageAllocator
- continue;
- }
- if (pageInBucket.CanAllocate(targetBucket))
- {
- Page* page = &pageInBucket;
- if (findPreReservedHeapPages)
- {
- VerboseHeapTrace(_u("PRE-RESERVE: Found page for splitting in Pre Reserved Segment\n"));
- }
- VerboseHeapTrace(_u("Found page to split. Moving from bucket %d to %d\n"), b, targetBucket);
- return AddPageToBucket(page, targetBucket);
- }
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- return nullptr;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- BVIndex Heap<TAlloc, TPreReservedAlloc>::GetIndexInPage(__in Page* page, __in char* address)
- {
- Assert(page->address <= address && address < page->address + AutoSystemInfo::PageSize);
- return (BVIndex) ((address - page->address) / Page::Alignment);
- }
- #pragma endregion
- /**
- * Free List methods
- */
- #pragma region "Freeing methods"
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::FreeAllocation(Allocation* object)
- {
- Page* page = object->page;
- void* segment = page->segment;
- size_t pageSize = AutoSystemInfo::PageSize;
- unsigned int length = GetChunkSizeForBytes(object->size);
- BVIndex index = GetIndexInPage(page, object->address);
- uint freeBitsCount = page->freeBitVector.Count();
- // Make sure that the section under interest or the whole page has not already been freed
- if (page->IsEmpty() || page->freeBitVector.TestAnyInRange(index, length))
- {
- CustomHeap_BadPageState_unrecoverable_error((ULONG_PTR)this);
- return false;
- }
- if (page->inFullList)
- {
- VerboseHeapTrace(_u("Recycling page 0x%p because address 0x%p of size %d was freed\n"), page->address, object->address, object->size);
-
- // If the object being freed is equal to the page size, we're
- // going to remove it anyway so don't add it to a bucket
- if (object->size != pageSize)
- {
- AddPageToBucket(page, page->currentBucket, true);
- }
- else
- {
- EnsureAllocationWriteable(object);
- // Fill the old buffer with debug breaks
- char* localAddr = this->codePageAllocators->AllocLocal(object->address, object->size, page->segment);
- if (!localAddr)
- {
- MemoryOperationLastError::RecordError(JSERR_FatalMemoryExhaustion);
- return false;
- }
- FillDebugBreak((BYTE*)localAddr, object->size);
- this->codePageAllocators->FreeLocal(localAddr, page->segment);
- void* pageAddress = page->address;
- this->fullPages[page->currentBucket].RemoveElement(this->auxiliaryAllocator, page);
- // The page is not in any bucket- just update the stats, free the allocation
- // and dump the page- we don't need to update free object size since the object
- // size is equal to the page size so they cancel each other out
- #if DBG_DUMP
- this->totalAllocationSize -= pageSize;
- #endif
- this->auxiliaryAllocator->Free(object, sizeof(Allocation));
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- this->codePageAllocators->ReleasePages(pageAddress, 1, segment);
- }
- VerboseHeapTrace(_u("FastPath: freeing page-sized object directly\n"));
- return true;
- }
- }
- // If the page is about to become empty then we should not need
- // to set it to executable and we don't expect to restore the
- // previous protection settings.
- if (freeBitsCount == BVUnit::BitsPerWord - length)
- {
- EnsureAllocationWriteable(object);
- FreeAllocationHelper(object, index, length);
- Assert(page->IsEmpty());
- this->buckets[page->currentBucket].RemoveElement(this->auxiliaryAllocator, page);
- return false;
- }
- else
- {
- EnsureAllocationExecuteWriteable(object);
- FreeAllocationHelper(object, index, length);
- // after freeing part of the page, the page should be in PAGE_EXECUTE_READWRITE protection, and turning to PAGE_EXECUTE_READ (always with TARGETS_NO_UPDATE state)
- DWORD protectFlags = 0;
- if (AutoSystemInfo::Data.IsCFGEnabled())
- {
- protectFlags = PAGE_EXECUTE_RO_TARGETS_NO_UPDATE;
- }
- else
- {
- protectFlags = PAGE_EXECUTE_READ;
- }
- this->codePageAllocators->ProtectPages(page->address, 1, segment, protectFlags, PAGE_EXECUTE_READWRITE);
- return true;
- }
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeAllocationHelper(Allocation* object, BVIndex index, uint length)
- {
- Page* page = object->page;
- // Fill the old buffer with debug breaks
- char* localAddr = this->codePageAllocators->AllocLocal(object->address, object->size, page->segment);
- if (localAddr)
- {
- FillDebugBreak((BYTE*)localAddr, object->size);
- this->codePageAllocators->FreeLocal(localAddr, page->segment);
- }
- else
- {
- MemoryOperationLastError::RecordError(JSERR_FatalMemoryExhaustion);
- return;
- }
- VerboseHeapTrace(_u("Setting %d bits starting at bit %d, Free bit vector in page was "), length, index);
- #if VERBOSE_HEAP
- page->freeBitVector.DumpWord();
- #endif
- VerboseHeapTrace(_u("\n"));
- page->freeBitVector.SetRange(index, length);
- VerboseHeapTrace(_u("Free bit vector in page: "), length, index);
- #if VERBOSE_HEAP
- page->freeBitVector.DumpWord();
- #endif
- VerboseHeapTrace(_u("\n"));
- #if DBG_DUMP
- this->freeObjectSize += object->size;
- this->freesSinceLastCompact += object->size;
- #endif
- this->auxiliaryAllocator->Free(object, sizeof(Allocation));
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeDecommittedBuckets()
- {
- // CodePageAllocators is locked in FreeAll
- Assert(inDtor);
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, &this->decommittedPages, iter)
- {
- this->codePageAllocators->TrackDecommittedPages(page.address, 1, page.segment);
- iter.RemoveCurrent(this->auxiliaryAllocator);
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreePage(Page* page)
- {
- // CodePageAllocators is locked in FreeAll
- Assert(inDtor);
- DWORD pageSize = AutoSystemInfo::PageSize;
- EnsurePageWriteable(page);
- size_t freeSpace = page->freeBitVector.Count() * Page::Alignment;
- VerboseHeapTrace(_u("Removing page in bucket %d, freeSpace: %d\n"), page->currentBucket, freeSpace);
- this->codePageAllocators->ReleasePages(page->address, 1, page->segment);
- #if DBG_DUMP
- this->freeObjectSize -= freeSpace;
- this->totalAllocationSize -= pageSize;
- #endif
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeBucket(DListBase<Page>* bucket, bool freeOnlyEmptyPages)
- {
- // CodePageAllocators is locked in FreeAll
- Assert(inDtor);
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, bucket, pageIter)
- {
- // Templatize this to remove branches/make code more compact?
- if (!freeOnlyEmptyPages || page.IsEmpty())
- {
- FreePage(&page);
- pageIter.RemoveCurrent(this->auxiliaryAllocator);
- }
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeBuckets(bool freeOnlyEmptyPages)
- {
- // CodePageAllocators is locked in FreeAll
- Assert(inDtor);
- for (int i = 0; i < NumBuckets; i++)
- {
- FreeBucket(&this->buckets[i], freeOnlyEmptyPages);
- FreeBucket(&this->fullPages[i], freeOnlyEmptyPages);
- }
- #if DBG_DUMP
- this->allocationsSinceLastCompact = 0;
- this->freesSinceLastCompact = 0;
- #endif
- }
- template<typename TAlloc, typename TPreReservedAlloc>
- bool Heap<TAlloc, TPreReservedAlloc>::UpdateFullPages()
- {
- bool updated = false;
- if (this->codePageAllocators->HasSecondaryAllocStateChanged(&lastSecondaryAllocStateChangedCount))
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- for (int bucket = 0; bucket < BucketId::NumBuckets; bucket++)
- {
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, &(this->fullPages[bucket]), bucketIter)
- {
- Assert(page.inFullList);
- if (!this->ShouldBeInFullList(&page))
- {
- VerboseHeapTrace(_u("Recycling page 0x%p because XDATA was freed\n"), page.address);
- bucketIter.MoveCurrentTo(&(this->buckets[bucket]));
- page.inFullList = false;
- updated = true;
- }
- }
- NEXT_DLISTBASE_ENTRY_EDITING;
- }
- }
- return updated;
- }
- #if PDATA_ENABLED
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::FreeXdata(XDataAllocation* xdata, void* segment)
- {
- Assert(!xdata->IsFreed());
- {
- AutoCriticalSection autoLock(&this->codePageAllocators->cs);
- this->codePageAllocators->ReleaseSecondary(*xdata, segment);
- xdata->Free();
- }
- }
- #endif
- #if DBG_DUMP
- template<typename TAlloc, typename TPreReservedAlloc>
- void Heap<TAlloc, TPreReservedAlloc>::DumpStats()
- {
- HeapTrace(_u("Total allocation size: %d\n"), totalAllocationSize);
- HeapTrace(_u("Total free size: %d\n"), freeObjectSize);
- HeapTrace(_u("Total allocations since last compact: %d\n"), allocationsSinceLastCompact);
- HeapTrace(_u("Total frees since last compact: %d\n"), freesSinceLastCompact);
- HeapTrace(_u("Large object count: %d\n"), this->largeObjectAllocations.Count());
- HeapTrace(_u("Buckets: \n"));
- for (int i = 0; i < BucketId::NumBuckets; i++)
- {
- Output::Print(_u("\t%d => %u ["), (1 << (i + 7)), buckets[i].Count());
- FOREACH_DLISTBASE_ENTRY_EDITING(Page, page, &this->buckets[i], bucketIter)
- {
- BVUnit usedBitVector = page.freeBitVector;
- usedBitVector.ComplimentAll(); // Get the actual used bit vector
- Output::Print(_u(" %u "), usedBitVector.Count() * Page::Alignment); // Print out the space used in this page
- }
- NEXT_DLISTBASE_ENTRY_EDITING
- Output::Print(_u("] {{%u}}\n"), this->fullPages[i].Count());
- }
- }
- #endif
- #pragma endregion
- /**
- * Helper methods
- */
- #pragma region "Helpers"
- inline unsigned int log2(size_t number)
- {
- const unsigned int b[] = {0x2, 0xC, 0xF0, 0xFF00, 0xFFFF0000};
- const unsigned int S[] = {1, 2, 4, 8, 16};
- unsigned int result = 0;
- for (int i = 4; i >= 0; i--)
- {
- if (number & b[i])
- {
- number >>= S[i];
- result |= S[i];
- }
- }
- return result;
- }
- inline BucketId GetBucketForSize(size_t bytes)
- {
- if (bytes > Page::MaxAllocationSize)
- {
- return BucketId::LargeObjectList;
- }
- BucketId bucket = (BucketId) (log2(bytes / Page::sizePerBit));
- Assert(bucket < BucketId::LargeObjectList);
- Assert(bucket >= BucketId::SmallObjectList);
- return bucket;
- }
- // Fills the specified buffer with "debug break" instruction encoding.
- // If there is any space left after that due to alignment, fill it with 0.
- // static
- void FillDebugBreak(_Out_writes_bytes_all_(byteCount) BYTE* buffer, _In_ size_t byteCount)
- {
- #if defined(_M_ARM)
- // On ARM there is breakpoint instruction (BKPT) which is 0xBEii, where ii (immediate 8) can be any value, 0xBE in particular.
- // While it could be easier to put 0xBE (same way as 0xCC on x86), BKPT is not recommended -- it may cause unexpected side effects.
- // So, use same sequence are C++ compiler uses (0xDEFE), this is recognized by debugger as __debugbreak.
- // This is 2 bytes, and in case there is a gap of 1 byte in the end, fill it with 0 (there is no 1 byte long THUMB instruction).
- CompileAssert(sizeof(char16) == 2);
- char16 pattern = 0xDEFE;
- BYTE * writeBuffer = buffer;
- wmemset((char16 *)writeBuffer, pattern, byteCount / 2);
- if (byteCount % 2)
- {
- // Note: this is valid scenario: in JIT mode, we may not be 2-byte-aligned in the end of unwind info.
- *(writeBuffer + byteCount - 1) = 0; // Fill last remaining byte.
- }
- #elif defined(_M_ARM64)
- CompileAssert(sizeof(DWORD) == 4);
- DWORD pattern = 0xd4200000 | (0xf000 << 5);
- for (size_t i = 0; i < byteCount / 4; i++)
- {
- reinterpret_cast<DWORD*>(buffer)[i] = pattern;
- }
- for (size_t i = (byteCount / 4) * 4; i < byteCount; i++)
- {
- // Note: this is valid scenario: in JIT mode, we may not be 2-byte-aligned in the end of unwind info.
- buffer[i] = 0; // Fill last remaining bytes.
- }
- #else
- // On Intel just use "INT 3" instruction which is 0xCC.
- memset(buffer, 0xCC, byteCount);
- #endif
- }
- template class Heap<VirtualAllocWrapper, PreReservedVirtualAllocWrapper>;
- #if ENABLE_OOP_NATIVE_CODEGEN
- template class Heap<SectionAllocWrapper, PreReservedSectionAllocWrapper>;
- #endif
- #pragma endregion
- template<>
- char *
- CodePageAllocators<VirtualAllocWrapper, PreReservedVirtualAllocWrapper>::AllocLocal(char * remoteAddr, size_t size, void * segment)
- {
- return remoteAddr;
- }
- template<>
- void
- CodePageAllocators<VirtualAllocWrapper, PreReservedVirtualAllocWrapper>::FreeLocal(char * localAddr, void * segment)
- {
- // do nothing in case we are in proc
- }
- #if ENABLE_OOP_NATIVE_CODEGEN
- template<>
- char *
- CodePageAllocators<SectionAllocWrapper, PreReservedSectionAllocWrapper>::AllocLocal(char * remoteAddr, size_t size, void * segment)
- {
- AutoCriticalSection autoLock(&this->cs);
- Assert(segment);
- LPVOID address = nullptr;
- if (IsPreReservedSegment(segment))
- {
- address = ((SegmentBase<PreReservedSectionAllocWrapper>*)segment)->GetAllocator()->GetVirtualAllocator()->AllocLocal(remoteAddr, size);
- }
- else
- {
- address = ((SegmentBase<SectionAllocWrapper>*)segment)->GetAllocator()->GetVirtualAllocator()->AllocLocal(remoteAddr, size);
- }
- return (char*)address;
- }
- template<>
- void
- CodePageAllocators<SectionAllocWrapper, PreReservedSectionAllocWrapper>::FreeLocal(char * localAddr, void * segment)
- {
- AutoCriticalSection autoLock(&this->cs);
- Assert(segment);
- if (IsPreReservedSegment(segment))
- {
- ((SegmentBase<PreReservedSectionAllocWrapper>*)segment)->GetAllocator()->GetVirtualAllocator()->FreeLocal(localAddr);
- }
- else
- {
- ((SegmentBase<SectionAllocWrapper>*)segment)->GetAllocator()->GetVirtualAllocator()->FreeLocal(localAddr);
- }
- }
- #endif
- } // namespace CustomHeap
- } // namespace Memory
- #endif // ENABLE_NATIVE_CODEGEN || DYNAMIC_INTERPRETER_THUNK
|