| // Copyright 2016 The SwiftShader Authors. All Rights Reserved. |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| #include "PoolAlloc.h" |
| |
| #ifndef _MSC_VER |
| #include <stdint.h> |
| #endif |
| #include <stdio.h> |
| |
| #include "InitializeGlobals.h" |
| #include "osinclude.h" |
| |
| OS_TLSIndex PoolIndex = OS_INVALID_TLS_INDEX; |
| |
| bool InitializePoolIndex() |
| { |
| assert(PoolIndex == OS_INVALID_TLS_INDEX); |
| |
| PoolIndex = OS_AllocTLSIndex(); |
| return PoolIndex != OS_INVALID_TLS_INDEX; |
| } |
| |
| void FreePoolIndex() |
| { |
| assert(PoolIndex != OS_INVALID_TLS_INDEX); |
| |
| OS_FreeTLSIndex(PoolIndex); |
| PoolIndex = OS_INVALID_TLS_INDEX; |
| } |
| |
| TPoolAllocator* GetGlobalPoolAllocator() |
| { |
| assert(PoolIndex != OS_INVALID_TLS_INDEX); |
| return static_cast<TPoolAllocator*>(OS_GetTLSValue(PoolIndex)); |
| } |
| |
| void SetGlobalPoolAllocator(TPoolAllocator* poolAllocator) |
| { |
| assert(PoolIndex != OS_INVALID_TLS_INDEX); |
| OS_SetTLSValue(PoolIndex, poolAllocator); |
| } |
| |
| // |
| // Implement the functionality of the TPoolAllocator class, which |
| // is documented in PoolAlloc.h. |
| // |
| TPoolAllocator::TPoolAllocator(int growthIncrement, int allocationAlignment) : |
| alignment(allocationAlignment) |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| , pageSize(growthIncrement), |
| freeList(0), |
| inUseList(0), |
| numCalls(0), |
| totalBytes(0) |
| #endif |
| { |
| // |
| // Adjust alignment to be at least pointer aligned and |
| // power of 2. |
| // |
| size_t minAlign = sizeof(void*); |
| alignment &= ~(minAlign - 1); |
| if (alignment < minAlign) |
| alignment = minAlign; |
| size_t a = 1; |
| while (a < alignment) |
| a <<= 1; |
| alignment = a; |
| alignmentMask = a - 1; |
| |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| // |
| // Don't allow page sizes we know are smaller than all common |
| // OS page sizes. |
| // |
| if (pageSize < 4*1024) |
| pageSize = 4*1024; |
| |
| // |
| // A large currentPageOffset indicates a new page needs to |
| // be obtained to allocate memory. |
| // |
| currentPageOffset = pageSize; |
| |
| // |
| // Align header skip |
| // |
| headerSkip = minAlign; |
| if (headerSkip < sizeof(tHeader)) { |
| headerSkip = (sizeof(tHeader) + alignmentMask) & ~alignmentMask; |
| } |
| #else // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| mStack.push_back({}); |
| #endif |
| } |
| |
| TPoolAllocator::~TPoolAllocator() |
| { |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| while (inUseList) { |
| tHeader* next = inUseList->nextPage; |
| inUseList->~tHeader(); |
| delete [] reinterpret_cast<char*>(inUseList); |
| inUseList = next; |
| } |
| |
| // We should not check the guard blocks |
| // here, because we did it already when the block was |
| // placed into the free list. |
| // |
| while (freeList) { |
| tHeader* next = freeList->nextPage; |
| delete [] reinterpret_cast<char*>(freeList); |
| freeList = next; |
| } |
| #else // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| for (auto& allocs : mStack) { |
| for (auto alloc : allocs) { |
| free(alloc); |
| } |
| } |
| mStack.clear(); |
| #endif |
| } |
| |
| // Support MSVC++ 6.0 |
| const unsigned char TAllocation::guardBlockBeginVal = 0xfb; |
| const unsigned char TAllocation::guardBlockEndVal = 0xfe; |
| const unsigned char TAllocation::userDataFill = 0xcd; |
| |
| #ifdef GUARD_BLOCKS |
| const size_t TAllocation::guardBlockSize = 16; |
| #else |
| const size_t TAllocation::guardBlockSize = 0; |
| #endif |
| |
| // |
| // Check a single guard block for damage |
| // |
| void TAllocation::checkGuardBlock(unsigned char* blockMem, unsigned char val, const char* locText) const |
| { |
| #ifdef GUARD_BLOCKS |
| for (size_t x = 0; x < guardBlockSize; x++) { |
| if (blockMem[x] != val) { |
| char assertMsg[80]; |
| |
| // We don't print the assert message. It's here just to be helpful. |
| #if defined(_MSC_VER) |
| _snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %Iu byte allocation at 0x%p\n", |
| locText, size, data()); |
| #else |
| snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %zu byte allocation at 0x%p\n", |
| locText, size, data()); |
| #endif |
| assert(0 && "PoolAlloc: Damage in guard block"); |
| } |
| } |
| #endif |
| } |
| |
| |
| void TPoolAllocator::push() |
| { |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| tAllocState state = { currentPageOffset, inUseList }; |
| |
| mStack.push_back(state); |
| |
| // |
| // Indicate there is no current page to allocate from. |
| // |
| currentPageOffset = pageSize; |
| #else // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| mStack.push_back({}); |
| #endif |
| } |
| |
| // |
| // Do a mass-deallocation of all the individual allocations |
| // that have occurred since the last push(), or since the |
| // last pop(), or since the object's creation. |
| // |
| // The deallocated pages are saved for future allocations. |
| // |
| void TPoolAllocator::pop() |
| { |
| if (mStack.size() < 1) |
| return; |
| |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| tHeader* page = mStack.back().page; |
| currentPageOffset = mStack.back().offset; |
| |
| while (inUseList != page) { |
| // invoke destructor to free allocation list |
| inUseList->~tHeader(); |
| |
| tHeader* nextInUse = inUseList->nextPage; |
| if (inUseList->pageCount > 1) |
| delete [] reinterpret_cast<char*>(inUseList); |
| else { |
| inUseList->nextPage = freeList; |
| freeList = inUseList; |
| } |
| inUseList = nextInUse; |
| } |
| |
| mStack.pop_back(); |
| #else // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| for (auto alloc : mStack.back()) { |
| free(alloc); |
| } |
| mStack.pop_back(); |
| #endif |
| } |
| |
| // |
| // Do a mass-deallocation of all the individual allocations |
| // that have occurred. |
| // |
| void TPoolAllocator::popAll() |
| { |
| while (mStack.size() > 0) |
| pop(); |
| } |
| |
| void* TPoolAllocator::allocate(size_t numBytes) |
| { |
| #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| // |
| // Just keep some interesting statistics. |
| // |
| ++numCalls; |
| totalBytes += numBytes; |
| |
| // If we are using guard blocks, all allocations are bracketed by |
| // them: [guardblock][allocation][guardblock]. numBytes is how |
| // much memory the caller asked for. allocationSize is the total |
| // size including guard blocks. In release build, |
| // guardBlockSize=0 and this all gets optimized away. |
| size_t allocationSize = TAllocation::allocationSize(numBytes); |
| // Detect integer overflow. |
| if (allocationSize < numBytes) |
| return 0; |
| |
| // |
| // Do the allocation, most likely case first, for efficiency. |
| // This step could be moved to be inline sometime. |
| // |
| if (allocationSize <= pageSize - currentPageOffset) { |
| // |
| // Safe to allocate from currentPageOffset. |
| // |
| unsigned char* memory = reinterpret_cast<unsigned char *>(inUseList) + currentPageOffset; |
| currentPageOffset += allocationSize; |
| currentPageOffset = (currentPageOffset + alignmentMask) & ~alignmentMask; |
| |
| return initializeAllocation(inUseList, memory, numBytes); |
| } |
| |
| if (allocationSize > pageSize - headerSkip) { |
| // |
| // Do a multi-page allocation. Don't mix these with the others. |
| // The OS is efficient and allocating and free-ing multiple pages. |
| // |
| size_t numBytesToAlloc = allocationSize + headerSkip; |
| // Detect integer overflow. |
| if (numBytesToAlloc < allocationSize) |
| return 0; |
| |
| tHeader* memory = reinterpret_cast<tHeader*>(::new char[numBytesToAlloc]); |
| if (memory == 0) |
| return 0; |
| |
| // Use placement-new to initialize header |
| new(memory) tHeader(inUseList, (numBytesToAlloc + pageSize - 1) / pageSize); |
| inUseList = memory; |
| |
| currentPageOffset = pageSize; // make next allocation come from a new page |
| |
| // No guard blocks for multi-page allocations (yet) |
| return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(memory) + headerSkip); |
| } |
| |
| // |
| // Need a simple page to allocate from. |
| // |
| tHeader* memory; |
| if (freeList) { |
| memory = freeList; |
| freeList = freeList->nextPage; |
| } else { |
| memory = reinterpret_cast<tHeader*>(::new char[pageSize]); |
| if (memory == 0) |
| return 0; |
| } |
| |
| // Use placement-new to initialize header |
| new(memory) tHeader(inUseList, 1); |
| inUseList = memory; |
| |
| unsigned char* ret = reinterpret_cast<unsigned char *>(inUseList) + headerSkip; |
| currentPageOffset = (headerSkip + allocationSize + alignmentMask) & ~alignmentMask; |
| |
| return initializeAllocation(inUseList, ret, numBytes); |
| #else // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) |
| void *alloc = malloc(numBytes + alignmentMask); |
| mStack.back().push_back(alloc); |
| |
| intptr_t intAlloc = reinterpret_cast<intptr_t>(alloc); |
| intAlloc = (intAlloc + alignmentMask) & ~alignmentMask; |
| return reinterpret_cast<void *>(intAlloc); |
| #endif |
| } |
| |
| |
| // |
| // Check all allocations in a list for damage by calling check on each. |
| // |
| void TAllocation::checkAllocList() const |
| { |
| for (const TAllocation* alloc = this; alloc != 0; alloc = alloc->prevAlloc) |
| alloc->check(); |
| } |