| // Copyright 2016 The SwiftShader Authors. All Rights Reserved. | 
 | // | 
 | // Licensed under the Apache License, Version 2.0 (the "License"); | 
 | // you may not use this file except in compliance with the License. | 
 | // You may obtain a copy of the License at | 
 | // | 
 | //    http://www.apache.org/licenses/LICENSE-2.0 | 
 | // | 
 | // Unless required by applicable law or agreed to in writing, software | 
 | // distributed under the License is distributed on an "AS IS" BASIS, | 
 | // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
 | // See the License for the specific language governing permissions and | 
 | // limitations under the License. | 
 |  | 
 | #include "PoolAlloc.h" | 
 |  | 
 | #ifndef _MSC_VER | 
 | #include <stdint.h> | 
 | #endif | 
 | #include <stdio.h> | 
 | #include <stdlib.h> | 
 |  | 
 | #include "InitializeGlobals.h" | 
 | #include "osinclude.h" | 
 |  | 
 | OS_TLSIndex PoolIndex = OS_INVALID_TLS_INDEX; | 
 |  | 
 | bool InitializePoolIndex() | 
 | { | 
 | 	assert(PoolIndex == OS_INVALID_TLS_INDEX); | 
 |  | 
 | 	PoolIndex = OS_AllocTLSIndex(); | 
 | 	return PoolIndex != OS_INVALID_TLS_INDEX; | 
 | } | 
 |  | 
 | void FreePoolIndex() | 
 | { | 
 | 	assert(PoolIndex != OS_INVALID_TLS_INDEX); | 
 |  | 
 | 	OS_FreeTLSIndex(PoolIndex); | 
 | 	PoolIndex = OS_INVALID_TLS_INDEX; | 
 | } | 
 |  | 
 | TPoolAllocator* GetGlobalPoolAllocator() | 
 | { | 
 | 	assert(PoolIndex != OS_INVALID_TLS_INDEX); | 
 | 	return static_cast<TPoolAllocator*>(OS_GetTLSValue(PoolIndex)); | 
 | } | 
 |  | 
 | void SetGlobalPoolAllocator(TPoolAllocator* poolAllocator) | 
 | { | 
 | 	assert(PoolIndex != OS_INVALID_TLS_INDEX); | 
 | 	OS_SetTLSValue(PoolIndex, poolAllocator); | 
 | } | 
 |  | 
 | // | 
 | // Implement the functionality of the TPoolAllocator class, which | 
 | // is documented in PoolAlloc.h. | 
 | // | 
 | TPoolAllocator::TPoolAllocator(int growthIncrement, int allocationAlignment) : | 
 | 	alignment(allocationAlignment) | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	, pageSize(growthIncrement), | 
 | 	freeList(0), | 
 | 	inUseList(0), | 
 | 	numCalls(0), | 
 | 	totalBytes(0) | 
 | #endif | 
 | { | 
 | 	// | 
 | 	// Adjust alignment to be at least pointer aligned and | 
 | 	// power of 2. | 
 | 	// | 
 | 	size_t minAlign = sizeof(void*); | 
 | 	alignment &= ~(minAlign - 1); | 
 | 	if (alignment < minAlign) | 
 | 		alignment = minAlign; | 
 | 	size_t a = 1; | 
 | 	while (a < alignment) | 
 | 		a <<= 1; | 
 | 	alignment = a; | 
 | 	alignmentMask = a - 1; | 
 |  | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	// | 
 | 	// Don't allow page sizes we know are smaller than all common | 
 | 	// OS page sizes. | 
 | 	// | 
 | 	if (pageSize < 4*1024) | 
 | 		pageSize = 4*1024; | 
 |  | 
 | 	// | 
 | 	// A large currentPageOffset indicates a new page needs to | 
 | 	// be obtained to allocate memory. | 
 | 	// | 
 | 	currentPageOffset = pageSize; | 
 |  | 
 | 	// | 
 | 	// Align header skip | 
 | 	// | 
 | 	headerSkip = minAlign; | 
 | 	if (headerSkip < sizeof(tHeader)) { | 
 | 		headerSkip = (sizeof(tHeader) + alignmentMask) & ~alignmentMask; | 
 | 	} | 
 | #else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	mStack.push_back({}); | 
 | #endif | 
 | } | 
 |  | 
 | TPoolAllocator::~TPoolAllocator() | 
 | { | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	while (inUseList) { | 
 | 		tHeader* next = inUseList->nextPage; | 
 | 		inUseList->~tHeader(); | 
 | 		delete [] reinterpret_cast<char*>(inUseList); | 
 | 		inUseList = next; | 
 | 	} | 
 |  | 
 | 	// We should not check the guard blocks | 
 | 	// here, because we did it already when the block was | 
 | 	// placed into the free list. | 
 | 	// | 
 | 	while (freeList) { | 
 | 		tHeader* next = freeList->nextPage; | 
 | 		delete [] reinterpret_cast<char*>(freeList); | 
 | 		freeList = next; | 
 | 	} | 
 | #else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	for (auto& allocs : mStack) { | 
 | 		for (auto alloc : allocs) { | 
 | 			free(alloc); | 
 | 		} | 
 | 	} | 
 | 	mStack.clear(); | 
 | #endif | 
 | } | 
 |  | 
 | // Support MSVC++ 6.0 | 
 | const unsigned char TAllocation::guardBlockBeginVal = 0xfb; | 
 | const unsigned char TAllocation::guardBlockEndVal   = 0xfe; | 
 | const unsigned char TAllocation::userDataFill       = 0xcd; | 
 |  | 
 | #ifdef GUARD_BLOCKS | 
 | 	const size_t TAllocation::guardBlockSize = 16; | 
 | #else | 
 | 	const size_t TAllocation::guardBlockSize = 0; | 
 | #endif | 
 |  | 
 | // | 
 | // Check a single guard block for damage | 
 | // | 
 | void TAllocation::checkGuardBlock(unsigned char* blockMem, unsigned char val, const char* locText) const | 
 | { | 
 | #ifdef GUARD_BLOCKS | 
 | 	for (size_t x = 0; x < guardBlockSize; x++) { | 
 | 		if (blockMem[x] != val) { | 
 | 			char assertMsg[80]; | 
 |  | 
 | 			// We don't print the assert message.  It's here just to be helpful. | 
 | 			#if defined(_MSC_VER) | 
 | 				_snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %Iu byte allocation at 0x%p\n", | 
 | 						  locText, size, data()); | 
 | 			#else | 
 | 				snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %zu byte allocation at 0x%p\n", | 
 | 						 locText, size, data()); | 
 | 			#endif | 
 | 			assert(0 && "PoolAlloc: Damage in guard block"); | 
 | 		} | 
 | 	} | 
 | #endif | 
 | } | 
 |  | 
 |  | 
 | void TPoolAllocator::push() | 
 | { | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	tAllocState state = { currentPageOffset, inUseList }; | 
 |  | 
 | 	mStack.push_back(state); | 
 |  | 
 | 	// | 
 | 	// Indicate there is no current page to allocate from. | 
 | 	// | 
 | 	currentPageOffset = pageSize; | 
 | #else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	mStack.push_back({}); | 
 | #endif | 
 | } | 
 |  | 
 | // | 
 | // Do a mass-deallocation of all the individual allocations | 
 | // that have occurred since the last push(), or since the | 
 | // last pop(), or since the object's creation. | 
 | // | 
 | // The deallocated pages are saved for future allocations. | 
 | // | 
 | void TPoolAllocator::pop() | 
 | { | 
 | 	if (mStack.size() < 1) | 
 | 		return; | 
 |  | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	tHeader* page = mStack.back().page; | 
 | 	currentPageOffset = mStack.back().offset; | 
 |  | 
 | 	while (inUseList != page) { | 
 | 		// invoke destructor to free allocation list | 
 | 		inUseList->~tHeader(); | 
 |  | 
 | 		tHeader* nextInUse = inUseList->nextPage; | 
 | 		if (inUseList->pageCount > 1) | 
 | 			delete [] reinterpret_cast<char*>(inUseList); | 
 | 		else { | 
 | 			inUseList->nextPage = freeList; | 
 | 			freeList = inUseList; | 
 | 		} | 
 | 		inUseList = nextInUse; | 
 | 	} | 
 |  | 
 | 	mStack.pop_back(); | 
 | #else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	for (auto alloc : mStack.back()) { | 
 | 		free(alloc); | 
 | 	} | 
 | 	mStack.pop_back(); | 
 | #endif | 
 | } | 
 |  | 
 | // | 
 | // Do a mass-deallocation of all the individual allocations | 
 | // that have occurred. | 
 | // | 
 | void TPoolAllocator::popAll() | 
 | { | 
 | 	while (mStack.size() > 0) | 
 | 		pop(); | 
 | } | 
 |  | 
 | void* TPoolAllocator::allocate(size_t numBytes) | 
 | { | 
 | #if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	// | 
 | 	// Just keep some interesting statistics. | 
 | 	// | 
 | 	++numCalls; | 
 | 	totalBytes += numBytes; | 
 |  | 
 | 	// If we are using guard blocks, all allocations are bracketed by | 
 | 	// them: [guardblock][allocation][guardblock].  numBytes is how | 
 | 	// much memory the caller asked for.  allocationSize is the total | 
 | 	// size including guard blocks.  In release build, | 
 | 	// guardBlockSize=0 and this all gets optimized away. | 
 | 	size_t allocationSize = TAllocation::allocationSize(numBytes); | 
 | 	// Detect integer overflow. | 
 | 	if (allocationSize < numBytes) | 
 | 		return 0; | 
 |  | 
 | 	// | 
 | 	// Do the allocation, most likely case first, for efficiency. | 
 | 	// This step could be moved to be inline sometime. | 
 | 	// | 
 | 	if (allocationSize <= pageSize - currentPageOffset) { | 
 | 		// | 
 | 		// Safe to allocate from currentPageOffset. | 
 | 		// | 
 | 		unsigned char* memory = reinterpret_cast<unsigned char *>(inUseList) + currentPageOffset; | 
 | 		currentPageOffset += allocationSize; | 
 | 		currentPageOffset = (currentPageOffset + alignmentMask) & ~alignmentMask; | 
 |  | 
 | 		return initializeAllocation(inUseList, memory, numBytes); | 
 | 	} | 
 |  | 
 | 	if (allocationSize > pageSize - headerSkip) { | 
 | 		// | 
 | 		// Do a multi-page allocation.  Don't mix these with the others. | 
 | 		// The OS is efficient and allocating and free-ing multiple pages. | 
 | 		// | 
 | 		size_t numBytesToAlloc = allocationSize + headerSkip; | 
 | 		// Detect integer overflow. | 
 | 		if (numBytesToAlloc < allocationSize) | 
 | 			return 0; | 
 |  | 
 | 		tHeader* memory = reinterpret_cast<tHeader*>(::new char[numBytesToAlloc]); | 
 | 		if (memory == 0) | 
 | 			return 0; | 
 |  | 
 | 		// Use placement-new to initialize header | 
 | 		new(memory) tHeader(inUseList, (numBytesToAlloc + pageSize - 1) / pageSize); | 
 | 		inUseList = memory; | 
 |  | 
 | 		currentPageOffset = pageSize;  // make next allocation come from a new page | 
 |  | 
 | 		// No guard blocks for multi-page allocations (yet) | 
 | 		return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(memory) + headerSkip); | 
 | 	} | 
 |  | 
 | 	// | 
 | 	// Need a simple page to allocate from. | 
 | 	// | 
 | 	tHeader* memory; | 
 | 	if (freeList) { | 
 | 		memory = freeList; | 
 | 		freeList = freeList->nextPage; | 
 | 	} else { | 
 | 		memory = reinterpret_cast<tHeader*>(::new char[pageSize]); | 
 | 		if (memory == 0) | 
 | 			return 0; | 
 | 	} | 
 |  | 
 | 	// Use placement-new to initialize header | 
 | 	new(memory) tHeader(inUseList, 1); | 
 | 	inUseList = memory; | 
 |  | 
 | 	unsigned char* ret = reinterpret_cast<unsigned char *>(inUseList) + headerSkip; | 
 | 	currentPageOffset = (headerSkip + allocationSize + alignmentMask) & ~alignmentMask; | 
 |  | 
 | 	return initializeAllocation(inUseList, ret, numBytes); | 
 | #else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC) | 
 | 	void *alloc = malloc(numBytes + alignmentMask); | 
 | 	mStack.back().push_back(alloc); | 
 |  | 
 | 	intptr_t intAlloc = reinterpret_cast<intptr_t>(alloc); | 
 | 	intAlloc = (intAlloc + alignmentMask) & ~alignmentMask; | 
 | 	return reinterpret_cast<void *>(intAlloc); | 
 | #endif | 
 | } | 
 |  | 
 |  | 
 | // | 
 | // Check all allocations in a list for damage by calling check on each. | 
 | // | 
 | void TAllocation::checkAllocList() const | 
 | { | 
 | 	for (const TAllocation* alloc = this; alloc != 0; alloc = alloc->prevAlloc) | 
 | 		alloc->check(); | 
 | } |