blob: 161cb0a21783e7bc27dec6108deff4046e71537a [file] [log] [blame] [edit]
// Copyright 2018 The SwiftShader Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "VkDevice.hpp"
#include "VkConfig.hpp"
#include "VkDescriptorSetLayout.hpp"
#include "VkFence.hpp"
#include "VkQueue.hpp"
#include "VkSemaphore.hpp"
#include "VkTimelineSemaphore.hpp"
#include "Debug/Context.hpp"
#include "Debug/Server.hpp"
#include "Device/Blitter.hpp"
#include "System/Debug.hpp"
#include <chrono>
#include <climits>
#include <new> // Must #include this to use "placement new"
namespace {
using time_point = std::chrono::time_point<std::chrono::system_clock, std::chrono::nanoseconds>;
time_point now()
{
return std::chrono::time_point_cast<std::chrono::nanoseconds>(std::chrono::system_clock::now());
}
const time_point getEndTimePoint(uint64_t timeout, bool &infiniteTimeout)
{
const time_point start = now();
const uint64_t max_timeout = (LLONG_MAX - start.time_since_epoch().count());
infiniteTimeout = (timeout > max_timeout);
return start + std::chrono::nanoseconds(std::min(max_timeout, timeout));
}
} // anonymous namespace
namespace vk {
void Device::SamplingRoutineCache::updateSnapshot()
{
marl::lock lock(mutex);
if(snapshotNeedsUpdate)
{
snapshot.clear();
for(auto it : cache)
{
snapshot[it.key()] = it.data();
}
snapshotNeedsUpdate = false;
}
}
Device::SamplerIndexer::~SamplerIndexer()
{
ASSERT(map.empty());
}
uint32_t Device::SamplerIndexer::index(const SamplerState &samplerState)
{
marl::lock lock(mutex);
auto it = map.find(samplerState);
if(it != map.end())
{
it->second.count++;
return it->second.id;
}
nextID++;
map.emplace(samplerState, Identifier{ nextID, 1 });
return nextID;
}
void Device::SamplerIndexer::remove(const SamplerState &samplerState)
{
marl::lock lock(mutex);
auto it = map.find(samplerState);
ASSERT(it != map.end());
auto count = --it->second.count;
if(count == 0)
{
map.erase(it);
}
}
const SamplerState *Device::SamplerIndexer::find(uint32_t id)
{
marl::lock lock(mutex);
auto it = std::find_if(std::begin(map), std::end(map),
[&id](auto &&p) { return p.second.id == id; });
return (it != std::end(map)) ? &(it->first) : nullptr;
}
Device::Device(const VkDeviceCreateInfo *pCreateInfo, void *mem, PhysicalDevice *physicalDevice, const VkPhysicalDeviceFeatures *enabledFeatures, const std::shared_ptr<marl::Scheduler> &scheduler)
: physicalDevice(physicalDevice)
, queues(reinterpret_cast<Queue *>(mem))
, enabledExtensionCount(pCreateInfo->enabledExtensionCount)
, enabledFeatures(enabledFeatures ? *enabledFeatures : VkPhysicalDeviceFeatures{})
, // "Setting pEnabledFeatures to NULL and not including a VkPhysicalDeviceFeatures2 in the pNext member of VkDeviceCreateInfo is equivalent to setting all members of the structure to VK_FALSE."
scheduler(scheduler)
{
for(uint32_t i = 0; i < pCreateInfo->queueCreateInfoCount; i++)
{
const VkDeviceQueueCreateInfo &queueCreateInfo = pCreateInfo->pQueueCreateInfos[i];
queueCount += queueCreateInfo.queueCount;
}
uint32_t queueID = 0;
for(uint32_t i = 0; i < pCreateInfo->queueCreateInfoCount; i++)
{
const VkDeviceQueueCreateInfo &queueCreateInfo = pCreateInfo->pQueueCreateInfos[i];
for(uint32_t j = 0; j < queueCreateInfo.queueCount; j++, queueID++)
{
new(&queues[queueID]) Queue(this, scheduler.get());
}
}
extensions = reinterpret_cast<ExtensionName *>(static_cast<uint8_t *>(mem) + (sizeof(Queue) * queueCount));
for(uint32_t i = 0; i < enabledExtensionCount; i++)
{
strncpy(extensions[i], pCreateInfo->ppEnabledExtensionNames[i], VK_MAX_EXTENSION_NAME_SIZE);
}
if(pCreateInfo->enabledLayerCount)
{
// "The ppEnabledLayerNames and enabledLayerCount members of VkDeviceCreateInfo are deprecated and their values must be ignored by implementations."
UNSUPPORTED("enabledLayerCount");
}
// FIXME (b/119409619): use an allocator here so we can control all memory allocations
blitter.reset(new sw::Blitter());
samplingRoutineCache.reset(new SamplingRoutineCache());
samplerIndexer.reset(new SamplerIndexer());
#ifdef ENABLE_VK_DEBUGGER
static auto port = getenv("VK_DEBUGGER_PORT");
if(port)
{
// Construct the debugger context and server - this may block for a
// debugger connection, allowing breakpoints to be set before they're
// executed.
debugger.context = vk::dbg::Context::create();
debugger.server = vk::dbg::Server::create(debugger.context, atoi(port));
}
#endif // ENABLE_VK_DEBUGGER
#ifdef SWIFTSHADER_DEVICE_MEMORY_REPORT
const VkBaseInStructure *extensionCreateInfo = reinterpret_cast<const VkBaseInStructure *>(pCreateInfo->pNext);
while(extensionCreateInfo)
{
if(extensionCreateInfo->sType == VK_STRUCTURE_TYPE_DEVICE_DEVICE_MEMORY_REPORT_CREATE_INFO_EXT)
{
auto deviceMemoryReportCreateInfo = reinterpret_cast<const VkDeviceDeviceMemoryReportCreateInfoEXT *>(pCreateInfo->pNext);
if(deviceMemoryReportCreateInfo->pfnUserCallback != nullptr)
{
deviceMemoryReportCallbacks.emplace_back(deviceMemoryReportCreateInfo->pfnUserCallback, deviceMemoryReportCreateInfo->pUserData);
}
}
extensionCreateInfo = extensionCreateInfo->pNext;
}
#endif // SWIFTSHADER_DEVICE_MEMORY_REPORT
}
void Device::destroy(const VkAllocationCallbacks *pAllocator)
{
for(uint32_t i = 0; i < queueCount; i++)
{
queues[i].~Queue();
}
vk::freeHostMemory(queues, pAllocator);
}
size_t Device::ComputeRequiredAllocationSize(const VkDeviceCreateInfo *pCreateInfo)
{
uint32_t queueCount = 0;
for(uint32_t i = 0; i < pCreateInfo->queueCreateInfoCount; i++)
{
queueCount += pCreateInfo->pQueueCreateInfos[i].queueCount;
}
return (sizeof(Queue) * queueCount) + (pCreateInfo->enabledExtensionCount * sizeof(ExtensionName));
}
bool Device::hasExtension(const char *extensionName) const
{
for(uint32_t i = 0; i < enabledExtensionCount; i++)
{
if(strncmp(extensions[i], extensionName, VK_MAX_EXTENSION_NAME_SIZE) == 0)
{
return true;
}
}
return false;
}
VkQueue Device::getQueue(uint32_t queueFamilyIndex, uint32_t queueIndex) const
{
ASSERT(queueFamilyIndex == 0);
return queues[queueIndex];
}
VkResult Device::waitForFences(uint32_t fenceCount, const VkFence *pFences, VkBool32 waitAll, uint64_t timeout)
{
bool infiniteTimeout = false;
const time_point end_ns = getEndTimePoint(timeout, infiniteTimeout);
if(waitAll != VK_FALSE) // All fences must be signaled
{
for(uint32_t i = 0; i < fenceCount; i++)
{
if(timeout == 0)
{
if(Cast(pFences[i])->getStatus() != VK_SUCCESS) // At least one fence is not signaled
{
return VK_TIMEOUT;
}
}
else if(infiniteTimeout)
{
if(Cast(pFences[i])->wait() != VK_SUCCESS) // At least one fence is not signaled
{
return VK_TIMEOUT;
}
}
else
{
if(Cast(pFences[i])->wait(end_ns) != VK_SUCCESS) // At least one fence is not signaled
{
return VK_TIMEOUT;
}
}
}
return VK_SUCCESS;
}
else // At least one fence must be signaled
{
marl::containers::vector<marl::Event, 8> events;
for(uint32_t i = 0; i < fenceCount; i++)
{
events.push_back(Cast(pFences[i])->getCountedEvent()->event());
}
auto any = marl::Event::any(events.begin(), events.end());
if(timeout == 0)
{
return any.isSignalled() ? VK_SUCCESS : VK_TIMEOUT;
}
else if(infiniteTimeout)
{
any.wait();
return VK_SUCCESS;
}
else
{
return any.wait_until(end_ns) ? VK_SUCCESS : VK_TIMEOUT;
}
}
}
VkResult Device::waitForSemaphores(const VkSemaphoreWaitInfo *pWaitInfo, uint64_t timeout)
{
bool infiniteTimeout = false;
const time_point end_ns = getEndTimePoint(timeout, infiniteTimeout);
if(pWaitInfo->flags & VK_SEMAPHORE_WAIT_ANY_BIT)
{
TimelineSemaphore any = TimelineSemaphore();
for(uint32_t i = 0; i < pWaitInfo->semaphoreCount; i++)
{
TimelineSemaphore *semaphore = DynamicCast<TimelineSemaphore>(pWaitInfo->pSemaphores[i]);
uint64_t waitValue = pWaitInfo->pValues[i];
if(semaphore->getCounterValue() == waitValue)
{
return VK_SUCCESS;
}
semaphore->addDependent(any, waitValue);
}
if(infiniteTimeout)
{
any.wait(1ull);
return VK_SUCCESS;
}
else
{
if(any.wait(1, end_ns) == VK_SUCCESS)
{
return VK_SUCCESS;
}
}
return VK_TIMEOUT;
}
else
{
ASSERT(pWaitInfo->flags == 0);
for(uint32_t i = 0; i < pWaitInfo->semaphoreCount; i++)
{
TimelineSemaphore *semaphore = DynamicCast<TimelineSemaphore>(pWaitInfo->pSemaphores[i]);
uint64_t value = pWaitInfo->pValues[i];
if(infiniteTimeout)
{
semaphore->wait(value);
}
else if(semaphore->wait(pWaitInfo->pValues[i], end_ns) != VK_SUCCESS)
{
return VK_TIMEOUT;
}
}
return VK_SUCCESS;
}
}
VkResult Device::waitIdle()
{
for(uint32_t i = 0; i < queueCount; i++)
{
queues[i].waitIdle();
}
return VK_SUCCESS;
}
void Device::getDescriptorSetLayoutSupport(const VkDescriptorSetLayoutCreateInfo *pCreateInfo,
VkDescriptorSetLayoutSupport *pSupport) const
{
// From Vulkan Spec 13.2.1 Descriptor Set Layout, in description of vkGetDescriptorSetLayoutSupport:
// "This command does not consider other limits such as maxPerStageDescriptor*, and so a descriptor
// set layout that is supported according to this command must still satisfy the pipeline layout limits
// such as maxPerStageDescriptor* in order to be used in a pipeline layout."
// We have no "strange" limitations to enforce beyond the device limits, so we can safely always claim support.
pSupport->supported = VK_TRUE;
}
void Device::updateDescriptorSets(uint32_t descriptorWriteCount, const VkWriteDescriptorSet *pDescriptorWrites,
uint32_t descriptorCopyCount, const VkCopyDescriptorSet *pDescriptorCopies)
{
for(uint32_t i = 0; i < descriptorWriteCount; i++)
{
DescriptorSetLayout::WriteDescriptorSet(this, pDescriptorWrites[i]);
}
for(uint32_t i = 0; i < descriptorCopyCount; i++)
{
DescriptorSetLayout::CopyDescriptorSet(pDescriptorCopies[i]);
}
}
void Device::getRequirements(VkMemoryDedicatedRequirements *requirements) const
{
requirements->prefersDedicatedAllocation = VK_FALSE;
requirements->requiresDedicatedAllocation = VK_FALSE;
}
Device::SamplingRoutineCache *Device::getSamplingRoutineCache() const
{
return samplingRoutineCache.get();
}
void Device::updateSamplingRoutineSnapshotCache()
{
samplingRoutineCache->updateSnapshot();
}
uint32_t Device::indexSampler(const SamplerState &samplerState)
{
return samplerIndexer->index(samplerState);
}
void Device::removeSampler(const SamplerState &samplerState)
{
samplerIndexer->remove(samplerState);
}
const SamplerState *Device::findSampler(uint32_t samplerId) const
{
return samplerIndexer->find(samplerId);
}
VkResult Device::setDebugUtilsObjectName(const VkDebugUtilsObjectNameInfoEXT *pNameInfo)
{
// Optionally maps user-friendly name to an object
return VK_SUCCESS;
}
VkResult Device::setDebugUtilsObjectTag(const VkDebugUtilsObjectTagInfoEXT *pTagInfo)
{
// Optionally attach arbitrary data to an object
return VK_SUCCESS;
}
void Device::registerImageView(ImageView *imageView)
{
if(imageView != nullptr)
{
marl::lock lock(imageViewSetMutex);
imageViewSet.insert(imageView);
}
}
void Device::unregisterImageView(ImageView *imageView)
{
if(imageView != nullptr)
{
marl::lock lock(imageViewSetMutex);
auto it = imageViewSet.find(imageView);
if(it != imageViewSet.end())
{
imageViewSet.erase(it);
}
}
}
void Device::prepareForSampling(ImageView *imageView)
{
if(imageView != nullptr)
{
marl::lock lock(imageViewSetMutex);
auto it = imageViewSet.find(imageView);
if(it != imageViewSet.end())
{
imageView->prepareForSampling();
}
}
}
void Device::contentsChanged(ImageView *imageView, Image::ContentsChangedContext context)
{
if(imageView != nullptr)
{
marl::lock lock(imageViewSetMutex);
auto it = imageViewSet.find(imageView);
if(it != imageViewSet.end())
{
imageView->contentsChanged(context);
}
}
}
#ifdef SWIFTSHADER_DEVICE_MEMORY_REPORT
void Device::emitDeviceMemoryReport(VkDeviceMemoryReportEventTypeEXT type, uint64_t memoryObjectId, VkDeviceSize size, VkObjectType objectType, uint64_t objectHandle, uint32_t heapIndex)
{
if(deviceMemoryReportCallbacks.empty()) return;
const VkDeviceMemoryReportCallbackDataEXT callbackData = {
VK_STRUCTURE_TYPE_DEVICE_MEMORY_REPORT_CALLBACK_DATA_EXT, // sType
nullptr, // pNext
0, // flags
type, // type
memoryObjectId, // memoryObjectId
size, // size
objectType, // objectType
objectHandle, // objectHandle
heapIndex, // heapIndex
};
for(const auto &callback : deviceMemoryReportCallbacks)
{
callback.first(&callbackData, callback.second);
}
}
#endif // SWIFTSHADER_DEVICE_MEMORY_REPORT
} // namespace vk