clang-format the src/Vulkan directory

Bug: b/144825072

Change-Id: I1bd5196b34a7974a41dcb95814a1ae8643b26f22
Reviewed-on: https://swiftshader-review.googlesource.com/c/SwiftShader/+/39658
Kokoro-Presubmit: kokoro <noreply+kokoro@google.com>
Tested-by: Ben Clayton <bclayton@google.com>
Reviewed-by: Nicolas Capens <nicolascapens@google.com>
diff --git a/src/Vulkan/VkPipelineCache.cpp b/src/Vulkan/VkPipelineCache.cpp
index 40fa5c3..9ee5723 100644
--- a/src/Vulkan/VkPipelineCache.cpp
+++ b/src/Vulkan/VkPipelineCache.cpp
@@ -17,12 +17,12 @@
 
 namespace vk {
 
-PipelineCache::SpirvShaderKey::SpecializationInfo::SpecializationInfo(const VkSpecializationInfo* specializationInfo)
+PipelineCache::SpirvShaderKey::SpecializationInfo::SpecializationInfo(const VkSpecializationInfo *specializationInfo)
 {
 	if(specializationInfo)
 	{
-		auto ptr = reinterpret_cast<VkSpecializationInfo*>(
-			allocate(sizeof(VkSpecializationInfo), REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY));
+		auto ptr = reinterpret_cast<VkSpecializationInfo *>(
+		    allocate(sizeof(VkSpecializationInfo), REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY));
 
 		info = std::shared_ptr<VkSpecializationInfo>(ptr, Deleter());
 
@@ -30,8 +30,8 @@
 		if(specializationInfo->mapEntryCount > 0)
 		{
 			size_t entriesSize = specializationInfo->mapEntryCount * sizeof(VkSpecializationMapEntry);
-			VkSpecializationMapEntry* mapEntries = reinterpret_cast<VkSpecializationMapEntry*>(
-				allocate(entriesSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY));
+			VkSpecializationMapEntry *mapEntries = reinterpret_cast<VkSpecializationMapEntry *>(
+			    allocate(entriesSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY));
 			memcpy(mapEntries, specializationInfo->pMapEntries, entriesSize);
 			info->pMapEntries = mapEntries;
 		}
@@ -39,7 +39,7 @@
 		info->dataSize = specializationInfo->dataSize;
 		if(specializationInfo->dataSize > 0)
 		{
-			void* data = allocate(specializationInfo->dataSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY);
+			void *data = allocate(specializationInfo->dataSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY);
 			memcpy(data, specializationInfo->pData, specializationInfo->dataSize);
 			info->pData = data;
 		}
@@ -50,17 +50,17 @@
 	}
 }
 
-void PipelineCache::SpirvShaderKey::SpecializationInfo::Deleter::operator() (VkSpecializationInfo* info) const
+void PipelineCache::SpirvShaderKey::SpecializationInfo::Deleter::operator()(VkSpecializationInfo *info) const
 {
 	if(info)
 	{
-		deallocate(const_cast<VkSpecializationMapEntry*>(info->pMapEntries), DEVICE_MEMORY);
-		deallocate(const_cast<void*>(info->pData), DEVICE_MEMORY);
+		deallocate(const_cast<VkSpecializationMapEntry *>(info->pMapEntries), DEVICE_MEMORY);
+		deallocate(const_cast<void *>(info->pData), DEVICE_MEMORY);
 		deallocate(info, DEVICE_MEMORY);
 	}
 }
 
-bool PipelineCache::SpirvShaderKey::SpecializationInfo::operator<(const SpecializationInfo& specializationInfo) const
+bool PipelineCache::SpirvShaderKey::SpecializationInfo::operator<(const SpecializationInfo &specializationInfo) const
 {
 	if(info && specializationInfo.info)
 	{
@@ -97,17 +97,17 @@
 }
 
 PipelineCache::SpirvShaderKey::SpirvShaderKey(const VkShaderStageFlagBits pipelineStage,
-	                                          const std::string& entryPointName,
-	                                          const std::vector<uint32_t>& insns,
-	                                          const vk::RenderPass *renderPass,
-	                                          const uint32_t subpassIndex,
-	                                          const VkSpecializationInfo* specializationInfo) :
-	pipelineStage(pipelineStage),
-	entryPointName(entryPointName),
-	insns(insns),
-	renderPass(renderPass),
-	subpassIndex(subpassIndex),
-	specializationInfo(specializationInfo)
+                                              const std::string &entryPointName,
+                                              const std::vector<uint32_t> &insns,
+                                              const vk::RenderPass *renderPass,
+                                              const uint32_t subpassIndex,
+                                              const VkSpecializationInfo *specializationInfo)
+    : pipelineStage(pipelineStage)
+    , entryPointName(entryPointName)
+    , insns(insns)
+    , renderPass(renderPass)
+    , subpassIndex(subpassIndex)
+    , specializationInfo(specializationInfo)
 {
 }
 
@@ -153,10 +153,11 @@
 	return (specializationInfo < other.specializationInfo);
 }
 
-PipelineCache::PipelineCache(const VkPipelineCacheCreateInfo* pCreateInfo, void* mem) :
-	dataSize(ComputeRequiredAllocationSize(pCreateInfo)), data(reinterpret_cast<uint8_t*>(mem))
+PipelineCache::PipelineCache(const VkPipelineCacheCreateInfo *pCreateInfo, void *mem)
+    : dataSize(ComputeRequiredAllocationSize(pCreateInfo))
+    , data(reinterpret_cast<uint8_t *>(mem))
 {
-	CacheHeader* header = reinterpret_cast<CacheHeader*>(mem);
+	CacheHeader *header = reinterpret_cast<CacheHeader *>(mem);
 	header->headerLength = sizeof(CacheHeader);
 	header->headerVersion = VK_PIPELINE_CACHE_HEADER_VERSION_ONE;
 	header->vendorID = VENDOR_ID;
@@ -175,17 +176,17 @@
 	computePrograms.clear();
 }
 
-void PipelineCache::destroy(const VkAllocationCallbacks* pAllocator)
+void PipelineCache::destroy(const VkAllocationCallbacks *pAllocator)
 {
 	vk::deallocate(data, pAllocator);
 }
 
-size_t PipelineCache::ComputeRequiredAllocationSize(const VkPipelineCacheCreateInfo* pCreateInfo)
+size_t PipelineCache::ComputeRequiredAllocationSize(const VkPipelineCacheCreateInfo *pCreateInfo)
 {
 	return pCreateInfo->initialDataSize + sizeof(CacheHeader);
 }
 
-VkResult PipelineCache::getData(size_t* pDataSize, void* pData)
+VkResult PipelineCache::getData(size_t *pDataSize, void *pData)
 {
 	if(!pData)
 	{
@@ -207,11 +208,11 @@
 	return VK_SUCCESS;
 }
 
-VkResult PipelineCache::merge(uint32_t srcCacheCount, const VkPipelineCache* pSrcCaches)
+VkResult PipelineCache::merge(uint32_t srcCacheCount, const VkPipelineCache *pSrcCaches)
 {
 	for(uint32_t i = 0; i < srcCacheCount; i++)
 	{
-		PipelineCache* srcCache = Cast(pSrcCaches[i]);
+		PipelineCache *srcCache = Cast(pSrcCaches[i]);
 
 		{
 			std::unique_lock<std::mutex> lock(spirvShadersMutex);
@@ -227,24 +228,24 @@
 	return VK_SUCCESS;
 }
 
-const std::shared_ptr<sw::SpirvShader>* PipelineCache::operator[](const PipelineCache::SpirvShaderKey& key) const
+const std::shared_ptr<sw::SpirvShader> *PipelineCache::operator[](const PipelineCache::SpirvShaderKey &key) const
 {
 	auto it = spirvShaders.find(key);
 	return (it != spirvShaders.end()) ? &(it->second) : nullptr;
 }
 
-void PipelineCache::insert(const PipelineCache::SpirvShaderKey& key, const std::shared_ptr<sw::SpirvShader> &shader)
+void PipelineCache::insert(const PipelineCache::SpirvShaderKey &key, const std::shared_ptr<sw::SpirvShader> &shader)
 {
 	spirvShaders[key] = shader;
 }
 
-const std::shared_ptr<sw::ComputeProgram>* PipelineCache::operator[](const PipelineCache::ComputeProgramKey& key) const
+const std::shared_ptr<sw::ComputeProgram> *PipelineCache::operator[](const PipelineCache::ComputeProgramKey &key) const
 {
 	auto it = computePrograms.find(key);
 	return (it != computePrograms.end()) ? &(it->second) : nullptr;
 }
 
-void PipelineCache::insert(const PipelineCache::ComputeProgramKey& key, const std::shared_ptr<sw::ComputeProgram> &computeProgram)
+void PipelineCache::insert(const PipelineCache::ComputeProgramKey &key, const std::shared_ptr<sw::ComputeProgram> &computeProgram)
 {
 	computePrograms[key] = computeProgram;
 }