blob: 17fd908e6958b508597505d6303c1383d05f035b [file] [log] [blame]
// Copyright 2018 The SwiftShader Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "VkDeviceMemory.hpp"
#include "VkBuffer.hpp"
#include "VkImage.hpp"
#include "Device/Blitter.hpp"
#include "Device/Surface.hpp"
#include <cstring>
namespace vk
{
Image::Image(const VkImageCreateInfo* pCreateInfo, void* mem) :
flags(pCreateInfo->flags),
imageType(pCreateInfo->imageType),
format(pCreateInfo->format),
extent(pCreateInfo->extent),
mipLevels(pCreateInfo->mipLevels),
arrayLayers(pCreateInfo->arrayLayers),
samples(pCreateInfo->samples),
tiling(pCreateInfo->tiling)
{
}
void Image::destroy(const VkAllocationCallbacks* pAllocator)
{
}
size_t Image::ComputeRequiredAllocationSize(const VkImageCreateInfo* pCreateInfo)
{
return 0;
}
const VkMemoryRequirements Image::getMemoryRequirements() const
{
VkMemoryRequirements memoryRequirements;
memoryRequirements.alignment = vk::REQUIRED_MEMORY_ALIGNMENT;
memoryRequirements.memoryTypeBits = vk::MEMORY_TYPE_GENERIC_BIT;
memoryRequirements.size = getStorageSize(flags);
return memoryRequirements;
}
void Image::bind(VkDeviceMemory pDeviceMemory, VkDeviceSize pMemoryOffset)
{
deviceMemory = Cast(pDeviceMemory);
memoryOffset = pMemoryOffset;
}
void Image::copyTo(VkImage dstImage, const VkImageCopy& pRegion)
{
// Image copy does not perform any conversion, it simply copies memory from
// an image to another image that has the same number of bytes per pixel.
Image* dst = Cast(dstImage);
int srcBytesPerTexel = bytesPerTexel(pRegion.srcSubresource.aspectMask);
ASSERT(srcBytesPerTexel == dst->bytesPerTexel(pRegion.dstSubresource.aspectMask));
if(!((pRegion.srcSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) ||
(pRegion.srcSubresource.aspectMask == VK_IMAGE_ASPECT_DEPTH_BIT) ||
(pRegion.srcSubresource.aspectMask == VK_IMAGE_ASPECT_STENCIL_BIT)) ||
(pRegion.srcSubresource.mipLevel != 0))
{
UNIMPLEMENTED();
}
if(!((pRegion.dstSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) ||
(pRegion.dstSubresource.aspectMask == VK_IMAGE_ASPECT_DEPTH_BIT) ||
(pRegion.dstSubresource.aspectMask == VK_IMAGE_ASPECT_STENCIL_BIT)) ||
(pRegion.dstSubresource.mipLevel != 0))
{
UNIMPLEMENTED();
}
const char* srcMem = static_cast<const char*>(
getTexelPointer(pRegion.srcOffset, pRegion.srcSubresource.baseArrayLayer, pRegion.srcSubresource.aspectMask));
char* dstMem = static_cast<char*>(
dst->getTexelPointer(pRegion.dstOffset, pRegion.dstSubresource.baseArrayLayer, pRegion.dstSubresource.aspectMask));
int srcRowPitchBytes = rowPitchBytes(pRegion.srcSubresource.aspectMask);
int srcSlicePitchBytes = slicePitchBytes(pRegion.srcSubresource.aspectMask);
int dstRowPitchBytes = dst->rowPitchBytes(pRegion.dstSubresource.aspectMask);
int dstSlicePitchBytes = dst->slicePitchBytes(pRegion.dstSubresource.aspectMask);
bool isSinglePlane = (pRegion.extent.depth == 1);
bool isSingleLine = (pRegion.extent.height == 1) && isSinglePlane;
bool isEntireLine = (pRegion.extent.width == extent.width) &&
(pRegion.extent.width == dst->extent.width) &&
(srcRowPitchBytes == dstRowPitchBytes);
bool isEntirePlane = isEntireLine &&
(pRegion.extent.height == extent.height) &&
(pRegion.extent.height == dst->extent.height) &&
(srcSlicePitchBytes == dstSlicePitchBytes);
if(isSingleLine)
{
memcpy(dstMem, srcMem, pRegion.extent.width * srcBytesPerTexel); // Copy one line
}
else if(isEntireLine && isSinglePlane)
{
memcpy(dstMem, srcMem, pRegion.extent.height * srcRowPitchBytes); // Copy one plane
}
else if(isEntirePlane)
{
memcpy(dstMem, srcMem, pRegion.extent.depth * srcSlicePitchBytes); // Copy multiple planes
}
else if(isEntireLine) // Copy plane by plane
{
for(uint32_t z = 0; z < pRegion.extent.depth; z++, dstMem += dstSlicePitchBytes, srcMem += srcSlicePitchBytes)
{
memcpy(dstMem, srcMem, pRegion.extent.height * srcRowPitchBytes);
}
}
else // Copy line by line
{
for(uint32_t z = 0; z < pRegion.extent.depth; z++)
{
for(uint32_t y = 0; y < pRegion.extent.height; y++, dstMem += dstRowPitchBytes, srcMem += srcRowPitchBytes)
{
memcpy(dstMem, srcMem, pRegion.extent.width * srcBytesPerTexel);
}
}
}
}
void Image::copy(VkBuffer buffer, const VkBufferImageCopy& region, bool bufferIsSource)
{
if((region.imageExtent.width != extent.width) ||
(region.imageExtent.height != extent.height) ||
(region.imageExtent.depth != extent.depth) ||
!((region.imageSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) ||
(region.imageSubresource.aspectMask == VK_IMAGE_ASPECT_DEPTH_BIT) ||
(region.imageSubresource.aspectMask == VK_IMAGE_ASPECT_STENCIL_BIT)) ||
(region.imageSubresource.mipLevel != 0) ||
(region.imageOffset.x != 0) ||
(region.imageOffset.y != 0) ||
(region.imageOffset.z != 0) ||
(region.bufferRowLength != extent.width) ||
(region.bufferImageHeight != extent.height))
{
UNIMPLEMENTED();
}
VkDeviceSize copySize = slicePitchBytes(region.imageSubresource.aspectMask) * region.imageExtent.depth;
VkDeviceSize layerSize = slicePitchBytes(region.imageSubresource.aspectMask) * extent.depth;
char* bufferMemory = static_cast<char*>(Cast(buffer)->getOffsetPointer(region.bufferOffset));
char* imageMemory = static_cast<char*>(deviceMemory->getOffsetPointer(getMemoryOffset(region.imageSubresource.aspectMask)));
char* srcMemory = bufferIsSource ? bufferMemory : imageMemory;
char* dstMemory = bufferIsSource ? imageMemory : bufferMemory;
uint32_t firstLayer = region.imageSubresource.baseArrayLayer;
uint32_t lastLayer = firstLayer + region.imageSubresource.layerCount - 1;
for(uint32_t layer = firstLayer; layer <= lastLayer; layer++)
{
memcpy(dstMemory, srcMemory, copySize);
srcMemory += layerSize;
dstMemory += layerSize;
}
}
void Image::copyTo(VkBuffer dstBuffer, const VkBufferImageCopy& region)
{
copy(dstBuffer, region, false);
}
void Image::copyFrom(VkBuffer srcBuffer, const VkBufferImageCopy& region)
{
copy(srcBuffer, region, true);
}
void* Image::getTexelPointer(const VkOffset3D& offset, uint32_t baseArrayLayer, const VkImageAspectFlags& flags) const
{
return deviceMemory->getOffsetPointer(texelOffsetBytesInStorage(offset, baseArrayLayer, flags) + getMemoryOffset(flags));
}
VkDeviceSize Image::texelOffsetBytesInStorage(const VkOffset3D& offset, uint32_t baseArrayLayer, const VkImageAspectFlags& flags) const
{
return (baseArrayLayer * extent.depth + offset.z) * slicePitchBytes(flags) + offset.y * rowPitchBytes(flags) + offset.x * bytesPerTexel(flags);
}
VkDeviceSize Image::getMemoryOffset(const VkImageAspectFlags& flags) const
{
switch(format)
{
case VK_FORMAT_D16_UNORM_S8_UINT:
case VK_FORMAT_D24_UNORM_S8_UINT:
case VK_FORMAT_D32_SFLOAT_S8_UINT:
if(flags == VK_IMAGE_ASPECT_STENCIL_BIT)
{
// Offset by depth buffer to get to stencil buffer
return memoryOffset + getStorageSize(VK_IMAGE_ASPECT_DEPTH_BIT);
}
break;
default:
break;
}
return memoryOffset;
}
int Image::rowPitchBytes(const VkImageAspectFlags& flags) const
{
// Depth and Stencil pitch should be computed separately
ASSERT((flags & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT)) !=
(VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT));
return sw::Surface::pitchB(extent.width, getBorder(), getFormat(flags), false);
}
int Image::slicePitchBytes(const VkImageAspectFlags& flags) const
{
// Depth and Stencil slice should be computed separately
ASSERT((flags & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT)) !=
(VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT));
return sw::Surface::sliceB(extent.width, extent.height, getBorder(), getFormat(flags), false);
}
int Image::bytesPerTexel(const VkImageAspectFlags& flags) const
{
// Depth and Stencil bytes should be computed separately
ASSERT((flags & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT)) !=
(VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT));
return sw::Surface::bytes(getFormat(flags));
}
VkFormat Image::getFormat(const VkImageAspectFlags& flags) const
{
switch(flags)
{
case VK_IMAGE_ASPECT_DEPTH_BIT:
switch(format)
{
case VK_FORMAT_D16_UNORM_S8_UINT:
return VK_FORMAT_D16_UNORM;
case VK_FORMAT_D24_UNORM_S8_UINT:
return VK_FORMAT_X8_D24_UNORM_PACK32; // FIXME: This will allocate an extra byte per pixel
case VK_FORMAT_D32_SFLOAT_S8_UINT:
return VK_FORMAT_D32_SFLOAT;
default:
break;
}
break;
case VK_IMAGE_ASPECT_STENCIL_BIT:
switch(format)
{
case VK_FORMAT_D16_UNORM_S8_UINT:
case VK_FORMAT_D24_UNORM_S8_UINT:
case VK_FORMAT_D32_SFLOAT_S8_UINT:
return VK_FORMAT_S8_UINT;
default:
break;
}
break;
default:
break;
}
return format;
}
int Image::getBorder() const
{
return ((flags & VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT) && (imageType == VK_IMAGE_TYPE_2D)) ? 1 : 0;
}
VkDeviceSize Image::getStorageSize(const VkImageAspectFlags& flags) const
{
if(mipLevels > 1)
{
UNIMPLEMENTED();
}
int slicePitchB = 0;
if(sw::Surface::isDepth(format) && sw::Surface::isStencil(format))
{
switch(flags)
{
case VK_IMAGE_ASPECT_DEPTH_BIT:
case VK_IMAGE_ASPECT_STENCIL_BIT:
slicePitchB = slicePitchBytes(flags);
break;
default:
// Allow allocating both depth and stencil contiguously
slicePitchB = (slicePitchBytes(VK_IMAGE_ASPECT_DEPTH_BIT) + slicePitchBytes(VK_IMAGE_ASPECT_STENCIL_BIT));
break;
}
}
else
{
slicePitchB = slicePitchBytes(flags);
}
return arrayLayers * extent.depth * slicePitchB;
}
void Image::clear(const VkClearValue& clearValue, const VkRect2D& renderArea, const VkImageSubresourceRange& subresourceRange)
{
if(!((subresourceRange.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) ||
(subresourceRange.aspectMask == VK_IMAGE_ASPECT_DEPTH_BIT) ||
(subresourceRange.aspectMask == VK_IMAGE_ASPECT_STENCIL_BIT)) ||
(subresourceRange.baseMipLevel != 0) ||
(subresourceRange.levelCount != 1) ||
(subresourceRange.baseArrayLayer != 0) ||
(subresourceRange.layerCount != 1))
{
UNIMPLEMENTED();
}
// Set the proper format for the clear value, as described here:
// https://www.khronos.org/registry/vulkan/specs/1.1-extensions/html/vkspec.html#clears-values
VkFormat clearFormat = VK_FORMAT_R32G32B32A32_SFLOAT;
if(sw::Surface::isSignedNonNormalizedInteger(format))
{
clearFormat = VK_FORMAT_R32G32B32A32_SINT;
}
else if(sw::Surface::isUnsignedNonNormalizedInteger(format))
{
clearFormat = VK_FORMAT_R32G32B32A32_UINT;
}
const sw::Rect rect(renderArea.offset.x, renderArea.offset.y,
renderArea.offset.x + renderArea.extent.width,
renderArea.offset.y + renderArea.extent.height);
const sw::SliceRect dRect(rect);
sw::Surface* surface = sw::Surface::create(extent.width, extent.height, extent.depth,
getFormat(subresourceRange.aspectMask), deviceMemory->getOffsetPointer(memoryOffset),
rowPitchBytes(subresourceRange.aspectMask), slicePitchBytes(subresourceRange.aspectMask));
sw::Blitter blitter;
blitter.clear((void*)clearValue.color.float32, clearFormat, surface, dRect, 0xF);
delete surface;
}
} // namespace vk