mirror of
https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git
synced 2024-11-26 00:24:35 +00:00
Code refactor (lower scope, local var, default constructor and destructor)
This commit is contained in:
parent
895b080a3c
commit
666b38f11f
@ -3801,11 +3801,9 @@ VmaRawList<T>::VmaRawList(const VkAllocationCallbacks* pAllocationCallbacks) :
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
VmaRawList<T>::~VmaRawList()
|
||||
{
|
||||
// Intentionally not calling Clear, because that would be unnecessary
|
||||
// computations to return all items to m_ItemAllocator as free.
|
||||
}
|
||||
VmaRawList<T>::~VmaRawList() = default;
|
||||
// Intentionally not calling Clear, because that would be unnecessary
|
||||
// computations to return all items to m_ItemAllocator as free.
|
||||
|
||||
template<typename T>
|
||||
void VmaRawList<T>::Clear()
|
||||
@ -4225,7 +4223,7 @@ public:
|
||||
static ItemType* GetPrev(const ItemType* item) { return ItemTypeTraits::GetPrev(item); }
|
||||
static ItemType* GetNext(const ItemType* item) { return ItemTypeTraits::GetNext(item); }
|
||||
// Movable, not copyable.
|
||||
VmaIntrusiveLinkedList() { }
|
||||
VmaIntrusiveLinkedList() = default;
|
||||
VmaIntrusiveLinkedList(const VmaIntrusiveLinkedList<ItemTypeTraits>& src) = delete;
|
||||
VmaIntrusiveLinkedList(VmaIntrusiveLinkedList<ItemTypeTraits>&& src) :
|
||||
m_Front(src.m_Front), m_Back(src.m_Back), m_Count(src.m_Count)
|
||||
@ -12226,13 +12224,14 @@ void VmaBlockVector::Defragment(
|
||||
|
||||
const VkDeviceSize maxBytesToMove = defragmentOnGpu ? maxGpuBytesToMove : maxCpuBytesToMove;
|
||||
const uint32_t maxAllocationsToMove = defragmentOnGpu ? maxGpuAllocationsToMove : maxCpuAllocationsToMove;
|
||||
pCtx->res = pCtx->GetAlgorithm()->Defragment(pCtx->defragmentationMoves, maxBytesToMove, maxAllocationsToMove, flags);
|
||||
VmaDefragmentationAlgorithm* algo = pCtx->GetAlgorithm();
|
||||
pCtx->res = algo->Defragment(pCtx->defragmentationMoves, maxBytesToMove, maxAllocationsToMove, flags);
|
||||
|
||||
// Accumulate statistics.
|
||||
if(pStats != VMA_NULL)
|
||||
{
|
||||
const VkDeviceSize bytesMoved = pCtx->GetAlgorithm()->GetBytesMoved();
|
||||
const uint32_t allocationsMoved = pCtx->GetAlgorithm()->GetAllocationsMoved();
|
||||
const VkDeviceSize bytesMoved = algo->GetBytesMoved();
|
||||
const uint32_t allocationsMoved = algo->GetAllocationsMoved();
|
||||
pStats->bytesMoved += bytesMoved;
|
||||
pStats->allocationsMoved += allocationsMoved;
|
||||
VMA_ASSERT(bytesMoved <= maxBytesToMove);
|
||||
@ -13073,9 +13072,9 @@ void VmaDefragmentationAlgorithm_Fast::PostprocessMetadata()
|
||||
freeSize, // size
|
||||
VMA_NULL, // hAllocation
|
||||
VMA_SUBALLOCATION_TYPE_FREE };
|
||||
VmaSuballocationList::iterator precedingFreeIt = pMetadata->m_Suballocations.insert(it, suballoc);
|
||||
if(freeSize >= VMA_MIN_FREE_SUBALLOCATION_SIZE_TO_REGISTER)
|
||||
{
|
||||
VmaSuballocationList::iterator precedingFreeIt = pMetadata->m_Suballocations.insert(it, suballoc);
|
||||
pMetadata->m_FreeSuballocationsBySize.push_back(precedingFreeIt);
|
||||
}
|
||||
}
|
||||
@ -13095,9 +13094,9 @@ void VmaDefragmentationAlgorithm_Fast::PostprocessMetadata()
|
||||
VMA_NULL, // hAllocation
|
||||
VMA_SUBALLOCATION_TYPE_FREE };
|
||||
VMA_ASSERT(it == pMetadata->m_Suballocations.end());
|
||||
VmaSuballocationList::iterator trailingFreeIt = pMetadata->m_Suballocations.insert(it, suballoc);
|
||||
if(freeSize > VMA_MIN_FREE_SUBALLOCATION_SIZE_TO_REGISTER)
|
||||
{
|
||||
VmaSuballocationList::iterator trailingFreeIt = pMetadata->m_Suballocations.insert(it, suballoc);
|
||||
pMetadata->m_FreeSuballocationsBySize.push_back(trailingFreeIt);
|
||||
}
|
||||
}
|
||||
@ -16634,12 +16633,13 @@ VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForBufferInfo(
|
||||
|
||||
const VkDevice hDev = allocator->m_hDevice;
|
||||
VkBuffer hBuffer = VK_NULL_HANDLE;
|
||||
VkResult res = allocator->GetVulkanFunctions().vkCreateBuffer(
|
||||
const VmaVulkanFunctions* funcs = &allocator->GetVulkanFunctions();
|
||||
VkResult res = funcs->vkCreateBuffer(
|
||||
hDev, pBufferCreateInfo, allocator->GetAllocationCallbacks(), &hBuffer);
|
||||
if(res == VK_SUCCESS)
|
||||
{
|
||||
VkMemoryRequirements memReq = {};
|
||||
allocator->GetVulkanFunctions().vkGetBufferMemoryRequirements(
|
||||
funcs->vkGetBufferMemoryRequirements(
|
||||
hDev, hBuffer, &memReq);
|
||||
|
||||
res = vmaFindMemoryTypeIndex(
|
||||
@ -16648,7 +16648,7 @@ VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForBufferInfo(
|
||||
pAllocationCreateInfo,
|
||||
pMemoryTypeIndex);
|
||||
|
||||
allocator->GetVulkanFunctions().vkDestroyBuffer(
|
||||
funcs->vkDestroyBuffer(
|
||||
hDev, hBuffer, allocator->GetAllocationCallbacks());
|
||||
}
|
||||
return res;
|
||||
@ -16667,12 +16667,13 @@ VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForImageInfo(
|
||||
|
||||
const VkDevice hDev = allocator->m_hDevice;
|
||||
VkImage hImage = VK_NULL_HANDLE;
|
||||
VkResult res = allocator->GetVulkanFunctions().vkCreateImage(
|
||||
const VmaVulkanFunctions* funcs = &allocator->GetVulkanFunctions();
|
||||
VkResult res = funcs->vkCreateImage(
|
||||
hDev, pImageCreateInfo, allocator->GetAllocationCallbacks(), &hImage);
|
||||
if(res == VK_SUCCESS)
|
||||
{
|
||||
VkMemoryRequirements memReq = {};
|
||||
allocator->GetVulkanFunctions().vkGetImageMemoryRequirements(
|
||||
funcs->vkGetImageMemoryRequirements(
|
||||
hDev, hImage, &memReq);
|
||||
|
||||
res = vmaFindMemoryTypeIndex(
|
||||
@ -16681,7 +16682,7 @@ VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForImageInfo(
|
||||
pAllocationCreateInfo,
|
||||
pMemoryTypeIndex);
|
||||
|
||||
allocator->GetVulkanFunctions().vkDestroyImage(
|
||||
funcs->vkDestroyImage(
|
||||
hDev, hImage, allocator->GetAllocationCallbacks());
|
||||
}
|
||||
return res;
|
||||
|
Loading…
Reference in New Issue
Block a user