From 738e2879b3cc64a7200f64e7a838704db82550b4 Mon Sep 17 00:00:00 2001 From: Mikko Rasa Date: Tue, 25 Jan 2022 19:42:35 +0200 Subject: [PATCH] Overhaul the Vulkan backend's memory allocator It now sub-allocates from larger chunks of memory to avoid hitting Vulkan's allocation count limit. --- source/backends/vulkan/memoryallocator.cpp | 276 ++++++++++++++++++--- source/backends/vulkan/memoryallocator.h | 45 +++- 2 files changed, 273 insertions(+), 48 deletions(-) diff --git a/source/backends/vulkan/memoryallocator.cpp b/source/backends/vulkan/memoryallocator.cpp index b15344c7..227b54b6 100644 --- a/source/backends/vulkan/memoryallocator.cpp +++ b/source/backends/vulkan/memoryallocator.cpp @@ -20,52 +20,220 @@ MemoryAllocator::MemoryAllocator(Device &d): VkPhysicalDeviceMemoryProperties mem_props; vk.GetPhysicalDeviceMemoryProperties(mem_props); + for(unsigned i=0; i::iterator MemoryAllocator::lower_bound_by_size(vector &indices, size_t size) +{ + return lower_bound(indices, size, [this](unsigned j, unsigned s){ return blocks[j].size=direct_alloc_threshold) + { + Block block; + block.region = create_region(pool_index, size, true); + block.size = size; + block.allocated = true; + + blocks.push_back(block); + return blocks.size()-1; + } - Allocation alloc; - vk.AllocateMemory(alloc_info, alloc.memory); + if(pool.can_consolidate && blocks[pool.free_blocks.back()].size=size+min_alignment) + { + unsigned tail_index = split_block(block_index, size); + pool.free_blocks.insert(lower_bound_by_size(pool.free_blocks, blocks[tail_index].size), tail_index); + } + + blocks[block_index].allocated = true; - return allocations.size()-1; + return block_index; +} + +unsigned MemoryAllocator::split_block(unsigned index, size_t head_size) +{ + blocks.emplace_back(); + Block &block = blocks[index]; + Block &tail = blocks.back(); + unsigned tail_index = blocks.size()-1; + + tail.region = block.region; + tail.offset = block.offset+head_size; + tail.size = block.size-head_size; + tail.prev = index; + tail.next = block.next; + + block.size = head_size; + block.next = tail_index; + + return tail_index; +} + +void MemoryAllocator::consolidate(unsigned pool_index) +{ + Pool &pool = pools[pool_index]; + + vector merged_blocks; + unsigned i = 0; + for(unsigned j=0; j=0 && !blocks[block.next].allocated) + { + merge_block_with_next(block_index); + + while(block.next>=0 && !blocks[block.next].allocated) + merge_block_with_next(block_index); + + merged_blocks.insert(lower_bound_by_size(merged_blocks, block.size), block_index); + } + } + else + continue; + } + + if(j!=i) + pool.free_blocks[i] = block_index; + ++i; + } + + pool.free_blocks.resize(i+merged_blocks.size()); + + if(!merged_blocks.empty()) + { + unsigned j = merged_blocks.size(); + for(unsigned k=pool.free_blocks.size()-1; j; --k) + { + if(!i || blocks[merged_blocks[j-1]].size>blocks[pool.free_blocks[i-1]].size) + pool.free_blocks[k] = merged_blocks[--j]; + else + pool.free_blocks[k] = pool.free_blocks[--i]; + } + } +} + +void MemoryAllocator::merge_block_with_next(unsigned index) +{ + Block &block = blocks[index]; + + Block &next = blocks[block.next]; + block.size += next.size; + block.next = next.next; + if(block.next>=0) + blocks[block.next].prev = index; + + next = Block(); } unsigned MemoryAllocator::allocate(VkBuffer buffer, MemoryType type) @@ -75,11 +243,12 @@ unsigned MemoryAllocator::allocate(VkBuffer buffer, MemoryType type) VkMemoryRequirements requirements; vk.GetBufferMemoryRequirements(buffer, requirements); - unsigned index = allocate(requirements.size, requirements.memoryTypeBits, type); + unsigned block_index = allocate(requirements.size, requirements.alignment, requirements.memoryTypeBits, type); - vk.BindBufferMemory(buffer, allocations[index].memory, 0); + Block &block = blocks[block_index]; + vk.BindBufferMemory(buffer, regions[block.region].memory, block.offset); - return index+1; + return block_index+1; } unsigned MemoryAllocator::allocate(VkImage image, MemoryType type) @@ -89,50 +258,75 @@ unsigned MemoryAllocator::allocate(VkImage image, MemoryType type) VkMemoryRequirements requirements; vk.GetImageMemoryRequirements(image, requirements); - unsigned index = allocate(requirements.size, requirements.memoryTypeBits, type); + unsigned block_index = allocate(requirements.size, requirements.alignment, requirements.memoryTypeBits, type); - vk.BindImageMemory(image, allocations[index].memory, 0); + Block &block = blocks[block_index]; + vk.BindImageMemory(image, regions[block.region].memory, block.offset); - return index+1; + return block_index+1; } void MemoryAllocator::release(unsigned id) { - if(!id || id>allocations.size() || !allocations[id-1].memory) + if(!id || id>blocks.size() || !blocks[id-1].allocated) throw key_error(id); - const VulkanFunctions &vk = device.get_functions(); + unsigned block_index = id-1; + Block &block = blocks[block_index]; + + block.allocated = false; - vk.FreeMemory(allocations[id-1].memory); + Region ®ion = regions[block.region]; + if(region.direct) + { + const VulkanFunctions &vk = device.get_functions(); + + vk.FreeMemory(region.memory); + region = Region(); + block = Block(); + return; + } + + Pool &pool = pools[region.pool]; + pool.free_blocks.insert(lower_bound_by_size(pool.free_blocks, block.size), block_index); + if((block.prev>=0 && !blocks[block.prev].allocated) || (block.next>=0 && !blocks[block.next].allocated)) + pool.can_consolidate = true; } void *MemoryAllocator::map(unsigned id) { - if(!id || id>allocations.size() || !allocations[id-1].memory) + if(!id || id>blocks.size() || !blocks[id-1].allocated) throw key_error(id); - Allocation &alloc = allocations[id-1]; - - const VulkanFunctions &vk = device.get_functions(); + Block &block = blocks[id-1]; + Region ®ion = regions[block.region]; + if(!region.mapped_address) + { + const VulkanFunctions &vk = device.get_functions(); + vk.MapMemory(region.memory, 0, region.size, 0, ®ion.mapped_address); + } - vk.MapMemory(alloc.memory, 0, alloc.size, 0, &alloc.mapped_address); + ++region.map_count; - return alloc.mapped_address; + return static_cast(region.mapped_address)+block.offset; } void MemoryAllocator::unmap(unsigned id) { - if(!id || id>allocations.size() || !allocations[id-1].memory) + if(!id || id>blocks.size() || !blocks[id-1].allocated) throw key_error(id); - Allocation &alloc = allocations[id-1]; - if(!alloc.mapped_address) - throw invalid_operation("MemoryAllocator::unmap"); + Block &block = blocks[id-1]; + Region ®ion = regions[block.region]; - const VulkanFunctions &vk = device.get_functions(); - - vk.UnmapMemory(alloc.memory); - alloc.mapped_address = 0; + if(!regions[block.region].mapped_address) + throw invalid_operation("MemoryAllocator::unmap"); + else if(!--region.map_count) + { + const VulkanFunctions &vk = device.get_functions(); + vk.UnmapMemory(region.memory); + region.mapped_address = 0; + } } } // namespace GL diff --git a/source/backends/vulkan/memoryallocator.h b/source/backends/vulkan/memoryallocator.h index 230f21a2..272f54e0 100644 --- a/source/backends/vulkan/memoryallocator.h +++ b/source/backends/vulkan/memoryallocator.h @@ -2,6 +2,7 @@ #define MSP_GL_VULKAN_MEMORYALLOCATOR_H_ #include +#include #include #include "handles.h" @@ -18,28 +19,58 @@ enum MemoryType STREAMING_MEMORY }; -class MemoryAllocator +class MemoryAllocator: public NonCopyable { private: - struct Allocation + struct Pool { - VkDeviceMemory memory = 0; MemoryType type = UNKNOWN_MEMORY; + std::vector free_blocks; + bool can_consolidate = false; + }; + + struct Region + { + int pool = -1; + bool direct = false; + VkDeviceMemory memory = 0; std::size_t size = 0; void *mapped_address = 0; + unsigned map_count = 0; + }; + + struct Block + { + int region = -1; + bool allocated = false; + std::size_t offset = 0; + std::size_t size = 0; + int prev = -1; + int next = -1; }; Device &device; VkPhysicalDevice phys_device; - std::vector memory_types; - std::vector allocations; + std::size_t total_device_memory = 0; + std::size_t default_region_size = 0; + std::size_t direct_alloc_threshold = 0; + std::size_t min_alignment = 256; + std::vector pools; + std::vector regions; + std::vector blocks; public: MemoryAllocator(Device &); + ~MemoryAllocator(); private: - unsigned find_memory_type_index(unsigned, MemoryType); - unsigned allocate(std::size_t, unsigned, MemoryType); + unsigned find_memory_pool(unsigned, MemoryType); + unsigned create_region(unsigned, size_t, bool); + std::vector::iterator lower_bound_by_size(std::vector &, std::size_t); + unsigned allocate(std::size_t, std::size_t, unsigned, MemoryType); + unsigned split_block(unsigned, std::size_t); + void consolidate(unsigned); + void merge_block_with_next(unsigned); public: unsigned allocate(VkBuffer, MemoryType); -- 2.45.2