Update VulkanMemoryAllocator/Hpp

This commit is contained in:
BearOso 2023-05-01 12:24:38 -05:00
parent 54aae6f5cf
commit b2d92e86ac
3 changed files with 168 additions and 62 deletions

View File

@ -2614,6 +2614,10 @@ VMA_CALL_PRE void VMA_CALL_POST vmaFreeStatsString(
#include <bit> // For std::popcount #include <bit> // For std::popcount
#endif #endif
#if VMA_STATS_STRING_ENABLED
#include <cstdio> // For snprintf
#endif
/******************************************************************************* /*******************************************************************************
CONFIGURATION SECTION CONFIGURATION SECTION
@ -2647,8 +2651,7 @@ VmaAllocatorCreateInfo::pVulkanFunctions. Other members can be null.
#endif #endif
#ifndef VMA_USE_STL_SHARED_MUTEX #ifndef VMA_USE_STL_SHARED_MUTEX
// Compiler conforms to C++17. #if __cplusplus >= 201703L || _MSVC_LANG >= 201703L // C++17
#if __cplusplus >= 201703L
#define VMA_USE_STL_SHARED_MUTEX 1 #define VMA_USE_STL_SHARED_MUTEX 1
// Visual studio defines __cplusplus properly only when passed additional parameter: /Zc:__cplusplus // Visual studio defines __cplusplus properly only when passed additional parameter: /Zc:__cplusplus
// Otherwise it is always 199711L, despite shared_mutex works since Visual Studio 2015 Update 2. // Otherwise it is always 199711L, despite shared_mutex works since Visual Studio 2015 Update 2.
@ -2693,6 +2696,14 @@ remove them if not needed.
#define VMA_NULL nullptr #define VMA_NULL nullptr
#endif #endif
#ifndef VMA_FALLTHROUGH
#if __cplusplus >= 201703L || _MSVC_LANG >= 201703L // C++17
#define VMA_FALLTHROUGH [[fallthrough]]
#else
#define VMA_FALLTHROUGH
#endif
#endif
// Normal assert to check for programmer's errors, especially in Debug configuration. // Normal assert to check for programmer's errors, especially in Debug configuration.
#ifndef VMA_ASSERT #ifndef VMA_ASSERT
#ifdef NDEBUG #ifdef NDEBUG
@ -2769,7 +2780,7 @@ static void* vma_aligned_alloc(size_t alignment, size_t size)
{ {
return _aligned_malloc(size, alignment); return _aligned_malloc(size, alignment);
} }
#elif __cplusplus >= 201703L // Compiler conforms to C++17. #elif __cplusplus >= 201703L || _MSVC_LANG >= 201703L // C++17
static void* vma_aligned_alloc(size_t alignment, size_t size) static void* vma_aligned_alloc(size_t alignment, size_t size)
{ {
return aligned_alloc(alignment, size); return aligned_alloc(alignment, size);
@ -2856,6 +2867,21 @@ static void vma_aligned_free(void* VMA_NULLABLE ptr)
#define VMA_DEBUG_LOG(str) VMA_DEBUG_LOG_FORMAT("%s", (str)) #define VMA_DEBUG_LOG(str) VMA_DEBUG_LOG_FORMAT("%s", (str))
#endif #endif
#ifndef VMA_CLASS_NO_COPY
#define VMA_CLASS_NO_COPY(className) \
private: \
className(const className&) = delete; \
className& operator=(const className&) = delete;
#endif
#ifndef VMA_CLASS_NO_COPY_NO_MOVE
#define VMA_CLASS_NO_COPY_NO_MOVE(className) \
private: \
className(const className&) = delete; \
className(className&&) = delete; \
className& operator=(const className&) = delete; \
className& operator=(className&&) = delete;
#endif
// Define this macro to 1 to enable functions: vmaBuildStatsString, vmaFreeStatsString. // Define this macro to 1 to enable functions: vmaBuildStatsString, vmaFreeStatsString.
#if VMA_STATS_STRING_ENABLED #if VMA_STATS_STRING_ENABLED
static inline void VmaUint32ToStr(char* VMA_NOT_NULL outStr, size_t strLen, uint32_t num) static inline void VmaUint32ToStr(char* VMA_NOT_NULL outStr, size_t strLen, uint32_t num)
@ -2875,7 +2901,9 @@ static void vma_aligned_free(void* VMA_NULLABLE ptr)
#ifndef VMA_MUTEX #ifndef VMA_MUTEX
class VmaMutex class VmaMutex
{ {
VMA_CLASS_NO_COPY_NO_MOVE(VmaMutex)
public: public:
VmaMutex() { }
void Lock() { m_Mutex.lock(); } void Lock() { m_Mutex.lock(); }
void Unlock() { m_Mutex.unlock(); } void Unlock() { m_Mutex.unlock(); }
bool TryLock() { return m_Mutex.try_lock(); } bool TryLock() { return m_Mutex.try_lock(); }
@ -3041,13 +3069,6 @@ tools like RenderDoc.
#define VMA_MAPPING_HYSTERESIS_ENABLED 1 #define VMA_MAPPING_HYSTERESIS_ENABLED 1
#endif #endif
#ifndef VMA_CLASS_NO_COPY
#define VMA_CLASS_NO_COPY(className) \
private: \
className(const className&) = delete; \
className& operator=(const className&) = delete;
#endif
#define VMA_VALIDATE(cond) do { if(!(cond)) { \ #define VMA_VALIDATE(cond) do { if(!(cond)) { \
VMA_ASSERT(0 && "Validation failed: " #cond); \ VMA_ASSERT(0 && "Validation failed: " #cond); \
return false; \ return false; \
@ -3540,7 +3561,7 @@ new element with value (key) should be inserted.
template <typename CmpLess, typename IterT, typename KeyT> template <typename CmpLess, typename IterT, typename KeyT>
static IterT VmaBinaryFindFirstNotLess(IterT beg, IterT end, const KeyT& key, const CmpLess& cmp) static IterT VmaBinaryFindFirstNotLess(IterT beg, IterT end, const KeyT& key, const CmpLess& cmp)
{ {
size_t down = 0, up = (end - beg); size_t down = 0, up = size_t(end - beg);
while (down < up) while (down < up)
{ {
const size_t mid = down + (up - down) / 2; // Overflow-safe midpoint calculation const size_t mid = down + (up - down) / 2; // Overflow-safe midpoint calculation
@ -3947,7 +3968,7 @@ static void VmaAddDetailedStatistics(VmaDetailedStatistics& inoutStats, const Vm
// Helper RAII class to lock a mutex in constructor and unlock it in destructor (at the end of scope). // Helper RAII class to lock a mutex in constructor and unlock it in destructor (at the end of scope).
struct VmaMutexLock struct VmaMutexLock
{ {
VMA_CLASS_NO_COPY(VmaMutexLock) VMA_CLASS_NO_COPY_NO_MOVE(VmaMutexLock)
public: public:
VmaMutexLock(VMA_MUTEX& mutex, bool useMutex = true) : VmaMutexLock(VMA_MUTEX& mutex, bool useMutex = true) :
m_pMutex(useMutex ? &mutex : VMA_NULL) m_pMutex(useMutex ? &mutex : VMA_NULL)
@ -3963,7 +3984,7 @@ private:
// Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for reading. // Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for reading.
struct VmaMutexLockRead struct VmaMutexLockRead
{ {
VMA_CLASS_NO_COPY(VmaMutexLockRead) VMA_CLASS_NO_COPY_NO_MOVE(VmaMutexLockRead)
public: public:
VmaMutexLockRead(VMA_RW_MUTEX& mutex, bool useMutex) : VmaMutexLockRead(VMA_RW_MUTEX& mutex, bool useMutex) :
m_pMutex(useMutex ? &mutex : VMA_NULL) m_pMutex(useMutex ? &mutex : VMA_NULL)
@ -3979,7 +4000,7 @@ private:
// Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for writing. // Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for writing.
struct VmaMutexLockWrite struct VmaMutexLockWrite
{ {
VMA_CLASS_NO_COPY(VmaMutexLockWrite) VMA_CLASS_NO_COPY_NO_MOVE(VmaMutexLockWrite)
public: public:
VmaMutexLockWrite(VMA_RW_MUTEX& mutex, bool useMutex) VmaMutexLockWrite(VMA_RW_MUTEX& mutex, bool useMutex)
: m_pMutex(useMutex ? &mutex : VMA_NULL) : m_pMutex(useMutex ? &mutex : VMA_NULL)
@ -4002,11 +4023,11 @@ private:
#ifndef _VMA_ATOMIC_TRANSACTIONAL_INCREMENT #ifndef _VMA_ATOMIC_TRANSACTIONAL_INCREMENT
// An object that increments given atomic but decrements it back in the destructor unless Commit() is called. // An object that increments given atomic but decrements it back in the destructor unless Commit() is called.
template<typename T> template<typename AtomicT>
struct AtomicTransactionalIncrement struct AtomicTransactionalIncrement
{ {
public: public:
typedef std::atomic<T> AtomicT; using T = decltype(AtomicT().load());
~AtomicTransactionalIncrement() ~AtomicTransactionalIncrement()
{ {
@ -4434,7 +4455,7 @@ allocator can create multiple blocks.
template<typename T> template<typename T>
class VmaPoolAllocator class VmaPoolAllocator
{ {
VMA_CLASS_NO_COPY(VmaPoolAllocator) VMA_CLASS_NO_COPY_NO_MOVE(VmaPoolAllocator)
public: public:
VmaPoolAllocator(const VkAllocationCallbacks* pAllocationCallbacks, uint32_t firstBlockCapacity); VmaPoolAllocator(const VkAllocationCallbacks* pAllocationCallbacks, uint32_t firstBlockCapacity);
~VmaPoolAllocator(); ~VmaPoolAllocator();
@ -4567,7 +4588,7 @@ struct VmaListItem
template<typename T> template<typename T>
class VmaRawList class VmaRawList
{ {
VMA_CLASS_NO_COPY(VmaRawList) VMA_CLASS_NO_COPY_NO_MOVE(VmaRawList)
public: public:
typedef VmaListItem<T> ItemType; typedef VmaListItem<T> ItemType;
@ -4830,7 +4851,7 @@ VmaListItem<T>* VmaRawList<T>::InsertAfter(ItemType* pItem, const T& value)
template<typename T, typename AllocatorT> template<typename T, typename AllocatorT>
class VmaList class VmaList
{ {
VMA_CLASS_NO_COPY(VmaList) VMA_CLASS_NO_COPY_NO_MOVE(VmaList)
public: public:
class reverse_iterator; class reverse_iterator;
class const_iterator; class const_iterator;
@ -5422,7 +5443,7 @@ void VmaStringBuilder::AddNumber(uint32_t num)
char* p = &buf[10]; char* p = &buf[10];
do do
{ {
*--p = '0' + (num % 10); *--p = '0' + (char)(num % 10);
num /= 10; num /= 10;
} while (num); } while (num);
Add(p); Add(p);
@ -5435,7 +5456,7 @@ void VmaStringBuilder::AddNumber(uint64_t num)
char* p = &buf[20]; char* p = &buf[20];
do do
{ {
*--p = '0' + (num % 10); *--p = '0' + (char)(num % 10);
num /= 10; num /= 10;
} while (num); } while (num);
Add(p); Add(p);
@ -5457,7 +5478,7 @@ VmaStringBuilder passed to the constructor.
*/ */
class VmaJsonWriter class VmaJsonWriter
{ {
VMA_CLASS_NO_COPY(VmaJsonWriter) VMA_CLASS_NO_COPY_NO_MOVE(VmaJsonWriter)
public: public:
// sb - string builder to write the document to. Must remain alive for the whole lifetime of this object. // sb - string builder to write the document to. Must remain alive for the whole lifetime of this object.
VmaJsonWriter(const VkAllocationCallbacks* pAllocationCallbacks, VmaStringBuilder& sb); VmaJsonWriter(const VkAllocationCallbacks* pAllocationCallbacks, VmaStringBuilder& sb);
@ -5649,7 +5670,6 @@ void VmaJsonWriter::ContinueString(const char* pStr)
break; break;
default: default:
VMA_ASSERT(0 && "Character not currently supported."); VMA_ASSERT(0 && "Character not currently supported.");
break;
} }
} }
} }
@ -5796,7 +5816,7 @@ static void VmaPrintDetailedStatistics(VmaJsonWriter& json, const VmaDetailedSta
class VmaMappingHysteresis class VmaMappingHysteresis
{ {
VMA_CLASS_NO_COPY(VmaMappingHysteresis) VMA_CLASS_NO_COPY_NO_MOVE(VmaMappingHysteresis)
public: public:
VmaMappingHysteresis() = default; VmaMappingHysteresis() = default;
@ -5903,7 +5923,7 @@ Thread-safety:
*/ */
class VmaDeviceMemoryBlock class VmaDeviceMemoryBlock
{ {
VMA_CLASS_NO_COPY(VmaDeviceMemoryBlock) VMA_CLASS_NO_COPY_NO_MOVE(VmaDeviceMemoryBlock)
public: public:
VmaBlockMetadata* m_pMetadata; VmaBlockMetadata* m_pMetadata;
@ -6126,6 +6146,7 @@ Thread-safe, synchronized internally.
*/ */
class VmaDedicatedAllocationList class VmaDedicatedAllocationList
{ {
VMA_CLASS_NO_COPY_NO_MOVE(VmaDedicatedAllocationList)
public: public:
VmaDedicatedAllocationList() {} VmaDedicatedAllocationList() {}
~VmaDedicatedAllocationList(); ~VmaDedicatedAllocationList();
@ -6311,6 +6332,7 @@ in a single VkDeviceMemory block.
*/ */
class VmaBlockMetadata class VmaBlockMetadata
{ {
VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockMetadata)
public: public:
// pAllocationCallbacks, if not null, must be owned externally - alive and unchanged for the whole lifetime of this object. // pAllocationCallbacks, if not null, must be owned externally - alive and unchanged for the whole lifetime of this object.
VmaBlockMetadata(const VkAllocationCallbacks* pAllocationCallbacks, VmaBlockMetadata(const VkAllocationCallbacks* pAllocationCallbacks,
@ -6377,7 +6399,7 @@ public:
protected: protected:
const VkAllocationCallbacks* GetAllocationCallbacks() const { return m_pAllocationCallbacks; } const VkAllocationCallbacks* GetAllocationCallbacks() const { return m_pAllocationCallbacks; }
VkDeviceSize GetBufferImageGranularity() const { return m_BufferImageGranularity; } VkDeviceSize GetBufferImageGranularity() const { return m_BufferImageGranularity; }
VkDeviceSize GetDebugMargin() const { return IsVirtual() ? 0 : VMA_DEBUG_MARGIN; } VkDeviceSize GetDebugMargin() const { return VkDeviceSize(IsVirtual() ? 0 : VMA_DEBUG_MARGIN); }
void DebugLogAllocation(VkDeviceSize offset, VkDeviceSize size, void* userData) const; void DebugLogAllocation(VkDeviceSize offset, VkDeviceSize size, void* userData) const;
#if VMA_STATS_STRING_ENABLED #if VMA_STATS_STRING_ENABLED
@ -6751,7 +6773,7 @@ class VmaBlockMetadata_Generic : public VmaBlockMetadata
{ {
friend class VmaDefragmentationAlgorithm_Generic; friend class VmaDefragmentationAlgorithm_Generic;
friend class VmaDefragmentationAlgorithm_Fast; friend class VmaDefragmentationAlgorithm_Fast;
VMA_CLASS_NO_COPY(VmaBlockMetadata_Generic) VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockMetadata_Generic)
public: public:
VmaBlockMetadata_Generic(const VkAllocationCallbacks* pAllocationCallbacks, VmaBlockMetadata_Generic(const VkAllocationCallbacks* pAllocationCallbacks,
VkDeviceSize bufferImageGranularity, bool isVirtual); VkDeviceSize bufferImageGranularity, bool isVirtual);
@ -7590,7 +7612,7 @@ GetSize() +-------+
*/ */
class VmaBlockMetadata_Linear : public VmaBlockMetadata class VmaBlockMetadata_Linear : public VmaBlockMetadata
{ {
VMA_CLASS_NO_COPY(VmaBlockMetadata_Linear) VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockMetadata_Linear)
public: public:
VmaBlockMetadata_Linear(const VkAllocationCallbacks* pAllocationCallbacks, VmaBlockMetadata_Linear(const VkAllocationCallbacks* pAllocationCallbacks,
VkDeviceSize bufferImageGranularity, bool isVirtual); VkDeviceSize bufferImageGranularity, bool isVirtual);
@ -9209,7 +9231,7 @@ m_LevelCount is the maximum number of levels to use in the current object.
*/ */
class VmaBlockMetadata_Buddy : public VmaBlockMetadata class VmaBlockMetadata_Buddy : public VmaBlockMetadata
{ {
VMA_CLASS_NO_COPY(VmaBlockMetadata_Buddy) VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockMetadata_Buddy)
public: public:
VmaBlockMetadata_Buddy(const VkAllocationCallbacks* pAllocationCallbacks, VmaBlockMetadata_Buddy(const VkAllocationCallbacks* pAllocationCallbacks,
VkDeviceSize bufferImageGranularity, bool isVirtual); VkDeviceSize bufferImageGranularity, bool isVirtual);
@ -9908,7 +9930,7 @@ void VmaBlockMetadata_Buddy::PrintDetailedMapNode(class VmaJsonWriter& json, con
// VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT for fastest alloc time possible. // VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT for fastest alloc time possible.
class VmaBlockMetadata_TLSF : public VmaBlockMetadata class VmaBlockMetadata_TLSF : public VmaBlockMetadata
{ {
VMA_CLASS_NO_COPY(VmaBlockMetadata_TLSF) VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockMetadata_TLSF)
public: public:
VmaBlockMetadata_TLSF(const VkAllocationCallbacks* pAllocationCallbacks, VmaBlockMetadata_TLSF(const VkAllocationCallbacks* pAllocationCallbacks,
VkDeviceSize bufferImageGranularity, bool isVirtual); VkDeviceSize bufferImageGranularity, bool isVirtual);
@ -10070,7 +10092,7 @@ void VmaBlockMetadata_TLSF::Init(VkDeviceSize size)
else else
m_ListsCount += 4; m_ListsCount += 4;
m_MemoryClasses = memoryClass + 2; m_MemoryClasses = memoryClass + uint8_t(2);
memset(m_InnerIsFreeBitmap, 0, MAX_MEMORY_CLASSES * sizeof(uint32_t)); memset(m_InnerIsFreeBitmap, 0, MAX_MEMORY_CLASSES * sizeof(uint32_t));
m_FreeList = vma_new_array(GetAllocationCallbacks(), Block*, m_ListsCount); m_FreeList = vma_new_array(GetAllocationCallbacks(), Block*, m_ListsCount);
@ -10263,7 +10285,7 @@ bool VmaBlockMetadata_TLSF::CreateAllocationRequest(
// Round up to the next block // Round up to the next block
VkDeviceSize sizeForNextList = allocSize; VkDeviceSize sizeForNextList = allocSize;
VkDeviceSize smallSizeStep = SMALL_BUFFER_SIZE / (IsVirtual() ? 1 << SECOND_LEVEL_INDEX : 4); VkDeviceSize smallSizeStep = VkDeviceSize(SMALL_BUFFER_SIZE / (IsVirtual() ? 1 << SECOND_LEVEL_INDEX : 4));
if (allocSize > SMALL_BUFFER_SIZE) if (allocSize > SMALL_BUFFER_SIZE)
{ {
sizeForNextList += (1ULL << (VMA_BITSCAN_MSB(allocSize) - SECOND_LEVEL_INDEX)); sizeForNextList += (1ULL << (VMA_BITSCAN_MSB(allocSize) - SECOND_LEVEL_INDEX));
@ -10672,7 +10694,7 @@ void VmaBlockMetadata_TLSF::DebugLogAllAllocations() const
uint8_t VmaBlockMetadata_TLSF::SizeToMemoryClass(VkDeviceSize size) const uint8_t VmaBlockMetadata_TLSF::SizeToMemoryClass(VkDeviceSize size) const
{ {
if (size > SMALL_BUFFER_SIZE) if (size > SMALL_BUFFER_SIZE)
return VMA_BITSCAN_MSB(size) - MEMORY_CLASS_SHIFT; return uint8_t(VMA_BITSCAN_MSB(size) - MEMORY_CLASS_SHIFT);
return 0; return 0;
} }
@ -10847,7 +10869,7 @@ Synchronized internally with a mutex.
class VmaBlockVector class VmaBlockVector
{ {
friend struct VmaDefragmentationContext_T; friend struct VmaDefragmentationContext_T;
VMA_CLASS_NO_COPY(VmaBlockVector) VMA_CLASS_NO_COPY_NO_MOVE(VmaBlockVector)
public: public:
VmaBlockVector( VmaBlockVector(
VmaAllocator hAllocator, VmaAllocator hAllocator,
@ -10966,7 +10988,7 @@ private:
#ifndef _VMA_DEFRAGMENTATION_CONTEXT #ifndef _VMA_DEFRAGMENTATION_CONTEXT
struct VmaDefragmentationContext_T struct VmaDefragmentationContext_T
{ {
VMA_CLASS_NO_COPY(VmaDefragmentationContext_T) VMA_CLASS_NO_COPY_NO_MOVE(VmaDefragmentationContext_T)
public: public:
VmaDefragmentationContext_T( VmaDefragmentationContext_T(
VmaAllocator hAllocator, VmaAllocator hAllocator,
@ -11053,7 +11075,7 @@ private:
struct VmaPool_T struct VmaPool_T
{ {
friend struct VmaPoolListItemTraits; friend struct VmaPoolListItemTraits;
VMA_CLASS_NO_COPY(VmaPool_T) VMA_CLASS_NO_COPY_NO_MOVE(VmaPool_T)
public: public:
VmaBlockVector m_BlockVector; VmaBlockVector m_BlockVector;
VmaDedicatedAllocationList m_DedicatedAllocations; VmaDedicatedAllocationList m_DedicatedAllocations;
@ -11095,6 +11117,9 @@ struct VmaPoolListItemTraits
#ifndef _VMA_CURRENT_BUDGET_DATA #ifndef _VMA_CURRENT_BUDGET_DATA
struct VmaCurrentBudgetData struct VmaCurrentBudgetData
{ {
VMA_CLASS_NO_COPY_NO_MOVE(VmaCurrentBudgetData)
public:
VMA_ATOMIC_UINT32 m_BlockCount[VK_MAX_MEMORY_HEAPS]; VMA_ATOMIC_UINT32 m_BlockCount[VK_MAX_MEMORY_HEAPS];
VMA_ATOMIC_UINT32 m_AllocationCount[VK_MAX_MEMORY_HEAPS]; VMA_ATOMIC_UINT32 m_AllocationCount[VK_MAX_MEMORY_HEAPS];
VMA_ATOMIC_UINT64 m_BlockBytes[VK_MAX_MEMORY_HEAPS]; VMA_ATOMIC_UINT64 m_BlockBytes[VK_MAX_MEMORY_HEAPS];
@ -11163,7 +11188,7 @@ Thread-safe wrapper over VmaPoolAllocator free list, for allocation of VmaAlloca
*/ */
class VmaAllocationObjectAllocator class VmaAllocationObjectAllocator
{ {
VMA_CLASS_NO_COPY(VmaAllocationObjectAllocator) VMA_CLASS_NO_COPY_NO_MOVE(VmaAllocationObjectAllocator)
public: public:
VmaAllocationObjectAllocator(const VkAllocationCallbacks* pAllocationCallbacks) VmaAllocationObjectAllocator(const VkAllocationCallbacks* pAllocationCallbacks)
: m_Allocator(pAllocationCallbacks, 1024) {} : m_Allocator(pAllocationCallbacks, 1024) {}
@ -11193,7 +11218,7 @@ void VmaAllocationObjectAllocator::Free(VmaAllocation hAlloc)
#ifndef _VMA_VIRTUAL_BLOCK_T #ifndef _VMA_VIRTUAL_BLOCK_T
struct VmaVirtualBlock_T struct VmaVirtualBlock_T
{ {
VMA_CLASS_NO_COPY(VmaVirtualBlock_T) VMA_CLASS_NO_COPY_NO_MOVE(VmaVirtualBlock_T)
public: public:
const bool m_AllocationCallbacksSpecified; const bool m_AllocationCallbacksSpecified;
const VkAllocationCallbacks m_AllocationCallbacks; const VkAllocationCallbacks m_AllocationCallbacks;
@ -11333,7 +11358,7 @@ void VmaVirtualBlock_T::BuildStatsString(bool detailedMap, VmaStringBuilder& sb)
// Main allocator object. // Main allocator object.
struct VmaAllocator_T struct VmaAllocator_T
{ {
VMA_CLASS_NO_COPY(VmaAllocator_T) VMA_CLASS_NO_COPY_NO_MOVE(VmaAllocator_T)
public: public:
bool m_UseMutex; bool m_UseMutex;
uint32_t m_VulkanApiVersion; uint32_t m_VulkanApiVersion;
@ -13197,29 +13222,24 @@ VkResult VmaDefragmentationContext_T::DefragmentPassEnd(VmaDefragmentationPassMo
m_PassStats.bytesFreed += freedBlockSize; m_PassStats.bytesFreed += freedBlockSize;
} }
switch (m_Algorithm) if(m_Algorithm == VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT &&
m_AlgorithmState != VMA_NULL)
{ {
case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: // Avoid unnecessary tries to allocate when new free block is available
{ StateExtensive& state = reinterpret_cast<StateExtensive*>(m_AlgorithmState)[vectorIndex];
if (m_AlgorithmState != VMA_NULL) if (state.firstFreeBlock != SIZE_MAX)
{ {
// Avoid unnecessary tries to allocate when new free block is available const size_t diff = prevCount - currentCount;
StateExtensive& state = reinterpret_cast<StateExtensive*>(m_AlgorithmState)[vectorIndex]; if (state.firstFreeBlock >= diff)
if (state.firstFreeBlock != SIZE_MAX)
{ {
const size_t diff = prevCount - currentCount; state.firstFreeBlock -= diff;
if (state.firstFreeBlock >= diff) if (state.firstFreeBlock != 0)
{ state.firstFreeBlock -= vector->GetBlock(state.firstFreeBlock - 1)->m_pMetadata->IsEmpty();
state.firstFreeBlock -= diff;
if (state.firstFreeBlock != 0)
state.firstFreeBlock -= vector->GetBlock(state.firstFreeBlock - 1)->m_pMetadata->IsEmpty();
}
else
state.firstFreeBlock = 0;
} }
else
state.firstFreeBlock = 0;
} }
} }
}
} }
moveInfo.moveCount = 0; moveInfo.moveCount = 0;
moveInfo.pMoves = VMA_NULL; moveInfo.pMoves = VMA_NULL;
@ -13360,8 +13380,8 @@ bool VmaDefragmentationContext_T::IncrementCounters(VkDeviceSize bytes)
// Early return when max found // Early return when max found
if (++m_PassStats.allocationsMoved >= m_MaxPassAllocations || m_PassStats.bytesMoved >= m_MaxPassBytes) if (++m_PassStats.allocationsMoved >= m_MaxPassAllocations || m_PassStats.bytesMoved >= m_MaxPassBytes)
{ {
VMA_ASSERT(m_PassStats.allocationsMoved == m_MaxPassAllocations || VMA_ASSERT((m_PassStats.allocationsMoved == m_MaxPassAllocations ||
m_PassStats.bytesMoved == m_MaxPassBytes && "Exceeded maximal pass threshold!"); m_PassStats.bytesMoved == m_MaxPassBytes) && "Exceeded maximal pass threshold!");
return true; return true;
} }
return false; return false;
@ -13780,6 +13800,7 @@ bool VmaDefragmentationContext_T::ComputeDefragmentation_Extensive(VmaBlockVecto
} }
else else
break; break;
VMA_FALLTHROUGH; // Fallthrough
} }
case StateExtensive::Operation::MoveBuffers: case StateExtensive::Operation::MoveBuffers:
{ {
@ -13804,6 +13825,7 @@ bool VmaDefragmentationContext_T::ComputeDefragmentation_Extensive(VmaBlockVecto
} }
else else
break; break;
VMA_FALLTHROUGH; // Fallthrough
} }
case StateExtensive::Operation::MoveAll: case StateExtensive::Operation::MoveAll:
{ {
@ -14201,6 +14223,12 @@ void VmaAllocator_T::ImportVulkanFunctions_Static()
m_VulkanFunctions.vkGetImageMemoryRequirements2KHR = (PFN_vkGetImageMemoryRequirements2)vkGetImageMemoryRequirements2; m_VulkanFunctions.vkGetImageMemoryRequirements2KHR = (PFN_vkGetImageMemoryRequirements2)vkGetImageMemoryRequirements2;
m_VulkanFunctions.vkBindBufferMemory2KHR = (PFN_vkBindBufferMemory2)vkBindBufferMemory2; m_VulkanFunctions.vkBindBufferMemory2KHR = (PFN_vkBindBufferMemory2)vkBindBufferMemory2;
m_VulkanFunctions.vkBindImageMemory2KHR = (PFN_vkBindImageMemory2)vkBindImageMemory2; m_VulkanFunctions.vkBindImageMemory2KHR = (PFN_vkBindImageMemory2)vkBindImageMemory2;
}
#endif
#if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000
if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0))
{
m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties2KHR = (PFN_vkGetPhysicalDeviceMemoryProperties2)vkGetPhysicalDeviceMemoryProperties2; m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties2KHR = (PFN_vkGetPhysicalDeviceMemoryProperties2)vkGetPhysicalDeviceMemoryProperties2;
} }
#endif #endif
@ -14253,7 +14281,7 @@ void VmaAllocator_T::ImportVulkanFunctions_Custom(const VmaVulkanFunctions* pVul
VMA_COPY_IF_NOT_NULL(vkBindImageMemory2KHR); VMA_COPY_IF_NOT_NULL(vkBindImageMemory2KHR);
#endif #endif
#if VMA_MEMORY_BUDGET #if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000
VMA_COPY_IF_NOT_NULL(vkGetPhysicalDeviceMemoryProperties2KHR); VMA_COPY_IF_NOT_NULL(vkGetPhysicalDeviceMemoryProperties2KHR);
#endif #endif
@ -14308,8 +14336,18 @@ void VmaAllocator_T::ImportVulkanFunctions_Dynamic()
VMA_FETCH_DEVICE_FUNC(vkGetImageMemoryRequirements2KHR, PFN_vkGetImageMemoryRequirements2, "vkGetImageMemoryRequirements2"); VMA_FETCH_DEVICE_FUNC(vkGetImageMemoryRequirements2KHR, PFN_vkGetImageMemoryRequirements2, "vkGetImageMemoryRequirements2");
VMA_FETCH_DEVICE_FUNC(vkBindBufferMemory2KHR, PFN_vkBindBufferMemory2, "vkBindBufferMemory2"); VMA_FETCH_DEVICE_FUNC(vkBindBufferMemory2KHR, PFN_vkBindBufferMemory2, "vkBindBufferMemory2");
VMA_FETCH_DEVICE_FUNC(vkBindImageMemory2KHR, PFN_vkBindImageMemory2, "vkBindImageMemory2"); VMA_FETCH_DEVICE_FUNC(vkBindImageMemory2KHR, PFN_vkBindImageMemory2, "vkBindImageMemory2");
}
#endif
#if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000
if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0))
{
VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2, "vkGetPhysicalDeviceMemoryProperties2"); VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2, "vkGetPhysicalDeviceMemoryProperties2");
} }
else if(m_UseExtMemoryBudget)
{
VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2, "vkGetPhysicalDeviceMemoryProperties2KHR");
}
#endif #endif
#if VMA_DEDICATED_ALLOCATION #if VMA_DEDICATED_ALLOCATION
@ -14328,8 +14366,12 @@ void VmaAllocator_T::ImportVulkanFunctions_Dynamic()
} }
#endif // #if VMA_BIND_MEMORY2 #endif // #if VMA_BIND_MEMORY2
#if VMA_MEMORY_BUDGET #if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000
if(m_UseExtMemoryBudget) if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0))
{
VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2KHR, "vkGetPhysicalDeviceMemoryProperties2");
}
else if(m_UseExtMemoryBudget)
{ {
VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2KHR, "vkGetPhysicalDeviceMemoryProperties2KHR"); VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2KHR, "vkGetPhysicalDeviceMemoryProperties2KHR");
} }
@ -15374,7 +15416,7 @@ VkResult VmaAllocator_T::CheckCorruption(uint32_t memoryTypeBits)
VkResult VmaAllocator_T::AllocateVulkanMemory(const VkMemoryAllocateInfo* pAllocateInfo, VkDeviceMemory* pMemory) VkResult VmaAllocator_T::AllocateVulkanMemory(const VkMemoryAllocateInfo* pAllocateInfo, VkDeviceMemory* pMemory)
{ {
AtomicTransactionalIncrement<uint32_t> deviceMemoryCountIncrement; AtomicTransactionalIncrement<VMA_ATOMIC_UINT32> deviceMemoryCountIncrement;
const uint64_t prevDeviceMemoryCount = deviceMemoryCountIncrement.Increment(&m_DeviceMemoryCount); const uint64_t prevDeviceMemoryCount = deviceMemoryCountIncrement.Increment(&m_DeviceMemoryCount);
#if VMA_DEBUG_DONT_EXCEED_MAX_MEMORY_ALLOCATION_COUNT #if VMA_DEBUG_DONT_EXCEED_MAX_MEMORY_ALLOCATION_COUNT
if(prevDeviceMemoryCount >= m_PhysicalDeviceProperties.limits.maxMemoryAllocationCount) if(prevDeviceMemoryCount >= m_PhysicalDeviceProperties.limits.maxMemoryAllocationCount)
@ -15531,6 +15573,7 @@ VkResult VmaAllocator_T::Map(VmaAllocation hAllocation, void** ppData)
} }
return res; return res;
} }
VMA_FALLTHROUGH; // Fallthrough
case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED:
return hAllocation->DedicatedAllocMap(this, ppData); return hAllocation->DedicatedAllocMap(this, ppData);
default: default:
@ -18165,6 +18208,12 @@ vkDestroyImage(allocator, img2, nullptr);
vkDestroyImage(allocator, img1, nullptr); vkDestroyImage(allocator, img1, nullptr);
\endcode \endcode
VMA also provides convenience functions that create a buffer or image and bind it to memory
represented by an existing #VmaAllocation:
vmaCreateAliasingBuffer(), vmaCreateAliasingBuffer2(),
vmaCreateAliasingImage(), vmaCreateAliasingImage2().
Versions with "2" offer additional parameter `allocationLocalOffset`.
Remember that using resources that alias in memory requires proper synchronization. Remember that using resources that alias in memory requires proper synchronization.
You need to issue a memory barrier to make sure commands that use `img1` and `img2` You need to issue a memory barrier to make sure commands that use `img1` and `img2`
don't overlap on GPU timeline. don't overlap on GPU timeline.
@ -19162,6 +19211,7 @@ else
// [Executed in runtime]: // [Executed in runtime]:
memcpy(stagingAllocInfo.pMappedData, myData, myDataSize); memcpy(stagingAllocInfo.pMappedData, myData, myDataSize);
vmaFlushAllocation(allocator, stagingAlloc, 0, VK_WHOLE_SIZE);
//vkCmdPipelineBarrier: VK_ACCESS_HOST_WRITE_BIT --> VK_ACCESS_TRANSFER_READ_BIT //vkCmdPipelineBarrier: VK_ACCESS_HOST_WRITE_BIT --> VK_ACCESS_TRANSFER_READ_BIT
VkBufferCopy bufCopy = { VkBufferCopy bufCopy = {
0, // srcOffset 0, // srcOffset

View File

@ -768,6 +768,24 @@ namespace VMA_HPP_NAMESPACE {
return result; return result;
} }
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_INLINE typename VULKAN_HPP_NAMESPACE::ResultValueType<VULKAN_HPP_NAMESPACE::Buffer>::type Allocator::createAliasingBuffer2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::BufferCreateInfo& bufferCreateInfo) const {
VULKAN_HPP_NAMESPACE::Buffer buffer;
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( vmaCreateAliasingBuffer2(m_allocator, static_cast<VmaAllocation>(allocation), static_cast<VkDeviceSize>(allocationLocalOffset), reinterpret_cast<const VkBufferCreateInfo*>(&bufferCreateInfo), reinterpret_cast<VkBuffer*>(&buffer)) );
resultCheck(result, VMA_HPP_NAMESPACE_STRING "::Allocator::createAliasingBuffer2");
return createResultValueType(result, buffer);
}
#endif
VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Allocator::createAliasingBuffer2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::BufferCreateInfo* bufferCreateInfo,
VULKAN_HPP_NAMESPACE::Buffer* buffer) const {
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( vmaCreateAliasingBuffer2(m_allocator, static_cast<VmaAllocation>(allocation), static_cast<VkDeviceSize>(allocationLocalOffset), reinterpret_cast<const VkBufferCreateInfo*>(bufferCreateInfo), reinterpret_cast<VkBuffer*>(buffer)) );
return result;
}
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_INLINE void Allocator::destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer, VULKAN_HPP_INLINE void Allocator::destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer,
Allocation allocation) const { Allocation allocation) const {
@ -829,6 +847,24 @@ namespace VMA_HPP_NAMESPACE {
return result; return result;
} }
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_INLINE typename VULKAN_HPP_NAMESPACE::ResultValueType<VULKAN_HPP_NAMESPACE::Image>::type Allocator::createAliasingImage2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::ImageCreateInfo& imageCreateInfo) const {
VULKAN_HPP_NAMESPACE::Image image;
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( vmaCreateAliasingImage2(m_allocator, static_cast<VmaAllocation>(allocation), static_cast<VkDeviceSize>(allocationLocalOffset), reinterpret_cast<const VkImageCreateInfo*>(&imageCreateInfo), reinterpret_cast<VkImage*>(&image)) );
resultCheck(result, VMA_HPP_NAMESPACE_STRING "::Allocator::createAliasingImage2");
return createResultValueType(result, image);
}
#endif
VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Allocator::createAliasingImage2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::ImageCreateInfo* imageCreateInfo,
VULKAN_HPP_NAMESPACE::Image* image) const {
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( vmaCreateAliasingImage2(m_allocator, static_cast<VmaAllocation>(allocation), static_cast<VkDeviceSize>(allocationLocalOffset), reinterpret_cast<const VkImageCreateInfo*>(imageCreateInfo), reinterpret_cast<VkImage*>(image)) );
return result;
}
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_INLINE void Allocator::destroyImage(VULKAN_HPP_NAMESPACE::Image image, VULKAN_HPP_INLINE void Allocator::destroyImage(VULKAN_HPP_NAMESPACE::Image image,
Allocation allocation) const { Allocation allocation) const {

View File

@ -649,6 +649,16 @@ namespace VMA_HPP_NAMESPACE {
const VULKAN_HPP_NAMESPACE::BufferCreateInfo* bufferCreateInfo, const VULKAN_HPP_NAMESPACE::BufferCreateInfo* bufferCreateInfo,
VULKAN_HPP_NAMESPACE::Buffer* buffer) const; VULKAN_HPP_NAMESPACE::Buffer* buffer) const;
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS typename VULKAN_HPP_NAMESPACE::ResultValueType<VULKAN_HPP_NAMESPACE::Buffer>::type createAliasingBuffer2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::BufferCreateInfo& bufferCreateInfo) const;
#endif
VULKAN_HPP_NODISCARD VULKAN_HPP_NAMESPACE::Result createAliasingBuffer2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::BufferCreateInfo* bufferCreateInfo,
VULKAN_HPP_NAMESPACE::Buffer* buffer) const;
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
void destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer, void destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer,
Allocation allocation) const; Allocation allocation) const;
@ -681,6 +691,16 @@ namespace VMA_HPP_NAMESPACE {
const VULKAN_HPP_NAMESPACE::ImageCreateInfo* imageCreateInfo, const VULKAN_HPP_NAMESPACE::ImageCreateInfo* imageCreateInfo,
VULKAN_HPP_NAMESPACE::Image* image) const; VULKAN_HPP_NAMESPACE::Image* image) const;
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS typename VULKAN_HPP_NAMESPACE::ResultValueType<VULKAN_HPP_NAMESPACE::Image>::type createAliasingImage2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::ImageCreateInfo& imageCreateInfo) const;
#endif
VULKAN_HPP_NODISCARD VULKAN_HPP_NAMESPACE::Result createAliasingImage2(Allocation allocation,
VULKAN_HPP_NAMESPACE::DeviceSize allocationLocalOffset,
const VULKAN_HPP_NAMESPACE::ImageCreateInfo* imageCreateInfo,
VULKAN_HPP_NAMESPACE::Image* image) const;
#ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
void destroyImage(VULKAN_HPP_NAMESPACE::Image image, void destroyImage(VULKAN_HPP_NAMESPACE::Image image,
Allocation allocation) const; Allocation allocation) const;