using Ryujinx.Common.Logging; using Ryujinx.Graphics.GAL; using Silk.NET.Vulkan; using System; using System.Collections.Generic; using System.Runtime.CompilerServices; using System.Runtime.InteropServices; using System.Threading; using VkBuffer = Silk.NET.Vulkan.Buffer; using VkFormat = Silk.NET.Vulkan.Format; namespace Ryujinx.Graphics.Vulkan { class BufferHolder : IDisposable, IMirrorable, IMirrorable { private const int MaxUpdateBufferSize = 0x10000; private const int SetCountThreshold = 100; private const int WriteCountThreshold = 50; private const int FlushCountThreshold = 5; public const int DeviceLocalSizeThreshold = 256 * 1024; // 256kb public const AccessFlags DefaultAccessFlags = AccessFlags.IndirectCommandReadBit | AccessFlags.ShaderReadBit | AccessFlags.ShaderWriteBit | AccessFlags.TransferReadBit | AccessFlags.TransferWriteBit | AccessFlags.UniformReadBit; private readonly VulkanRenderer _gd; private readonly Device _device; private MemoryAllocation _allocation; private Auto _buffer; private Auto _allocationAuto; private readonly bool _allocationImported; private ulong _bufferHandle; private CacheByRange _cachedConvertedBuffers; public int Size { get; } private IntPtr _map; private MultiFenceHolder _waitable; private bool _lastAccessIsWrite; private BufferAllocationType _baseType; private BufferAllocationType _currentType; private bool _swapQueued; public BufferAllocationType DesiredType { get; private set; } private int _setCount; private int _writeCount; private int _flushCount; private int _flushTemp; private int _lastFlushWrite = -1; private readonly ReaderWriterLockSlim _flushLock; private FenceHolder _flushFence; private int _flushWaiting; private List _swapActions; private byte[] _pendingData; private BufferMirrorRangeList _pendingDataRanges; private Dictionary _mirrors; private bool _useMirrors; public BufferHolder(VulkanRenderer gd, Device device, VkBuffer buffer, MemoryAllocation allocation, int size, BufferAllocationType type, BufferAllocationType currentType) { _gd = gd; _device = device; _allocation = allocation; _allocationAuto = new Auto(allocation); _waitable = new MultiFenceHolder(size); _buffer = new Auto(new DisposableBuffer(gd.Api, device, buffer), this, _waitable, _allocationAuto); _bufferHandle = buffer.Handle; Size = size; _map = allocation.HostPointer; _baseType = type; _currentType = currentType; DesiredType = currentType; _flushLock = new ReaderWriterLockSlim(); _useMirrors = gd.IsTBDR; } public BufferHolder(VulkanRenderer gd, Device device, VkBuffer buffer, Auto allocation, int size, BufferAllocationType type, BufferAllocationType currentType, int offset) { _gd = gd; _device = device; _allocation = allocation.GetUnsafe(); _allocationAuto = allocation; _allocationImported = true; _waitable = new MultiFenceHolder(size); _buffer = new Auto(new DisposableBuffer(gd.Api, device, buffer), this, _waitable, _allocationAuto); _bufferHandle = buffer.Handle; Size = size; _map = _allocation.HostPointer + offset; _baseType = type; _currentType = currentType; DesiredType = currentType; _flushLock = new ReaderWriterLockSlim(); } public BufferHolder(VulkanRenderer gd, Device device, VkBuffer buffer, int size, Auto[] storageAllocations) { _gd = gd; _device = device; _waitable = new MultiFenceHolder(size); _buffer = new Auto(new DisposableBuffer(gd.Api, device, buffer), _waitable, storageAllocations); _bufferHandle = buffer.Handle; Size = size; _baseType = BufferAllocationType.Sparse; _currentType = BufferAllocationType.Sparse; DesiredType = BufferAllocationType.Sparse; _flushLock = new ReaderWriterLockSlim(); } public bool TryBackingSwap(ref CommandBufferScoped? cbs) { if (_swapQueued && DesiredType != _currentType) { // Only swap if the buffer is not used in any queued command buffer. bool isRented = _buffer.HasRentedCommandBufferDependency(_gd.CommandBufferPool); if (!isRented && _gd.CommandBufferPool.OwnedByCurrentThread && !_flushLock.IsReadLockHeld && (_pendingData == null || cbs != null)) { var currentAllocation = _allocationAuto; var currentBuffer = _buffer; IntPtr currentMap = _map; (VkBuffer buffer, MemoryAllocation allocation, BufferAllocationType resultType) = _gd.BufferManager.CreateBacking(_gd, Size, DesiredType, false, false, _currentType); if (buffer.Handle != 0) { if (cbs != null) { ClearMirrors(cbs.Value, 0, Size); } _flushLock.EnterWriteLock(); ClearFlushFence(); _waitable = new MultiFenceHolder(Size); _allocation = allocation; _allocationAuto = new Auto(allocation); _buffer = new Auto(new DisposableBuffer(_gd.Api, _device, buffer), this, _waitable, _allocationAuto); _bufferHandle = buffer.Handle; _map = allocation.HostPointer; if (_map != IntPtr.Zero && currentMap != IntPtr.Zero) { // Copy data directly. Readbacks don't have to wait if this is done. unsafe { new Span((void*)currentMap, Size).CopyTo(new Span((void*)_map, Size)); } } else { cbs ??= _gd.CommandBufferPool.Rent(); CommandBufferScoped cbsV = cbs.Value; Copy(_gd, cbsV, currentBuffer, _buffer, 0, 0, Size); // Need to wait for the data to reach the new buffer before data can be flushed. _flushFence = _gd.CommandBufferPool.GetFence(cbsV.CommandBufferIndex); _flushFence.Get(); } Logger.Debug?.PrintMsg(LogClass.Gpu, $"Converted {Size} buffer {_currentType} to {resultType}"); _currentType = resultType; if (_swapActions != null) { foreach (var action in _swapActions) { action(); } _swapActions.Clear(); } currentBuffer.Dispose(); currentAllocation.Dispose(); _gd.PipelineInternal.SwapBuffer(currentBuffer, _buffer); _flushLock.ExitWriteLock(); } _swapQueued = false; return true; } return false; } _swapQueued = false; return true; } private void ConsiderBackingSwap() { if (_baseType == BufferAllocationType.Auto) { // When flushed, wait for a bit more info to make a decision. bool wasFlushed = _flushTemp > 0; int multiplier = wasFlushed ? 2 : 0; if (_writeCount >= (WriteCountThreshold << multiplier) || _setCount >= (SetCountThreshold << multiplier) || _flushCount >= (FlushCountThreshold << multiplier)) { if (_flushCount > 0 || _flushTemp-- > 0) { // Buffers that flush should ideally be mapped in host address space for easy copies. // If the buffer is large it will do better on GPU memory, as there will be more writes than data flushes (typically individual pages). // If it is small, then it's likely most of the buffer will be flushed so we want it on host memory, as access is cached. bool hostMappingSensitive = _gd.Vendor == Vendor.Nvidia; bool deviceLocalMapped = Size > DeviceLocalSizeThreshold || (wasFlushed && _writeCount > _flushCount * 10 && hostMappingSensitive) || _currentType == BufferAllocationType.DeviceLocalMapped; DesiredType = deviceLocalMapped ? BufferAllocationType.DeviceLocalMapped : BufferAllocationType.HostMapped; // It's harder for a buffer that is flushed to revert to another type of mapping. if (_flushCount > 0) { _flushTemp = 1000; } } else if (_writeCount >= (WriteCountThreshold << multiplier)) { // Buffers that are written often should ideally be in the device local heap. (Storage buffers) DesiredType = BufferAllocationType.DeviceLocal; } else if (_setCount > (SetCountThreshold << multiplier)) { // Buffers that have their data set often should ideally be host mapped. (Constant buffers) DesiredType = BufferAllocationType.HostMapped; } _lastFlushWrite = -1; _flushCount = 0; _writeCount = 0; _setCount = 0; } if (!_swapQueued && DesiredType != _currentType) { _swapQueued = true; _gd.PipelineInternal.AddBackingSwap(this); } } } public void Pin() { if (_baseType == BufferAllocationType.Auto) { _baseType = _currentType; } } public unsafe Auto CreateView(VkFormat format, int offset, int size, Action invalidateView) { var bufferViewCreateInfo = new BufferViewCreateInfo { SType = StructureType.BufferViewCreateInfo, Buffer = new VkBuffer(_bufferHandle), Format = format, Offset = (uint)offset, Range = (uint)size, }; _gd.Api.CreateBufferView(_device, bufferViewCreateInfo, null, out var bufferView).ThrowOnError(); (_swapActions ??= new List()).Add(invalidateView); return new Auto(new DisposableBufferView(_gd.Api, _device, bufferView), this, _waitable, _buffer); } public void InheritMetrics(BufferHolder other) { _setCount = other._setCount; _writeCount = other._writeCount; _flushCount = other._flushCount; _flushTemp = other._flushTemp; } public unsafe void InsertBarrier(CommandBuffer commandBuffer, bool isWrite) { // If the last access is write, we always need a barrier to be sure we will read or modify // the correct data. // If the last access is read, and current one is a write, we need to wait until the // read finishes to avoid overwriting data still in use. // Otherwise, if the last access is a read and the current one too, we don't need barriers. bool needsBarrier = isWrite || _lastAccessIsWrite; _lastAccessIsWrite = isWrite; if (needsBarrier) { MemoryBarrier memoryBarrier = new() { SType = StructureType.MemoryBarrier, SrcAccessMask = DefaultAccessFlags, DstAccessMask = DefaultAccessFlags, }; _gd.Api.CmdPipelineBarrier( commandBuffer, PipelineStageFlags.AllCommandsBit, PipelineStageFlags.AllCommandsBit, DependencyFlags.DeviceGroupBit, 1, memoryBarrier, 0, null, 0, null); } } private static ulong ToMirrorKey(int offset, int size) { return ((ulong)offset << 32) | (uint)size; } private static (int offset, int size) FromMirrorKey(ulong key) { return ((int)(key >> 32), (int)key); } private unsafe bool TryGetMirror(CommandBufferScoped cbs, ref int offset, int size, out Auto buffer) { size = Math.Min(size, Size - offset); // Does this binding need to be mirrored? if (!_pendingDataRanges.OverlapsWith(offset, size)) { buffer = null; return false; } var key = ToMirrorKey(offset, size); if (_mirrors.TryGetValue(key, out StagingBufferReserved reserved)) { buffer = reserved.Buffer.GetBuffer(); offset = reserved.Offset; return true; } // Is this mirror allowed to exist? Can't be used for write in any in-flight write. if (_waitable.IsBufferRangeInUse(offset, size, true)) { // Some of the data is not mirrorable, so upload the whole range. ClearMirrors(cbs, offset, size); buffer = null; return false; } // Build data for the new mirror. var baseData = new Span((void*)(_map + offset), size); var modData = _pendingData.AsSpan(offset, size); StagingBufferReserved? newMirror = _gd.BufferManager.StagingBuffer.TryReserveData(cbs, size); if (newMirror != null) { var mirror = newMirror.Value; _pendingDataRanges.FillData(baseData, modData, offset, new Span((void*)(mirror.Buffer._map + mirror.Offset), size)); if (_mirrors.Count == 0) { _gd.PipelineInternal.RegisterActiveMirror(this); } _mirrors.Add(key, mirror); buffer = mirror.Buffer.GetBuffer(); offset = mirror.Offset; return true; } else { // Data could not be placed on the mirror, likely out of space. Force the data to flush. ClearMirrors(cbs, offset, size); buffer = null; return false; } } public Auto GetBuffer() { return _buffer; } public Auto GetBuffer(CommandBuffer commandBuffer, bool isWrite = false, bool isSSBO = false) { if (isWrite) { _writeCount++; SignalWrite(0, Size); } else if (isSSBO) { // Always consider SSBO access for swapping to device local memory. _writeCount++; ConsiderBackingSwap(); } return _buffer; } public Auto GetBuffer(CommandBuffer commandBuffer, int offset, int size, bool isWrite = false) { if (isWrite) { _writeCount++; SignalWrite(offset, size); } return _buffer; } public Auto GetMirrorable(CommandBufferScoped cbs, ref int offset, int size, out bool mirrored) { if (_pendingData != null && TryGetMirror(cbs, ref offset, size, out Auto result)) { mirrored = true; return result; } mirrored = false; return _buffer; } Auto IMirrorable.GetMirrorable(CommandBufferScoped cbs, ref int offset, int size, out bool mirrored) { // Cannot mirror buffer views right now. throw new NotImplementedException(); } public void ClearMirrors() { // Clear mirrors without forcing a flush. This happens when the command buffer is switched, // as all reserved areas on the staging buffer are released. if (_pendingData != null) { _mirrors.Clear(); }; } public void ClearMirrors(CommandBufferScoped cbs, int offset, int size) { // Clear mirrors in the given range, and submit overlapping pending data. if (_pendingData != null) { bool hadMirrors = _mirrors.Count > 0 && RemoveOverlappingMirrors(offset, size); if (_pendingDataRanges.Count() != 0) { UploadPendingData(cbs, offset, size); } if (hadMirrors) { _gd.PipelineInternal.Rebind(_buffer, offset, size); } }; } public void UseMirrors() { _useMirrors = true; } private void UploadPendingData(CommandBufferScoped cbs, int offset, int size) { var ranges = _pendingDataRanges.FindOverlaps(offset, size); if (ranges != null) { _pendingDataRanges.Remove(offset, size); foreach (var range in ranges) { int rangeOffset = Math.Max(offset, range.Offset); int rangeSize = Math.Min(offset + size, range.End) - rangeOffset; if (_gd.PipelineInternal.CurrentCommandBuffer.CommandBuffer.Handle == cbs.CommandBuffer.Handle) { SetData(rangeOffset, _pendingData.AsSpan(rangeOffset, rangeSize), cbs, _gd.PipelineInternal.EndRenderPassDelegate, false); } else { SetData(rangeOffset, _pendingData.AsSpan(rangeOffset, rangeSize), cbs, null, false); } } } } public Auto GetAllocation() { return _allocationAuto; } public (DeviceMemory, ulong) GetDeviceMemoryAndOffset() { return (_allocation.Memory, _allocation.Offset); } public void SignalWrite(int offset, int size) { ConsiderBackingSwap(); if (offset == 0 && size == Size) { _cachedConvertedBuffers.Clear(); } else { _cachedConvertedBuffers.ClearRange(offset, size); } } public BufferHandle GetHandle() { var handle = _bufferHandle; return Unsafe.As(ref handle); } public IntPtr Map(int offset, int mappingSize) { return _map; } private void ClearFlushFence() { // Assumes _flushLock is held as writer. if (_flushFence != null) { if (_flushWaiting == 0) { _flushFence.Put(); } _flushFence = null; } } private void WaitForFlushFence() { if (_flushFence == null) { return; } // If storage has changed, make sure the fence has been reached so that the data is in place. _flushLock.ExitReadLock(); _flushLock.EnterWriteLock(); if (_flushFence != null) { var fence = _flushFence; Interlocked.Increment(ref _flushWaiting); // Don't wait in the lock. _flushLock.ExitWriteLock(); fence.Wait(); _flushLock.EnterWriteLock(); if (Interlocked.Decrement(ref _flushWaiting) == 0) { fence.Put(); } _flushFence = null; } // Assumes the _flushLock is held as reader, returns in same state. _flushLock.ExitWriteLock(); _flushLock.EnterReadLock(); } public PinnedSpan GetData(int offset, int size) { _flushLock.EnterReadLock(); WaitForFlushFence(); if (_lastFlushWrite != _writeCount) { // If it's on the same page as the last flush, ignore it. _lastFlushWrite = _writeCount; _flushCount++; } Span result; if (_map != IntPtr.Zero) { result = GetDataStorage(offset, size); // Need to be careful here, the buffer can't be unmapped while the data is being used. _buffer.IncrementReferenceCount(); _flushLock.ExitReadLock(); return PinnedSpan.UnsafeFromSpan(result, _buffer.DecrementReferenceCount); } BackgroundResource resource = _gd.BackgroundResources.Get(); if (_gd.CommandBufferPool.OwnedByCurrentThread) { _gd.FlushAllCommands(); result = resource.GetFlushBuffer().GetBufferData(_gd.CommandBufferPool, this, offset, size); } else { result = resource.GetFlushBuffer().GetBufferData(resource.GetPool(), this, offset, size); } _flushLock.ExitReadLock(); // Flush buffer is pinned until the next GetBufferData on the thread, which is fine for current uses. return PinnedSpan.UnsafeFromSpan(result); } public unsafe Span GetDataStorage(int offset, int size) { int mappingSize = Math.Min(size, Size - offset); if (_map != IntPtr.Zero) { return new Span((void*)(_map + offset), mappingSize); } throw new InvalidOperationException("The buffer is not host mapped."); } public bool RemoveOverlappingMirrors(int offset, int size) { List toRemove = null; foreach (var key in _mirrors.Keys) { (int keyOffset, int keySize) = FromMirrorKey(key); if (!(offset + size <= keyOffset || offset >= keyOffset + keySize)) { toRemove ??= new List(); toRemove.Add(key); } } if (toRemove != null) { foreach (var key in toRemove) { _mirrors.Remove(key); } return true; } return false; } public unsafe void SetData(int offset, ReadOnlySpan data, CommandBufferScoped? cbs = null, Action endRenderPass = null, bool allowCbsWait = true) { int dataSize = Math.Min(data.Length, Size - offset); if (dataSize == 0) { return; } _setCount++; bool allowMirror = _useMirrors && allowCbsWait && cbs != null && _currentType <= BufferAllocationType.HostMapped; if (_map != IntPtr.Zero) { // If persistently mapped, set the data directly if the buffer is not currently in use. bool isRented = _buffer.HasRentedCommandBufferDependency(_gd.CommandBufferPool); // If the buffer is rented, take a little more time and check if the use overlaps this handle. bool needsFlush = isRented && _waitable.IsBufferRangeInUse(offset, dataSize, false); if (!needsFlush) { WaitForFences(offset, dataSize); data[..dataSize].CopyTo(new Span((void*)(_map + offset), dataSize)); if (_pendingData != null) { bool removed = _pendingDataRanges.Remove(offset, dataSize); if (RemoveOverlappingMirrors(offset, dataSize) || removed) { // If any mirrors were removed, rebind the buffer range. _gd.PipelineInternal.Rebind(_buffer, offset, dataSize); } } SignalWrite(offset, dataSize); return; } } // If the buffer does not have an in-flight write (including an inline update), then upload data to a pendingCopy. if (allowMirror && !_waitable.IsBufferRangeInUse(offset, dataSize, true)) { if (_pendingData == null) { _pendingData = new byte[Size]; _mirrors = new Dictionary(); } data[..dataSize].CopyTo(_pendingData.AsSpan(offset, dataSize)); _pendingDataRanges.Add(offset, dataSize); // Remove any overlapping mirrors. RemoveOverlappingMirrors(offset, dataSize); // Tell the graphics device to rebind any constant buffer that overlaps the newly modified range, as it should access a mirror. _gd.PipelineInternal.Rebind(_buffer, offset, dataSize); return; } if (_pendingData != null) { _pendingDataRanges.Remove(offset, dataSize); } if (cbs != null && _gd.PipelineInternal.RenderPassActive && !(_buffer.HasCommandBufferDependency(cbs.Value) && _waitable.IsBufferRangeInUse(cbs.Value.CommandBufferIndex, offset, dataSize))) { // If the buffer hasn't been used on the command buffer yet, try to preload the data. // This avoids ending and beginning render passes on each buffer data upload. cbs = _gd.PipelineInternal.GetPreloadCommandBuffer(); endRenderPass = null; } if (cbs == null || !VulkanConfiguration.UseFastBufferUpdates || data.Length > MaxUpdateBufferSize || !TryPushData(cbs.Value, endRenderPass, offset, data)) { if (allowCbsWait) { _gd.BufferManager.StagingBuffer.PushData(_gd.CommandBufferPool, cbs, endRenderPass, this, offset, data); } else { bool rentCbs = cbs == null; if (rentCbs) { cbs = _gd.CommandBufferPool.Rent(); } if (!_gd.BufferManager.StagingBuffer.TryPushData(cbs.Value, endRenderPass, this, offset, data)) { // Need to do a slow upload. BufferHolder srcHolder = _gd.BufferManager.Create(_gd, dataSize, baseType: BufferAllocationType.HostMapped); srcHolder.SetDataUnchecked(0, data); var srcBuffer = srcHolder.GetBuffer(); var dstBuffer = this.GetBuffer(cbs.Value.CommandBuffer, true); Copy(_gd, cbs.Value, srcBuffer, dstBuffer, 0, offset, dataSize); srcHolder.Dispose(); } if (rentCbs) { cbs.Value.Dispose(); } } } } public unsafe void SetDataUnchecked(int offset, ReadOnlySpan data) { int dataSize = Math.Min(data.Length, Size - offset); if (dataSize == 0) { return; } if (_map != IntPtr.Zero) { data[..dataSize].CopyTo(new Span((void*)(_map + offset), dataSize)); } else { _gd.BufferManager.StagingBuffer.PushData(_gd.CommandBufferPool, null, null, this, offset, data); } } public unsafe void SetDataUnchecked(int offset, ReadOnlySpan data) where T : unmanaged { SetDataUnchecked(offset, MemoryMarshal.AsBytes(data)); } public void SetDataInline(CommandBufferScoped cbs, Action endRenderPass, int dstOffset, ReadOnlySpan data) { if (!TryPushData(cbs, endRenderPass, dstOffset, data)) { throw new ArgumentException($"Invalid offset 0x{dstOffset:X} or data size 0x{data.Length:X}."); } } private unsafe bool TryPushData(CommandBufferScoped cbs, Action endRenderPass, int dstOffset, ReadOnlySpan data) { if ((dstOffset & 3) != 0 || (data.Length & 3) != 0) { return false; } endRenderPass?.Invoke(); var dstBuffer = GetBuffer(cbs.CommandBuffer, dstOffset, data.Length, true).Get(cbs, dstOffset, data.Length, true).Value; _writeCount--; InsertBufferBarrier( _gd, cbs.CommandBuffer, dstBuffer, DefaultAccessFlags, AccessFlags.TransferWriteBit, PipelineStageFlags.AllCommandsBit, PipelineStageFlags.TransferBit, dstOffset, data.Length); fixed (byte* pData = data) { for (ulong offset = 0; offset < (ulong)data.Length;) { ulong size = Math.Min(MaxUpdateBufferSize, (ulong)data.Length - offset); _gd.Api.CmdUpdateBuffer(cbs.CommandBuffer, dstBuffer, (ulong)dstOffset + offset, size, pData + offset); offset += size; } } InsertBufferBarrier( _gd, cbs.CommandBuffer, dstBuffer, AccessFlags.TransferWriteBit, DefaultAccessFlags, PipelineStageFlags.TransferBit, PipelineStageFlags.AllCommandsBit, dstOffset, data.Length); return true; } public static unsafe void Copy( VulkanRenderer gd, CommandBufferScoped cbs, Auto src, Auto dst, int srcOffset, int dstOffset, int size, bool registerSrcUsage = true) { var srcBuffer = registerSrcUsage ? src.Get(cbs, srcOffset, size).Value : src.GetUnsafe().Value; var dstBuffer = dst.Get(cbs, dstOffset, size, true).Value; InsertBufferBarrier( gd, cbs.CommandBuffer, dstBuffer, DefaultAccessFlags, AccessFlags.TransferWriteBit, PipelineStageFlags.AllCommandsBit, PipelineStageFlags.TransferBit, dstOffset, size); var region = new BufferCopy((ulong)srcOffset, (ulong)dstOffset, (ulong)size); gd.Api.CmdCopyBuffer(cbs.CommandBuffer, srcBuffer, dstBuffer, 1, ®ion); InsertBufferBarrier( gd, cbs.CommandBuffer, dstBuffer, AccessFlags.TransferWriteBit, DefaultAccessFlags, PipelineStageFlags.TransferBit, PipelineStageFlags.AllCommandsBit, dstOffset, size); } public static unsafe void InsertBufferBarrier( VulkanRenderer gd, CommandBuffer commandBuffer, VkBuffer buffer, AccessFlags srcAccessMask, AccessFlags dstAccessMask, PipelineStageFlags srcStageMask, PipelineStageFlags dstStageMask, int offset, int size) { BufferMemoryBarrier memoryBarrier = new() { SType = StructureType.BufferMemoryBarrier, SrcAccessMask = srcAccessMask, DstAccessMask = dstAccessMask, SrcQueueFamilyIndex = Vk.QueueFamilyIgnored, DstQueueFamilyIndex = Vk.QueueFamilyIgnored, Buffer = buffer, Offset = (ulong)offset, Size = (ulong)size, }; gd.Api.CmdPipelineBarrier( commandBuffer, srcStageMask, dstStageMask, 0, 0, null, 1, memoryBarrier, 0, null); } public void WaitForFences() { _waitable.WaitForFences(_gd.Api, _device); } public void WaitForFences(int offset, int size) { _waitable.WaitForFences(_gd.Api, _device, offset, size); } private bool BoundToRange(int offset, ref int size) { if (offset >= Size) { return false; } size = Math.Min(Size - offset, size); return true; } public Auto GetBufferI8ToI16(CommandBufferScoped cbs, int offset, int size) { if (!BoundToRange(offset, ref size)) { return null; } var key = new I8ToI16CacheKey(_gd); if (!_cachedConvertedBuffers.TryGetValue(offset, size, key, out var holder)) { holder = _gd.BufferManager.Create(_gd, (size * 2 + 3) & ~3, baseType: BufferAllocationType.DeviceLocal); _gd.PipelineInternal.EndRenderPass(); _gd.HelperShader.ConvertI8ToI16(_gd, cbs, this, holder, offset, size); key.SetBuffer(holder.GetBuffer()); _cachedConvertedBuffers.Add(offset, size, key, holder); } return holder.GetBuffer(); } public Auto GetAlignedVertexBuffer(CommandBufferScoped cbs, int offset, int size, int stride, int alignment) { if (!BoundToRange(offset, ref size)) { return null; } var key = new AlignedVertexBufferCacheKey(_gd, stride, alignment); if (!_cachedConvertedBuffers.TryGetValue(offset, size, key, out var holder)) { int alignedStride = (stride + (alignment - 1)) & -alignment; holder = _gd.BufferManager.Create(_gd, (size / stride) * alignedStride, baseType: BufferAllocationType.DeviceLocal); _gd.PipelineInternal.EndRenderPass(); _gd.HelperShader.ChangeStride(_gd, cbs, this, holder, offset, size, stride, alignedStride); key.SetBuffer(holder.GetBuffer()); _cachedConvertedBuffers.Add(offset, size, key, holder); } return holder.GetBuffer(); } public Auto GetBufferTopologyConversion(CommandBufferScoped cbs, int offset, int size, IndexBufferPattern pattern, int indexSize) { if (!BoundToRange(offset, ref size)) { return null; } var key = new TopologyConversionCacheKey(_gd, pattern, indexSize); if (!_cachedConvertedBuffers.TryGetValue(offset, size, key, out var holder)) { // The destination index size is always I32. int indexCount = size / indexSize; int convertedCount = pattern.GetConvertedCount(indexCount); holder = _gd.BufferManager.Create(_gd, convertedCount * 4, baseType: BufferAllocationType.DeviceLocal); _gd.PipelineInternal.EndRenderPass(); _gd.HelperShader.ConvertIndexBuffer(_gd, cbs, this, holder, pattern, indexSize, offset, indexCount); key.SetBuffer(holder.GetBuffer()); _cachedConvertedBuffers.Add(offset, size, key, holder); } return holder.GetBuffer(); } public bool TryGetCachedConvertedBuffer(int offset, int size, ICacheKey key, out BufferHolder holder) { return _cachedConvertedBuffers.TryGetValue(offset, size, key, out holder); } public void AddCachedConvertedBuffer(int offset, int size, ICacheKey key, BufferHolder holder) { _cachedConvertedBuffers.Add(offset, size, key, holder); } public void AddCachedConvertedBufferDependency(int offset, int size, ICacheKey key, Dependency dependency) { _cachedConvertedBuffers.AddDependency(offset, size, key, dependency); } public void RemoveCachedConvertedBuffer(int offset, int size, ICacheKey key) { _cachedConvertedBuffers.Remove(offset, size, key); } public void Dispose() { _swapQueued = false; _gd.PipelineInternal?.FlushCommandsIfWeightExceeding(_buffer, (ulong)Size); _buffer.Dispose(); _cachedConvertedBuffers.Dispose(); if (_allocationImported) { _allocationAuto.DecrementReferenceCount(); } else { _allocationAuto?.Dispose(); } _flushLock.EnterWriteLock(); ClearFlushFence(); _flushLock.ExitWriteLock(); } } }