Files
VulkanDriver/src/vulkan/CommandPool.zig
Kbz-8 ee49c86ebb
Some checks failed
Build / build (push) Successful in 8m27s
Test / build_and_test (push) Failing after 25m19s
adding secondary command buffers
2026-03-18 03:22:46 +01:00

106 lines
4.1 KiB
Zig

const std = @import("std");
const vk = @import("vulkan");
const VkError = @import("error_set.zig").VkError;
const VulkanAllocator = @import("VulkanAllocator.zig");
const Dispatchable = @import("Dispatchable.zig").Dispatchable;
const CommandBuffer = @import("CommandBuffer.zig");
const Device = @import("Device.zig");
const Self = @This();
pub const ObjectType: vk.ObjectType = .command_pool;
/// Base capacity of the command buffer pool.
/// Every increase of the capacity will be by this amount.
pub const BUFFER_POOL_BASE_CAPACITY = 64;
owner: *Device,
flags: vk.CommandPoolCreateFlags,
queue_family_index: u32,
host_allocator: VulkanAllocator,
/// Contiguous dynamic array of command buffers with free ones
/// grouped at the end.
/// When freed swaps happen to keep the free buffers at the end.
buffers: std.ArrayList(*Dispatchable(CommandBuffer)),
/// Index of the first free command buffer.
first_free_buffer_index: usize,
vtable: *const VTable,
pub const VTable = struct {
createCommandBuffer: *const fn (*Self, std.mem.Allocator, *const vk.CommandBufferAllocateInfo) VkError!*CommandBuffer,
destroy: *const fn (*Self, std.mem.Allocator) void,
reset: *const fn (*Self, vk.CommandPoolResetFlags) VkError!void,
};
pub fn init(device: *Device, allocator: std.mem.Allocator, info: *const vk.CommandPoolCreateInfo) VkError!Self {
return .{
.owner = device,
.flags = info.flags,
.queue_family_index = info.queue_family_index,
.host_allocator = VulkanAllocator.from(allocator).clone(),
.buffers = std.ArrayList(*Dispatchable(CommandBuffer)).initCapacity(allocator, BUFFER_POOL_BASE_CAPACITY) catch return VkError.OutOfHostMemory,
.first_free_buffer_index = 0,
.vtable = undefined,
};
}
pub fn allocateCommandBuffers(self: *Self, info: *const vk.CommandBufferAllocateInfo) VkError![]*Dispatchable(CommandBuffer) {
const allocator = self.host_allocator.allocator();
if (self.buffers.items.len < self.first_free_buffer_index + info.command_buffer_count) {
while (self.buffers.capacity < self.buffers.items.len + info.command_buffer_count) {
self.buffers.ensureUnusedCapacity(allocator, BUFFER_POOL_BASE_CAPACITY) catch return VkError.OutOfHostMemory;
}
for (0..info.command_buffer_count) |_| {
const cmd = try self.vtable.createCommandBuffer(self, allocator, info);
const non_dis_cmd = try Dispatchable(CommandBuffer).wrap(allocator, cmd);
self.buffers.appendAssumeCapacity(non_dis_cmd);
}
}
const bound_up = self.first_free_buffer_index + info.command_buffer_count;
const slice = self.buffers.items[self.first_free_buffer_index..bound_up];
self.first_free_buffer_index += info.command_buffer_count;
return slice;
}
pub fn freeCommandBuffers(self: *Self, cmds: []*Dispatchable(CommandBuffer)) VkError!void {
// Ugly method but it works well
var len: usize = 0;
for (cmds) |cmd| {
if (std.mem.indexOfScalar(*Dispatchable(CommandBuffer), self.buffers.items, cmd)) |i| {
const save = self.buffers.orderedRemove(i);
self.buffers.appendAssumeCapacity(save);
len += 1;
}
}
const new_first_free_buffer_index, const has_overflown = @subWithOverflow(self.first_free_buffer_index, len);
if (has_overflown == 0) {
self.first_free_buffer_index = new_first_free_buffer_index;
}
}
pub fn destroy(self: *Self, allocator: std.mem.Allocator) void {
for (self.buffers.items) |non_dis_cmd| {
non_dis_cmd.intrusiveDestroy(allocator);
}
self.buffers.deinit(allocator);
self.vtable.destroy(self, allocator);
}
pub inline fn reset(self: *Self, flags: vk.CommandPoolResetFlags) VkError!void {
try self.vtable.reset(self, flags);
if (flags.release_resources_bit) {
const allocator = self.host_allocator.allocator();
for (self.buffers.items) |non_dis_cmd| {
non_dis_cmd.intrusiveDestroy(allocator);
}
self.buffers.shrinkAndFree(allocator, BUFFER_POOL_BASE_CAPACITY);
self.buffers.clearRetainingCapacity();
}
}