/*
* This file is part of libplacebo.
*
* libplacebo is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* libplacebo is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with libplacebo. If not, see .
*/
#include "gpu.h"
#include "formats.h"
#include "glsl/spirv.h"
#ifdef PL_HAVE_UNIX
#include
#endif
// Gives us enough queries for 8 results
#define QUERY_POOL_SIZE 16
struct pl_timer_t {
VkQueryPool qpool; // even=start, odd=stop
int index_write; // next index to write to
int index_read; // next index to read from
uint64_t pending; // bitmask of queries that are still running
};
static inline uint64_t timer_bit(int index)
{
return 1llu << (index / 2);
}
static void timer_destroy_cb(pl_gpu gpu, pl_timer timer)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_assert(!timer->pending);
vk->DestroyQueryPool(vk->dev, timer->qpool, PL_VK_ALLOC);
pl_free(timer);
}
static pl_timer vk_timer_create(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_timer timer = pl_alloc_ptr(NULL, timer);
*timer = (struct pl_timer_t) {0};
struct VkQueryPoolCreateInfo qinfo = {
.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO,
.queryType = VK_QUERY_TYPE_TIMESTAMP,
.queryCount = QUERY_POOL_SIZE,
};
VK(vk->CreateQueryPool(vk->dev, &qinfo, PL_VK_ALLOC, &timer->qpool));
return timer;
error:
timer_destroy_cb(gpu, timer);
return NULL;
}
static void vk_timer_destroy(pl_gpu gpu, pl_timer timer)
{
vk_gpu_idle_callback(gpu, (vk_cb) timer_destroy_cb, gpu, timer);
}
static uint64_t vk_timer_query(pl_gpu gpu, pl_timer timer)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
if (timer->index_read == timer->index_write)
return 0; // no more unprocessed results
vk_poll_commands(vk, 0);
if (timer->pending & timer_bit(timer->index_read))
return 0; // still waiting for results
VkResult res;
uint64_t ts[2] = {0};
res = vk->GetQueryPoolResults(vk->dev, timer->qpool, timer->index_read, 2,
sizeof(ts), &ts[0], sizeof(uint64_t),
VK_QUERY_RESULT_64_BIT);
switch (res) {
case VK_SUCCESS:
timer->index_read = (timer->index_read + 2) % QUERY_POOL_SIZE;
return (ts[1] - ts[0]) * vk->props.limits.timestampPeriod;
case VK_NOT_READY:
return 0;
default:
PL_VK_ASSERT(res, "Retrieving query pool results");
}
error:
return 0;
}
static void timer_begin(pl_gpu gpu, struct vk_cmd *cmd, pl_timer timer)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
if (!timer)
return;
if (!cmd->pool->props.timestampValidBits) {
PL_TRACE(gpu, "QF %d does not support timestamp queries", cmd->pool->qf);
return;
}
vk_poll_commands(vk, 0);
if (timer->pending & timer_bit(timer->index_write))
return; // next query is still running, skip this timer
VkQueueFlags reset_flags = VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT;
if (cmd->pool->props.queueFlags & reset_flags) {
// Use direct command buffer resets
vk->CmdResetQueryPool(cmd->buf, timer->qpool, timer->index_write, 2);
} else {
// Use host query reset
vk->ResetQueryPool(vk->dev, timer->qpool, timer->index_write, 2);
}
vk->CmdWriteTimestamp(cmd->buf, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT,
timer->qpool, timer->index_write);
p->cmd_timer = timer;
}
static inline bool supports_marks(struct vk_cmd *cmd) {
// Spec says debug markers are only available on graphics/compute queues
VkQueueFlags flags = cmd->pool->props.queueFlags;
return flags & (VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT);
}
struct vk_cmd *_begin_cmd(pl_gpu gpu, enum queue_type type, const char *label,
pl_timer timer)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_mutex_lock(&p->recording);
struct vk_cmdpool *pool;
switch (type) {
case ANY: pool = p->cmd ? p->cmd->pool : vk->pool_graphics; break;
case GRAPHICS: pool = vk->pool_graphics; break;
case COMPUTE: pool = vk->pool_compute; break;
case TRANSFER: pool = vk->pool_transfer; break;
default: pl_unreachable();
}
if (!p->cmd || p->cmd->pool != pool) {
vk_cmd_submit(&p->cmd);
p->cmd = vk_cmd_begin(pool, label);
if (!p->cmd) {
pl_mutex_unlock(&p->recording);
return NULL;
}
}
if (vk->CmdBeginDebugUtilsLabelEXT && supports_marks(p->cmd)) {
vk->CmdBeginDebugUtilsLabelEXT(p->cmd->buf, &(VkDebugUtilsLabelEXT) {
.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT,
.pLabelName = label,
});
}
timer_begin(gpu, p->cmd, timer);
return p->cmd;
}
static void timer_end_cb(void *ptimer, void *pindex)
{
pl_timer timer = ptimer;
int index = (uintptr_t) pindex;
timer->pending &= ~timer_bit(index);
}
bool _end_cmd(pl_gpu gpu, struct vk_cmd **pcmd, bool submit)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
bool ret = true;
if (!pcmd) {
if (submit) {
pl_mutex_lock(&p->recording);
ret = vk_cmd_submit(&p->cmd);
pl_mutex_unlock(&p->recording);
}
return ret;
}
struct vk_cmd *cmd = *pcmd;
pl_assert(p->cmd == cmd);
if (p->cmd_timer) {
pl_timer timer = p->cmd_timer;
vk->CmdWriteTimestamp(cmd->buf, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT,
timer->qpool, timer->index_write + 1);
timer->pending |= timer_bit(timer->index_write);
vk_cmd_callback(cmd, (vk_cb) timer_end_cb, timer,
(void *) (uintptr_t) timer->index_write);
timer->index_write = (timer->index_write + 2) % QUERY_POOL_SIZE;
if (timer->index_write == timer->index_read) {
// forcibly drop the least recent result to make space
timer->index_read = (timer->index_read + 2) % QUERY_POOL_SIZE;
}
p->cmd_timer = NULL;
}
if (vk->CmdEndDebugUtilsLabelEXT && supports_marks(cmd))
vk->CmdEndDebugUtilsLabelEXT(cmd->buf);
if (submit)
ret = vk_cmd_submit(&p->cmd);
pl_mutex_unlock(&p->recording);
return ret;
}
void vk_gpu_idle_callback(pl_gpu gpu, vk_cb cb, const void *priv, const void *arg)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_mutex_lock(&p->recording);
if (p->cmd) {
vk_cmd_callback(p->cmd, cb, priv, arg);
} else {
vk_dev_callback(vk, cb, priv, arg);
}
pl_mutex_unlock(&p->recording);
}
static void vk_gpu_destroy(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
vk_cmd_submit(&p->cmd);
vk_wait_idle(vk);
for (enum pl_tex_sample_mode s = 0; s < PL_TEX_SAMPLE_MODE_COUNT; s++) {
for (enum pl_tex_address_mode a = 0; a < PL_TEX_ADDRESS_MODE_COUNT; a++)
vk->DestroySampler(vk->dev, p->samplers[s][a], PL_VK_ALLOC);
}
pl_spirv_destroy(&p->spirv);
pl_mutex_destroy(&p->recording);
pl_free((void *) gpu);
}
pl_vulkan pl_vulkan_get(pl_gpu gpu)
{
const struct pl_gpu_fns *impl = PL_PRIV(gpu);
if (impl->destroy == vk_gpu_destroy) {
struct pl_vk *p = (struct pl_vk *) impl;
return p->vk->vulkan;
}
return NULL;
}
static pl_handle_caps vk_sync_handle_caps(struct vk_ctx *vk)
{
pl_handle_caps caps = 0;
for (int i = 0; vk_sync_handle_list[i]; i++) {
enum pl_handle_type type = vk_sync_handle_list[i];
VkPhysicalDeviceExternalSemaphoreInfo info = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_EXTERNAL_SEMAPHORE_INFO_KHR,
.handleType = vk_sync_handle_type(type),
};
VkExternalSemaphoreProperties props = {
.sType = VK_STRUCTURE_TYPE_EXTERNAL_SEMAPHORE_PROPERTIES_KHR,
};
vk->GetPhysicalDeviceExternalSemaphoreProperties(vk->physd, &info, &props);
VkExternalSemaphoreFeatureFlags flags = props.externalSemaphoreFeatures;
if ((props.compatibleHandleTypes & info.handleType) &&
(flags & VK_EXTERNAL_SEMAPHORE_FEATURE_EXPORTABLE_BIT_KHR))
{
caps |= type;
}
}
return caps;
}
static pl_handle_caps vk_tex_handle_caps(struct vk_ctx *vk, bool import)
{
pl_handle_caps caps = 0;
for (int i = 0; vk_mem_handle_list[i]; i++) {
enum pl_handle_type handle_type = vk_mem_handle_list[i];
if (handle_type == PL_HANDLE_DMA_BUF && !vk->GetImageDrmFormatModifierPropertiesEXT) {
PL_DEBUG(vk, "Tex caps for %s (0x%x) unsupported: no DRM modifiers",
vk_handle_name(vk_mem_handle_type(PL_HANDLE_DMA_BUF)),
(unsigned int) PL_HANDLE_DMA_BUF);
continue;
}
// Query whether creation of a "basic" dummy texture would work
VkPhysicalDeviceImageDrmFormatModifierInfoEXT drm_pinfo = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_DRM_FORMAT_MODIFIER_INFO_EXT,
.drmFormatModifier = DRM_FORMAT_MOD_LINEAR,
.sharingMode = VK_SHARING_MODE_EXCLUSIVE,
};
VkPhysicalDeviceExternalImageFormatInfoKHR ext_pinfo = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_EXTERNAL_IMAGE_FORMAT_INFO_KHR,
.handleType = vk_mem_handle_type(handle_type),
};
VkPhysicalDeviceImageFormatInfo2KHR pinfo = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_FORMAT_INFO_2_KHR,
.pNext = &ext_pinfo,
.format = VK_FORMAT_R8_UNORM,
.type = VK_IMAGE_TYPE_2D,
.tiling = VK_IMAGE_TILING_OPTIMAL,
.usage = VK_IMAGE_USAGE_TRANSFER_DST_BIT,
};
if (handle_type == PL_HANDLE_DMA_BUF) {
vk_link_struct(&pinfo, &drm_pinfo);
pinfo.tiling = VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT;
}
VkExternalImageFormatPropertiesKHR ext_props = {
.sType = VK_STRUCTURE_TYPE_EXTERNAL_IMAGE_FORMAT_PROPERTIES_KHR,
};
VkImageFormatProperties2KHR props = {
.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_PROPERTIES_2_KHR,
.pNext = &ext_props,
};
VkResult res;
res = vk->GetPhysicalDeviceImageFormatProperties2KHR(vk->physd, &pinfo, &props);
if (res != VK_SUCCESS) {
PL_DEBUG(vk, "Tex caps for %s (0x%x) unsupported: %s",
vk_handle_name(ext_pinfo.handleType),
(unsigned int) handle_type,
vk_res_str(res));
continue;
}
if (vk_external_mem_check(vk, &ext_props.externalMemoryProperties,
handle_type, import))
{
caps |= handle_type;
}
}
#ifdef VK_EXT_metal_objects
if (vk->ExportMetalObjectsEXT && import)
caps |= PL_HANDLE_MTL_TEX | PL_HANDLE_IOSURFACE;
#endif
return caps;
}
static const VkFilter filters[PL_TEX_SAMPLE_MODE_COUNT] = {
[PL_TEX_SAMPLE_NEAREST] = VK_FILTER_NEAREST,
[PL_TEX_SAMPLE_LINEAR] = VK_FILTER_LINEAR,
};
static inline struct pl_spirv_version get_spirv_version(const struct vk_ctx *vk)
{
if (vk->api_ver >= VK_API_VERSION_1_3) {
const VkPhysicalDeviceMaintenance4Features *device_maintenance4;
device_maintenance4 = vk_find_struct(&vk->features,
VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MAINTENANCE_4_FEATURES);
if (device_maintenance4 && device_maintenance4->maintenance4) {
return (struct pl_spirv_version) {
.env_version = VK_API_VERSION_1_3,
.spv_version = PL_SPV_VERSION(1, 6),
};
}
}
pl_assert(vk->api_ver >= VK_API_VERSION_1_2);
return (struct pl_spirv_version) {
.env_version = VK_API_VERSION_1_2,
.spv_version = PL_SPV_VERSION(1, 5),
};
}
static const struct pl_gpu_fns pl_fns_vk;
pl_gpu pl_gpu_create_vk(struct vk_ctx *vk)
{
pl_assert(vk->dev);
struct pl_gpu_t *gpu = pl_zalloc_obj(NULL, gpu, struct pl_vk);
gpu->log = vk->log;
struct pl_vk *p = PL_PRIV(gpu);
pl_mutex_init(&p->recording);
p->vk = vk;
p->impl = pl_fns_vk;
p->spirv = pl_spirv_create(vk->log, get_spirv_version(vk));
if (!p->spirv)
goto error;
// Query all device properties
VkPhysicalDevicePCIBusInfoPropertiesEXT pci_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PCI_BUS_INFO_PROPERTIES_EXT,
};
VkPhysicalDeviceIDPropertiesKHR id_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_ID_PROPERTIES_KHR,
.pNext = &pci_props,
};
VkPhysicalDevicePushDescriptorPropertiesKHR pushd_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PUSH_DESCRIPTOR_PROPERTIES_KHR,
.pNext = &id_props,
};
VkPhysicalDeviceSubgroupProperties group_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_SUBGROUP_PROPERTIES,
.pNext = &pushd_props,
};
VkPhysicalDeviceExternalMemoryHostPropertiesEXT host_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_EXTERNAL_MEMORY_HOST_PROPERTIES_EXT,
.pNext = &group_props,
};
VkPhysicalDeviceProperties2KHR props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2_KHR,
.pNext = &host_props,
};
bool is_portability = false;
#ifdef VK_KHR_portability_subset
VkPhysicalDevicePortabilitySubsetPropertiesKHR port_props = {
.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PORTABILITY_SUBSET_PROPERTIES_KHR,
.minVertexInputBindingStrideAlignment = 1,
};
for (int i = 0; i < vk->exts.num; i++) {
if (!strcmp(vk->exts.elem[i], VK_KHR_PORTABILITY_SUBSET_EXTENSION_NAME)) {
vk_link_struct(&props, &port_props);
is_portability = true;
break;
}
}
#endif
vk->GetPhysicalDeviceProperties2(vk->physd, &props);
VkPhysicalDeviceLimits limits = props.properties.limits;
// Determine GLSL features and limits
gpu->glsl = (struct pl_glsl_version) {
.version = 450,
.vulkan = true,
.compute = true,
.max_shmem_size = limits.maxComputeSharedMemorySize,
.max_group_threads = limits.maxComputeWorkGroupInvocations,
.max_group_size = {
limits.maxComputeWorkGroupSize[0],
limits.maxComputeWorkGroupSize[1],
limits.maxComputeWorkGroupSize[2],
},
};
VkShaderStageFlags req_stages = VK_SHADER_STAGE_FRAGMENT_BIT |
VK_SHADER_STAGE_COMPUTE_BIT;
VkSubgroupFeatureFlags req_flags = VK_SUBGROUP_FEATURE_BASIC_BIT |
VK_SUBGROUP_FEATURE_VOTE_BIT |
VK_SUBGROUP_FEATURE_ARITHMETIC_BIT |
VK_SUBGROUP_FEATURE_BALLOT_BIT |
VK_SUBGROUP_FEATURE_SHUFFLE_BIT;
if ((group_props.supportedStages & req_stages) == req_stages &&
(group_props.supportedOperations & req_flags) == req_flags)
{
gpu->glsl.subgroup_size = group_props.subgroupSize;
}
if (vk->features.features.shaderImageGatherExtended) {
gpu->glsl.min_gather_offset = limits.minTexelGatherOffset;
gpu->glsl.max_gather_offset = limits.maxTexelGatherOffset;
}
const size_t max_size = vk_malloc_avail(vk->ma, 0);
gpu->limits = (struct pl_gpu_limits) {
// pl_gpu
.thread_safe = true,
.callbacks = true,
// pl_buf
.max_buf_size = max_size,
.max_ubo_size = PL_MIN(limits.maxUniformBufferRange, max_size),
.max_ssbo_size = PL_MIN(limits.maxStorageBufferRange, max_size),
.max_vbo_size = vk_malloc_avail(vk->ma, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT),
.max_mapped_size = vk_malloc_avail(vk->ma, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT),
.max_buffer_texels = PL_MIN(limits.maxTexelBufferElements, max_size),
.align_host_ptr = host_props.minImportedHostPointerAlignment,
.host_cached = vk_malloc_avail(vk->ma, VK_MEMORY_PROPERTY_HOST_CACHED_BIT),
// pl_tex
.max_tex_1d_dim = limits.maxImageDimension1D,
.max_tex_2d_dim = limits.maxImageDimension2D,
.max_tex_3d_dim = limits.maxImageDimension3D,
.blittable_1d_3d = true,
.buf_transfer = true,
.align_tex_xfer_pitch = limits.optimalBufferCopyRowPitchAlignment,
.align_tex_xfer_offset = pl_lcm(limits.optimalBufferCopyOffsetAlignment, 4),
// pl_pass
.max_variable_comps = 0, // vulkan doesn't support these at all
.max_constants = SIZE_MAX,
.array_size_constants = !is_portability,
.max_pushc_size = limits.maxPushConstantsSize,
#ifdef VK_KHR_portability_subset
.align_vertex_stride = port_props.minVertexInputBindingStrideAlignment,
#else
.align_vertex_stride = 1,
#endif
.max_dispatch = {
limits.maxComputeWorkGroupCount[0],
limits.maxComputeWorkGroupCount[1],
limits.maxComputeWorkGroupCount[2],
},
.fragment_queues = vk->pool_graphics->num_queues,
.compute_queues = vk->pool_compute->num_queues,
};
gpu->export_caps.buf = vk_malloc_handle_caps(vk->ma, false);
gpu->import_caps.buf = vk_malloc_handle_caps(vk->ma, true);
gpu->export_caps.tex = vk_tex_handle_caps(vk, false);
gpu->import_caps.tex = vk_tex_handle_caps(vk, true);
gpu->export_caps.sync = vk_sync_handle_caps(vk);
gpu->import_caps.sync = 0; // Not supported yet
if (pl_gpu_supports_interop(gpu)) {
pl_static_assert(sizeof(gpu->uuid) == VK_UUID_SIZE);
memcpy(gpu->uuid, id_props.deviceUUID, sizeof(gpu->uuid));
gpu->pci.domain = pci_props.pciDomain;
gpu->pci.bus = pci_props.pciBus;
gpu->pci.device = pci_props.pciDevice;
gpu->pci.function = pci_props.pciFunction;
}
if (vk->CmdPushDescriptorSetKHR)
p->max_push_descriptors = pushd_props.maxPushDescriptors;
vk_setup_formats(gpu);
// Compute the correct minimum texture alignment
p->min_texel_alignment = 1;
for (int i = 0; i < gpu->num_formats; i++) {
if (gpu->formats[i]->emulated || gpu->formats[i]->opaque)
continue;
size_t texel_size = gpu->formats[i]->texel_size;
p->min_texel_alignment = pl_lcm(p->min_texel_alignment, texel_size);
}
PL_DEBUG(gpu, "Minimum texel alignment: %zu", p->min_texel_alignment);
// Initialize the samplers
for (enum pl_tex_sample_mode s = 0; s < PL_TEX_SAMPLE_MODE_COUNT; s++) {
for (enum pl_tex_address_mode a = 0; a < PL_TEX_ADDRESS_MODE_COUNT; a++) {
static const VkSamplerAddressMode modes[PL_TEX_ADDRESS_MODE_COUNT] = {
[PL_TEX_ADDRESS_CLAMP] = VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE,
[PL_TEX_ADDRESS_REPEAT] = VK_SAMPLER_ADDRESS_MODE_REPEAT,
[PL_TEX_ADDRESS_MIRROR] = VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT,
};
VkSamplerCreateInfo sinfo = {
.sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO,
.magFilter = filters[s],
.minFilter = filters[s],
.addressModeU = modes[a],
.addressModeV = modes[a],
.addressModeW = modes[a],
.maxAnisotropy = 1.0,
};
VK(vk->CreateSampler(vk->dev, &sinfo, PL_VK_ALLOC, &p->samplers[s][a]));
}
}
return pl_gpu_finalize(gpu);
error:
vk_gpu_destroy(gpu);
return NULL;
}
static void vk_sync_destroy(pl_gpu gpu, pl_sync sync)
{
if (!sync)
return;
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
struct pl_sync_vk *sync_vk = PL_PRIV(sync);
#ifdef PL_HAVE_UNIX
if (sync->handle_type == PL_HANDLE_FD) {
if (sync->wait_handle.fd > -1)
close(sync->wait_handle.fd);
if (sync->signal_handle.fd > -1)
close(sync->signal_handle.fd);
}
#endif
#ifdef PL_HAVE_WIN32
if (sync->handle_type == PL_HANDLE_WIN32) {
if (sync->wait_handle.handle != NULL)
CloseHandle(sync->wait_handle.handle);
if (sync->signal_handle.handle != NULL)
CloseHandle(sync->signal_handle.handle);
}
// PL_HANDLE_WIN32_KMT is just an identifier. It doesn't get closed.
#endif
vk->DestroySemaphore(vk->dev, sync_vk->wait, PL_VK_ALLOC);
vk->DestroySemaphore(vk->dev, sync_vk->signal, PL_VK_ALLOC);
pl_free((void *) sync);
}
void vk_sync_deref(pl_gpu gpu, pl_sync sync)
{
if (!sync)
return;
struct pl_sync_vk *sync_vk = PL_PRIV(sync);
if (pl_rc_deref(&sync_vk->rc))
vk_sync_destroy(gpu, sync);
}
static pl_sync vk_sync_create(pl_gpu gpu, enum pl_handle_type handle_type)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
struct pl_sync_t *sync = pl_zalloc_obj(NULL, sync, struct pl_sync_vk);
sync->handle_type = handle_type;
struct pl_sync_vk *sync_vk = PL_PRIV(sync);
pl_rc_init(&sync_vk->rc);
VkExportSemaphoreCreateInfoKHR einfo = {
.sType = VK_STRUCTURE_TYPE_EXPORT_SEMAPHORE_CREATE_INFO_KHR,
.handleTypes = vk_sync_handle_type(handle_type),
};
switch (handle_type) {
case PL_HANDLE_FD:
sync->wait_handle.fd = -1;
sync->signal_handle.fd = -1;
break;
case PL_HANDLE_WIN32:
case PL_HANDLE_WIN32_KMT:
sync->wait_handle.handle = NULL;
sync->signal_handle.handle = NULL;
break;
case PL_HANDLE_DMA_BUF:
case PL_HANDLE_HOST_PTR:
case PL_HANDLE_MTL_TEX:
case PL_HANDLE_IOSURFACE:
pl_unreachable();
}
const VkSemaphoreCreateInfo sinfo = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO,
.pNext = &einfo,
};
VK(vk->CreateSemaphore(vk->dev, &sinfo, PL_VK_ALLOC, &sync_vk->wait));
VK(vk->CreateSemaphore(vk->dev, &sinfo, PL_VK_ALLOC, &sync_vk->signal));
PL_VK_NAME(SEMAPHORE, sync_vk->wait, "sync wait");
PL_VK_NAME(SEMAPHORE, sync_vk->signal, "sync signal");
#ifdef PL_HAVE_UNIX
if (handle_type == PL_HANDLE_FD) {
VkSemaphoreGetFdInfoKHR finfo = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_GET_FD_INFO_KHR,
.semaphore = sync_vk->wait,
.handleType = einfo.handleTypes,
};
VK(vk->GetSemaphoreFdKHR(vk->dev, &finfo, &sync->wait_handle.fd));
finfo.semaphore = sync_vk->signal;
VK(vk->GetSemaphoreFdKHR(vk->dev, &finfo, &sync->signal_handle.fd));
}
#endif
#ifdef PL_HAVE_WIN32
if (handle_type == PL_HANDLE_WIN32 ||
handle_type == PL_HANDLE_WIN32_KMT)
{
VkSemaphoreGetWin32HandleInfoKHR handle_info = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_GET_WIN32_HANDLE_INFO_KHR,
.semaphore = sync_vk->wait,
.handleType = einfo.handleTypes,
};
VK(vk->GetSemaphoreWin32HandleKHR(vk->dev, &handle_info,
&sync->wait_handle.handle));
handle_info.semaphore = sync_vk->signal;
VK(vk->GetSemaphoreWin32HandleKHR(vk->dev, &handle_info,
&sync->signal_handle.handle));
}
#endif
return sync;
error:
vk_sync_destroy(gpu, sync);
return NULL;
}
void pl_vulkan_sem_destroy(pl_gpu gpu, VkSemaphore *semaphore)
{
VkSemaphore sem = *semaphore;
if (!sem)
return;
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
vk->DestroySemaphore(vk->dev, sem, PL_VK_ALLOC);
*semaphore = VK_NULL_HANDLE;
}
VkSemaphore pl_vulkan_sem_create(pl_gpu gpu, const struct pl_vulkan_sem_params *params)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_assert(PL_ISPOT(params->export_handle));
if ((params->export_handle & gpu->export_caps.sync) != params->export_handle) {
PL_ERR(gpu, "Invalid handle type 0x%"PRIx64" specified for "
"`pl_vulkan_sem_create`!", (uint64_t) params->export_handle);
return VK_NULL_HANDLE;
}
switch (params->export_handle) {
case PL_HANDLE_FD:
params->out_handle->fd = -1;
break;
case PL_HANDLE_WIN32:
case PL_HANDLE_WIN32_KMT:
params->out_handle->handle = NULL;
break;
case PL_HANDLE_DMA_BUF:
case PL_HANDLE_HOST_PTR:
case PL_HANDLE_MTL_TEX:
case PL_HANDLE_IOSURFACE:
pl_unreachable();
}
const VkExportSemaphoreCreateInfoKHR einfo = {
.sType = VK_STRUCTURE_TYPE_EXPORT_SEMAPHORE_CREATE_INFO_KHR,
.handleTypes = vk_sync_handle_type(params->export_handle),
};
const VkSemaphoreTypeCreateInfo stinfo = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_TYPE_CREATE_INFO,
.pNext = params->export_handle ? &einfo : NULL,
.semaphoreType = params->type,
.initialValue = params->initial_value,
};
const VkSemaphoreCreateInfo sinfo = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO,
.pNext = &stinfo,
};
VkSemaphore sem = VK_NULL_HANDLE;
VK(vk->CreateSemaphore(vk->dev, &sinfo, PL_VK_ALLOC, &sem));
PL_VK_NAME(SEMAPHORE, sem, PL_DEF(params->debug_tag, "pl_vulkan_sem"));
#ifdef PL_HAVE_UNIX
if (params->export_handle == PL_HANDLE_FD) {
VkSemaphoreGetFdInfoKHR finfo = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_GET_FD_INFO_KHR,
.handleType = einfo.handleTypes,
.semaphore = sem,
};
VK(vk->GetSemaphoreFdKHR(vk->dev, &finfo, ¶ms->out_handle->fd));
}
#endif
#ifdef PL_HAVE_WIN32
if (params->export_handle == PL_HANDLE_WIN32 ||
params->export_handle == PL_HANDLE_WIN32_KMT)
{
VkSemaphoreGetWin32HandleInfoKHR handle_info = {
.sType = VK_STRUCTURE_TYPE_SEMAPHORE_GET_WIN32_HANDLE_INFO_KHR,
.handleType = einfo.handleTypes,
.semaphore = sem,
};
VK(vk->GetSemaphoreWin32HandleKHR(vk->dev, &handle_info,
¶ms->out_handle->handle));
}
#endif
return sem;
error:
#ifdef PL_HAVE_UNIX
if (params->export_handle == PL_HANDLE_FD) {
if (params->out_handle->fd > -1)
close(params->out_handle->fd);
}
#endif
#ifdef PL_HAVE_WIN32
if (params->export_handle == PL_HANDLE_WIN32) {
if (params->out_handle->handle != NULL)
CloseHandle(params->out_handle->handle);
}
// PL_HANDLE_WIN32_KMT is just an identifier. It doesn't get closed.
#endif
vk->DestroySemaphore(vk->dev, sem, PL_VK_ALLOC);
return VK_NULL_HANDLE;
}
static void vk_gpu_flush(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
CMD_SUBMIT(NULL);
vk_rotate_queues(vk);
vk_malloc_garbage_collect(vk->ma);
}
static void vk_gpu_finish(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
CMD_SUBMIT(NULL);
vk_wait_idle(vk);
}
static bool vk_gpu_is_failed(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
return vk->failed;
}
struct vk_cmd *pl_vk_steal_cmd(pl_gpu gpu)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
pl_mutex_lock(&p->recording);
struct vk_cmd *cmd = p->cmd;
p->cmd = NULL;
pl_mutex_unlock(&p->recording);
struct vk_cmdpool *pool = vk->pool_graphics;
if (!cmd || cmd->pool != pool) {
vk_cmd_submit(&cmd);
cmd = vk_cmd_begin(pool, NULL);
}
return cmd;
}
void pl_vk_print_heap(pl_gpu gpu, enum pl_log_level lev)
{
struct pl_vk *p = PL_PRIV(gpu);
struct vk_ctx *vk = p->vk;
vk_malloc_print_stats(vk->ma, lev);
}
static const struct pl_gpu_fns pl_fns_vk = {
.destroy = vk_gpu_destroy,
.tex_create = vk_tex_create,
.tex_destroy = vk_tex_deref,
.tex_invalidate = vk_tex_invalidate,
.tex_clear_ex = vk_tex_clear_ex,
.tex_blit = vk_tex_blit,
.tex_upload = vk_tex_upload,
.tex_download = vk_tex_download,
.tex_poll = vk_tex_poll,
.tex_export = vk_tex_export,
.buf_create = vk_buf_create,
.buf_destroy = vk_buf_deref,
.buf_write = vk_buf_write,
.buf_read = vk_buf_read,
.buf_copy = vk_buf_copy,
.buf_export = vk_buf_export,
.buf_poll = vk_buf_poll,
.desc_namespace = vk_desc_namespace,
.pass_create = vk_pass_create,
.pass_destroy = vk_pass_destroy,
.pass_run = vk_pass_run,
.sync_create = vk_sync_create,
.sync_destroy = vk_sync_deref,
.timer_create = vk_timer_create,
.timer_destroy = vk_timer_destroy,
.timer_query = vk_timer_query,
.gpu_flush = vk_gpu_flush,
.gpu_finish = vk_gpu_finish,
.gpu_is_failed = vk_gpu_is_failed,
};