aboutsummaryrefslogtreecommitdiff
path: root/render/vulkan
diff options
context:
space:
mode:
authorManuel Stoeckl <code@mstoeckl.com>2022-11-19 08:44:56 -0500
committerManuel Stoeckl <code@mstoeckl.com>2022-11-21 17:55:46 -0500
commitb97ef593938e94f2c0c9c1d08f0b83dbb9a8c1c6 (patch)
tree0b9aef0752669205e7ac2d7eb546dc6fe17963e4 /render/vulkan
parent4ff46e6cf9463e594605928feeb7c55cf323b5e7 (diff)
render/vulkan: update shm texture data in one batch
Diffstat (limited to 'render/vulkan')
-rw-r--r--render/vulkan/texture.c150
1 files changed, 82 insertions, 68 deletions
diff --git a/render/vulkan/texture.c b/render/vulkan/texture.c
index 6c6e9825..45e3fec9 100644
--- a/render/vulkan/texture.c
+++ b/render/vulkan/texture.c
@@ -36,34 +36,47 @@ static VkImageAspectFlagBits mem_plane_aspect(unsigned i) {
// Will transition the texture to shaderReadOnlyOptimal layout for reading
// from fragment shader later on
static bool write_pixels(struct wlr_vk_texture *texture,
- uint32_t stride, uint32_t width, uint32_t height, uint32_t src_x,
- uint32_t src_y, uint32_t dst_x, uint32_t dst_y, const void *vdata,
+ uint32_t stride, const pixman_region32_t *region, const void *vdata,
VkImageLayout old_layout, VkPipelineStageFlags src_stage,
VkAccessFlags src_access) {
VkResult res;
struct wlr_vk_renderer *renderer = texture->renderer;
VkDevice dev = texture->renderer->dev->dev;
- // make sure assumptions are met
- assert(src_x + width <= texture->wlr_texture.width);
- assert(src_y + height <= texture->wlr_texture.height);
- assert(dst_x + width <= texture->wlr_texture.width);
- assert(dst_y + height <= texture->wlr_texture.height);
-
const struct wlr_pixel_format_info *format_info = drm_get_pixel_format_info(
texture->format->drm_format);
assert(format_info);
- // deferred upload by transfer; using staging buffer
- // calculate maximum side needed
uint32_t bsize = 0;
unsigned bytespb = format_info->bpp / 8;
- bsize += height * bytespb * width;
+
+ // deferred upload by transfer; using staging buffer
+ // calculate maximum side needed
+ int rects_len = 0;
+ const pixman_box32_t *rects = pixman_region32_rectangles(region, &rects_len);
+ for (int i = 0; i < rects_len; i++) {
+ pixman_box32_t rect = rects[i];
+ uint32_t width = rect.x2 - rect.x1;
+ uint32_t height = rect.y2 - rect.y1;
+
+ // make sure assumptions are met
+ assert((uint32_t)rect.x2 <= texture->wlr_texture.width);
+ assert((uint32_t)rect.y2 <= texture->wlr_texture.height);
+
+ bsize += height * bytespb * width;
+ }
+
+ VkBufferImageCopy *copies = calloc((size_t)rects_len, sizeof(*copies));
+ if (!copies) {
+ wlr_log(WLR_ERROR, "Failed to allocate image copy parameters");
+ return false;
+ }
// get staging buffer
struct wlr_vk_buffer_span span = vulkan_get_stage_span(renderer, bsize);
if (!span.buffer || span.alloc.size != bsize) {
wlr_log(WLR_ERROR, "Failed to retrieve staging buffer");
+ free(copies);
return false;
}
@@ -72,60 +85,72 @@ static bool write_pixels(struct wlr_vk_texture *texture,
bsize, 0, &vmap);
if (res != VK_SUCCESS) {
wlr_vk_error("vkMapMemory", res);
+ free(copies);
return false;
}
char *map = (char *)vmap;
- // record staging cb
- // will be executed before next frame
- VkCommandBuffer cb = vulkan_record_stage_cb(renderer);
- vulkan_change_layout(cb, texture->image,
- old_layout, src_stage, src_access,
- VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, VK_PIPELINE_STAGE_TRANSFER_BIT,
- VK_ACCESS_TRANSFER_WRITE_BIT);
-
// upload data
- const char *pdata = vdata; // data iterator
- uint32_t packed_stride = bytespb * width;
uint32_t buf_off = span.alloc.start + (map - (char *)vmap);
-
- // write data into staging buffer span
- pdata += stride * src_y;
- pdata += bytespb * src_x;
- if (src_x == 0 && width == texture->wlr_texture.width &&
- stride == packed_stride) {
- memcpy(map, pdata, packed_stride * height);
- map += packed_stride * height;
- } else {
- for (unsigned i = 0u; i < height; ++i) {
- memcpy(map, pdata, packed_stride);
- pdata += stride;
- map += packed_stride;
+ for (int i = 0; i < rects_len; i++) {
+ pixman_box32_t rect = rects[i];
+ uint32_t width = rect.x2 - rect.x1;
+ uint32_t height = rect.y2 - rect.y1;
+ uint32_t src_x = rect.x1;
+ uint32_t src_y = rect.y1;
+
+ uint32_t packed_stride = bytespb * width;
+
+ // write data into staging buffer span
+ const char *pdata = vdata; // data iterator
+ pdata += stride * src_y;
+ pdata += bytespb * src_x;
+ if (src_x == 0 && width == texture->wlr_texture.width &&
+ stride == packed_stride) {
+ memcpy(map, pdata, packed_stride * height);
+ map += packed_stride * height;
+ } else {
+ for (unsigned i = 0u; i < height; ++i) {
+ memcpy(map, pdata, packed_stride);
+ pdata += stride;
+ map += packed_stride;
+ }
}
+
+ copies[i] = (VkBufferImageCopy) {
+ .imageExtent.width = width,
+ .imageExtent.height = height,
+ .imageExtent.depth = 1,
+ .imageOffset.x = src_x,
+ .imageOffset.y = src_y,
+ .imageOffset.z = 0,
+ .bufferOffset = buf_off,
+ .bufferRowLength = width,
+ .bufferImageHeight = height,
+ .imageSubresource.mipLevel = 0,
+ .imageSubresource.baseArrayLayer = 0,
+ .imageSubresource.layerCount = 1,
+ .imageSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
+ };
+
+
+ buf_off += height * packed_stride;
}
assert((uint32_t)(map - (char *)vmap) == bsize);
vkUnmapMemory(dev, span.buffer->memory);
- VkBufferImageCopy copy = {
- .imageExtent.width = width,
- .imageExtent.height = height,
- .imageExtent.depth = 1,
- .imageOffset.x = dst_x,
- .imageOffset.y = dst_y,
- .imageOffset.z = 0,
- .bufferOffset = buf_off,
- .bufferRowLength = width,
- .bufferImageHeight = height,
- .imageSubresource.mipLevel = 0,
- .imageSubresource.baseArrayLayer = 0,
- .imageSubresource.layerCount = 1,
- .imageSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
- };
+ // record staging cb
+ // will be executed before next frame
+ VkCommandBuffer cb = vulkan_record_stage_cb(renderer);
+ vulkan_change_layout(cb, texture->image,
+ old_layout, src_stage, src_access,
+ VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, VK_PIPELINE_STAGE_TRANSFER_BIT,
+ VK_ACCESS_TRANSFER_WRITE_BIT);
vkCmdCopyBufferToImage(cb, span.buffer->buffer, texture->image,
- VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1, &copy);
+ VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, (uint32_t)rects_len, copies);
vulkan_change_layout(cb, texture->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, VK_PIPELINE_STAGE_TRANSFER_BIT,
VK_ACCESS_TRANSFER_WRITE_BIT,
@@ -133,6 +158,8 @@ static bool write_pixels(struct wlr_vk_texture *texture,
VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT, VK_ACCESS_SHADER_READ_BIT);
texture->last_used = renderer->frame;
+ free(copies);
+
return true;
}
@@ -155,22 +182,8 @@ static bool vulkan_texture_update_from_buffer(struct wlr_texture *wlr_texture,
goto out;
}
- int rects_len = 0;
- const pixman_box32_t *rects = pixman_region32_rectangles(damage, &rects_len);
-
- for (int i = 0; i < rects_len; i++) {
- pixman_box32_t rect = rects[i];
- uint32_t width = rect.x2 - rect.x1;
- uint32_t height = rect.y2 - rect.y1;
-
- // TODO: only map memory once
- ok = write_pixels(texture, stride, width, height, rect.x1, rect.y1,
- rect.x1, rect.y1, data, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL,
- VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, VK_ACCESS_SHADER_READ_BIT);
- if (!ok) {
- goto out;
- }
- }
+ ok = write_pixels(texture, stride, damage, data, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL,
+ VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, VK_ACCESS_SHADER_READ_BIT);
out:
wlr_buffer_end_data_ptr_access(buffer);
@@ -376,8 +389,9 @@ static struct wlr_texture *vulkan_texture_from_pixels(
vkUpdateDescriptorSets(dev, 1, &ds_write, 0, NULL);
// write data
- if (!write_pixels(texture, stride,
- width, height, 0, 0, 0, 0, data, VK_IMAGE_LAYOUT_UNDEFINED,
+ pixman_region32_t region;
+ pixman_region32_init_rect(&region, 0, 0, width, height);
+ if (!write_pixels(texture, stride, &region, data, VK_IMAGE_LAYOUT_UNDEFINED,
VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, 0)) {
goto error;
}