1/*
2 * Copyright © 2016 Dave Airlie
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 */
23
24
25#include <assert.h>
26#include <stdbool.h>
27
28#include "radv_meta.h"
29#include "radv_private.h"
30#include "nir/nir_builder.h"
31#include "sid.h"
32#include "vk_format.h"
33
34static nir_shader *
35build_resolve_compute_shader(struct radv_device *dev, bool is_integer, int samples)
36{
37	nir_builder b;
38	char name[64];
39	nir_if *outer_if = NULL;
40	const struct glsl_type *sampler_type = glsl_sampler_type(GLSL_SAMPLER_DIM_MS,
41								 false,
42								 false,
43								 GLSL_TYPE_FLOAT);
44	const struct glsl_type *img_type = glsl_sampler_type(GLSL_SAMPLER_DIM_2D,
45							     false,
46							     false,
47							     GLSL_TYPE_FLOAT);
48	snprintf(name, 64, "meta_resolve_cs-%d-%s", samples, is_integer ? "int" : "float");
49	nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_COMPUTE, NULL);
50	b.shader->info->name = ralloc_strdup(b.shader, name);
51	b.shader->info->cs.local_size[0] = 16;
52	b.shader->info->cs.local_size[1] = 16;
53	b.shader->info->cs.local_size[2] = 1;
54
55	nir_variable *input_img = nir_variable_create(b.shader, nir_var_uniform,
56						      sampler_type, "s_tex");
57	input_img->data.descriptor_set = 0;
58	input_img->data.binding = 0;
59
60	nir_variable *output_img = nir_variable_create(b.shader, nir_var_uniform,
61						       img_type, "out_img");
62	output_img->data.descriptor_set = 0;
63	output_img->data.binding = 1;
64	nir_ssa_def *invoc_id = nir_load_system_value(&b, nir_intrinsic_load_local_invocation_id, 0);
65	nir_ssa_def *wg_id = nir_load_system_value(&b, nir_intrinsic_load_work_group_id, 0);
66	nir_ssa_def *block_size = nir_imm_ivec4(&b,
67						b.shader->info->cs.local_size[0],
68						b.shader->info->cs.local_size[1],
69						b.shader->info->cs.local_size[2], 0);
70
71	nir_ssa_def *global_id = nir_iadd(&b, nir_imul(&b, wg_id, block_size), invoc_id);
72
73	nir_intrinsic_instr *src_offset = nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_push_constant);
74	src_offset->src[0] = nir_src_for_ssa(nir_imm_int(&b, 0));
75	src_offset->num_components = 2;
76	nir_ssa_dest_init(&src_offset->instr, &src_offset->dest, 2, 32, "src_offset");
77	nir_builder_instr_insert(&b, &src_offset->instr);
78
79	nir_intrinsic_instr *dst_offset = nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_push_constant);
80	dst_offset->src[0] = nir_src_for_ssa(nir_imm_int(&b, 8));
81	dst_offset->num_components = 2;
82	nir_ssa_dest_init(&dst_offset->instr, &dst_offset->dest, 2, 32, "dst_offset");
83	nir_builder_instr_insert(&b, &dst_offset->instr);
84
85	nir_ssa_def *img_coord = nir_iadd(&b, global_id, &src_offset->dest.ssa);
86	/* do a txf_ms on each sample */
87	nir_ssa_def *tmp;
88
89	nir_tex_instr *tex = nir_tex_instr_create(b.shader, 2);
90	tex->sampler_dim = GLSL_SAMPLER_DIM_MS;
91	tex->op = nir_texop_txf_ms;
92	tex->src[0].src_type = nir_tex_src_coord;
93	tex->src[0].src = nir_src_for_ssa(img_coord);
94	tex->src[1].src_type = nir_tex_src_ms_index;
95	tex->src[1].src = nir_src_for_ssa(nir_imm_int(&b, 0));
96	tex->dest_type = nir_type_float;
97	tex->is_array = false;
98	tex->coord_components = 2;
99	tex->texture = nir_deref_var_create(tex, input_img);
100	tex->sampler = NULL;
101
102	nir_ssa_dest_init(&tex->instr, &tex->dest, 4, 32, "tex");
103	nir_builder_instr_insert(&b, &tex->instr);
104
105	tmp = &tex->dest.ssa;
106	nir_variable *color =
107		nir_local_variable_create(b.impl, glsl_vec4_type(), "color");
108
109	if (!is_integer && samples > 1) {
110		nir_tex_instr *tex_all_same = nir_tex_instr_create(b.shader, 1);
111		tex_all_same->sampler_dim = GLSL_SAMPLER_DIM_MS;
112		tex_all_same->op = nir_texop_samples_identical;
113		tex_all_same->src[0].src_type = nir_tex_src_coord;
114		tex_all_same->src[0].src = nir_src_for_ssa(img_coord);
115		tex_all_same->dest_type = nir_type_float;
116		tex_all_same->is_array = false;
117		tex_all_same->coord_components = 2;
118		tex_all_same->texture = nir_deref_var_create(tex_all_same, input_img);
119		tex_all_same->sampler = NULL;
120
121		nir_ssa_dest_init(&tex_all_same->instr, &tex_all_same->dest, 1, 32, "tex");
122		nir_builder_instr_insert(&b, &tex_all_same->instr);
123
124		nir_ssa_def *all_same = nir_ine(&b, &tex_all_same->dest.ssa, nir_imm_int(&b, 0));
125		nir_if *if_stmt = nir_if_create(b.shader);
126		if_stmt->condition = nir_src_for_ssa(all_same);
127		nir_cf_node_insert(b.cursor, &if_stmt->cf_node);
128
129		b.cursor = nir_after_cf_list(&if_stmt->then_list);
130		for (int i = 1; i < samples; i++) {
131			nir_tex_instr *tex_add = nir_tex_instr_create(b.shader, 2);
132			tex_add->sampler_dim = GLSL_SAMPLER_DIM_MS;
133			tex_add->op = nir_texop_txf_ms;
134			tex_add->src[0].src_type = nir_tex_src_coord;
135			tex_add->src[0].src = nir_src_for_ssa(img_coord);
136			tex_add->src[1].src_type = nir_tex_src_ms_index;
137			tex_add->src[1].src = nir_src_for_ssa(nir_imm_int(&b, i));
138			tex_add->dest_type = nir_type_float;
139			tex_add->is_array = false;
140			tex_add->coord_components = 2;
141			tex_add->texture = nir_deref_var_create(tex_add, input_img);
142			tex_add->sampler = NULL;
143
144			nir_ssa_dest_init(&tex_add->instr, &tex_add->dest, 4, 32, "tex");
145			nir_builder_instr_insert(&b, &tex_add->instr);
146
147			tmp = nir_fadd(&b, tmp, &tex_add->dest.ssa);
148		}
149
150		tmp = nir_fdiv(&b, tmp, nir_imm_float(&b, samples));
151		nir_store_var(&b, color, tmp, 0xf);
152		b.cursor = nir_after_cf_list(&if_stmt->else_list);
153		outer_if = if_stmt;
154	}
155	nir_store_var(&b, color, &tex->dest.ssa, 0xf);
156
157	if (outer_if)
158		b.cursor = nir_after_cf_node(&outer_if->cf_node);
159
160	nir_ssa_def *newv = nir_load_var(&b, color);
161	nir_ssa_def *coord = nir_iadd(&b, global_id, &dst_offset->dest.ssa);
162	nir_intrinsic_instr *store = nir_intrinsic_instr_create(b.shader, nir_intrinsic_image_store);
163	store->src[0] = nir_src_for_ssa(coord);
164	store->src[1] = nir_src_for_ssa(nir_ssa_undef(&b, 1, 32));
165	store->src[2] = nir_src_for_ssa(newv);
166	store->variables[0] = nir_deref_var_create(store, output_img);
167	nir_builder_instr_insert(&b, &store->instr);
168	return b.shader;
169}
170
171
172static VkResult
173create_layout(struct radv_device *device)
174{
175	VkResult result;
176	/*
177	 * two descriptors one for the image being sampled
178	 * one for the buffer being written.
179	 */
180	VkDescriptorSetLayoutCreateInfo ds_create_info = {
181		.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO,
182		.bindingCount = 2,
183		.pBindings = (VkDescriptorSetLayoutBinding[]) {
184			{
185				.binding = 0,
186				.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE,
187				.descriptorCount = 1,
188				.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT,
189				.pImmutableSamplers = NULL
190			},
191			{
192				.binding = 1,
193				.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
194				.descriptorCount = 1,
195				.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT,
196				.pImmutableSamplers = NULL
197			},
198		}
199	};
200
201	result = radv_CreateDescriptorSetLayout(radv_device_to_handle(device),
202						&ds_create_info,
203						&device->meta_state.alloc,
204						&device->meta_state.resolve_compute.ds_layout);
205	if (result != VK_SUCCESS)
206		goto fail;
207
208
209	VkPipelineLayoutCreateInfo pl_create_info = {
210		.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO,
211		.setLayoutCount = 1,
212		.pSetLayouts = &device->meta_state.resolve_compute.ds_layout,
213		.pushConstantRangeCount = 1,
214		.pPushConstantRanges = &(VkPushConstantRange){VK_SHADER_STAGE_COMPUTE_BIT, 0, 16},
215	};
216
217	result = radv_CreatePipelineLayout(radv_device_to_handle(device),
218					  &pl_create_info,
219					  &device->meta_state.alloc,
220					  &device->meta_state.resolve_compute.p_layout);
221	if (result != VK_SUCCESS)
222		goto fail;
223	return VK_SUCCESS;
224fail:
225	return result;
226}
227
228static VkResult
229create_resolve_pipeline(struct radv_device *device,
230			int samples,
231			bool is_integer,
232			VkPipeline *pipeline)
233{
234	VkResult result;
235	struct radv_shader_module cs = { .nir = NULL };
236
237	cs.nir = build_resolve_compute_shader(device, is_integer, samples);
238
239	/* compute shader */
240
241	VkPipelineShaderStageCreateInfo pipeline_shader_stage = {
242		.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO,
243		.stage = VK_SHADER_STAGE_COMPUTE_BIT,
244		.module = radv_shader_module_to_handle(&cs),
245		.pName = "main",
246		.pSpecializationInfo = NULL,
247	};
248
249	VkComputePipelineCreateInfo vk_pipeline_info = {
250		.sType = VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO,
251		.stage = pipeline_shader_stage,
252		.flags = 0,
253		.layout = device->meta_state.resolve_compute.p_layout,
254	};
255
256	result = radv_CreateComputePipelines(radv_device_to_handle(device),
257					     radv_pipeline_cache_to_handle(&device->meta_state.cache),
258					     1, &vk_pipeline_info, NULL,
259					     pipeline);
260	if (result != VK_SUCCESS)
261		goto fail;
262
263	ralloc_free(cs.nir);
264	return VK_SUCCESS;
265fail:
266	ralloc_free(cs.nir);
267	return result;
268}
269
270VkResult
271radv_device_init_meta_resolve_compute_state(struct radv_device *device)
272{
273	struct radv_meta_state *state = &device->meta_state;
274	VkResult res;
275	memset(&device->meta_state.resolve_compute, 0, sizeof(device->meta_state.resolve_compute));
276
277	res = create_layout(device);
278	if (res != VK_SUCCESS)
279		return res;
280
281	for (uint32_t i = 0; i < MAX_SAMPLES_LOG2; ++i) {
282		uint32_t samples = 1 << i;
283
284		res = create_resolve_pipeline(device, samples, false,
285					      &state->resolve_compute.rc[i].pipeline);
286
287		res = create_resolve_pipeline(device, samples, true,
288					      &state->resolve_compute.rc[i].i_pipeline);
289
290	}
291
292	return res;
293}
294
295void
296radv_device_finish_meta_resolve_compute_state(struct radv_device *device)
297{
298	struct radv_meta_state *state = &device->meta_state;
299	for (uint32_t i = 0; i < MAX_SAMPLES_LOG2; ++i) {
300		radv_DestroyPipeline(radv_device_to_handle(device),
301				     state->resolve_compute.rc[i].pipeline,
302				     &state->alloc);
303
304		radv_DestroyPipeline(radv_device_to_handle(device),
305				     state->resolve_compute.rc[i].i_pipeline,
306				     &state->alloc);
307	}
308
309	radv_DestroyDescriptorSetLayout(radv_device_to_handle(device),
310					state->resolve_compute.ds_layout,
311					&state->alloc);
312	radv_DestroyPipelineLayout(radv_device_to_handle(device),
313				   state->resolve_compute.p_layout,
314				   &state->alloc);
315}
316
317void radv_meta_resolve_compute_image(struct radv_cmd_buffer *cmd_buffer,
318				     struct radv_image *src_image,
319				     VkImageLayout src_image_layout,
320				     struct radv_image *dest_image,
321				     VkImageLayout dest_image_layout,
322				     uint32_t region_count,
323				     const VkImageResolve *regions)
324{
325	struct radv_device *device = cmd_buffer->device;
326	struct radv_meta_saved_compute_state saved_state;
327	const uint32_t samples = src_image->samples;
328	const uint32_t samples_log2 = ffs(samples) - 1;
329	radv_meta_save_compute(&saved_state, cmd_buffer, 16);
330
331	for (uint32_t r = 0; r < region_count; ++r) {
332		const VkImageResolve *region = &regions[r];
333
334		assert(region->srcSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT);
335		assert(region->dstSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT);
336		assert(region->srcSubresource.layerCount == region->dstSubresource.layerCount);
337
338		const uint32_t src_base_layer =
339			radv_meta_get_iview_layer(src_image, &region->srcSubresource,
340						  &region->srcOffset);
341
342		const uint32_t dest_base_layer =
343			radv_meta_get_iview_layer(dest_image, &region->dstSubresource,
344						  &region->dstOffset);
345
346		const struct VkExtent3D extent =
347			radv_sanitize_image_extent(src_image->type, region->extent);
348		const struct VkOffset3D srcOffset =
349			radv_sanitize_image_offset(src_image->type, region->srcOffset);
350		const struct VkOffset3D dstOffset =
351			radv_sanitize_image_offset(dest_image->type, region->dstOffset);
352
353		for (uint32_t layer = 0; layer < region->srcSubresource.layerCount;
354		     ++layer) {
355
356			struct radv_image_view src_iview;
357			VkDescriptorSet set;
358			radv_image_view_init(&src_iview, cmd_buffer->device,
359					     &(VkImageViewCreateInfo) {
360						     .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
361							     .image = radv_image_to_handle(src_image),
362							     .viewType = radv_meta_get_view_type(src_image),
363							     .format = src_image->vk_format,
364							     .subresourceRange = {
365							     .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
366							     .baseMipLevel = region->srcSubresource.mipLevel,
367							     .levelCount = 1,
368							     .baseArrayLayer = src_base_layer + layer,
369							     .layerCount = 1,
370						     },
371					     },
372					     cmd_buffer, VK_IMAGE_USAGE_SAMPLED_BIT);
373
374			struct radv_image_view dest_iview;
375			radv_image_view_init(&dest_iview, cmd_buffer->device,
376					     &(VkImageViewCreateInfo) {
377						     .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
378							     .image = radv_image_to_handle(dest_image),
379							     .viewType = radv_meta_get_view_type(dest_image),
380							     .format = dest_image->vk_format,
381							     .subresourceRange = {
382							     .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
383							     .baseMipLevel = region->dstSubresource.mipLevel,
384							     .levelCount = 1,
385							     .baseArrayLayer = dest_base_layer + layer,
386							     .layerCount = 1,
387						     },
388							     },
389					     cmd_buffer, VK_IMAGE_USAGE_STORAGE_BIT);
390
391
392			radv_temp_descriptor_set_create(device, cmd_buffer,
393							device->meta_state.resolve_compute.ds_layout,
394							&set);
395
396			radv_UpdateDescriptorSets(radv_device_to_handle(device),
397						  2, /* writeCount */
398						  (VkWriteDescriptorSet[]) {
399						  {
400						  .sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET,
401						  .dstSet = set,
402						  .dstBinding = 0,
403						  .dstArrayElement = 0,
404						  .descriptorCount = 1,
405						  .descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE,
406						  .pImageInfo = (VkDescriptorImageInfo[]) {
407							  {
408								  .sampler = VK_NULL_HANDLE,
409								  .imageView = radv_image_view_to_handle(&src_iview),
410								  .imageLayout = VK_IMAGE_LAYOUT_GENERAL,
411							  },
412						  }
413					  },
414					  {
415						  .sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET,
416						  .dstSet = set,
417						  .dstBinding = 1,
418						  .dstArrayElement = 0,
419						  .descriptorCount = 1,
420						  .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
421						  .pImageInfo = (VkDescriptorImageInfo[]) {
422							  {
423								  .sampler = VK_NULL_HANDLE,
424								  .imageView = radv_image_view_to_handle(&dest_iview),
425								  .imageLayout = VK_IMAGE_LAYOUT_GENERAL,
426							  },
427						  }
428					  }
429				  }, 0, NULL);
430
431			radv_CmdBindDescriptorSets(radv_cmd_buffer_to_handle(cmd_buffer),
432						   VK_PIPELINE_BIND_POINT_COMPUTE,
433						   device->meta_state.resolve_compute.p_layout, 0, 1,
434						   &set, 0, NULL);
435
436			VkPipeline pipeline;
437			if (vk_format_is_int(src_image->vk_format))
438				pipeline = device->meta_state.resolve_compute.rc[samples_log2].i_pipeline;
439			else
440				pipeline = device->meta_state.resolve_compute.rc[samples_log2].pipeline;
441			if (cmd_buffer->state.compute_pipeline != radv_pipeline_from_handle(pipeline)) {
442				radv_CmdBindPipeline(radv_cmd_buffer_to_handle(cmd_buffer),
443						     VK_PIPELINE_BIND_POINT_COMPUTE, pipeline);
444			}
445
446			unsigned push_constants[4] = {
447				srcOffset.x,
448				srcOffset.y,
449				dstOffset.x,
450				dstOffset.y,
451			};
452			radv_CmdPushConstants(radv_cmd_buffer_to_handle(cmd_buffer),
453					      device->meta_state.resolve_compute.p_layout,
454					      VK_SHADER_STAGE_COMPUTE_BIT, 0, 16,
455					      push_constants);
456			radv_unaligned_dispatch(cmd_buffer, extent.width, extent.height, 1);
457			radv_temp_descriptor_set_destroy(cmd_buffer->device, set);
458		}
459	}
460	radv_meta_restore_compute(&saved_state, cmd_buffer, 16);
461}
462