1/*
2 * Copyright (C) 2009 Maciej Cencora.
3 * Copyright (C) 2008 Nicolai Haehnle.
4 *
5 * All Rights Reserved.
6 *
7 * Permission is hereby granted, free of charge, to any person obtaining
8 * a copy of this software and associated documentation files (the
9 * "Software"), to deal in the Software without restriction, including
10 * without limitation the rights to use, copy, modify, merge, publish,
11 * distribute, sublicense, and/or sell copies of the Software, and to
12 * permit persons to whom the Software is furnished to do so, subject to
13 * the following conditions:
14 *
15 * The above copyright notice and this permission notice (including the
16 * next paragraph) shall be included in all copies or substantial
17 * portions of the Software.
18 *
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
20 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
21 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
22 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
23 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
24 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
25 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 *
27 */
28
29#include "radeon_mipmap_tree.h"
30
31#include <errno.h>
32#include <unistd.h>
33
34#include "main/simple_list.h"
35#include "main/teximage.h"
36#include "main/texobj.h"
37#include "main/enums.h"
38#include "radeon_texture.h"
39#include "radeon_tile.h"
40
41static unsigned get_aligned_compressed_row_stride(
42		gl_format format,
43		unsigned width,
44		unsigned minStride)
45{
46	const unsigned blockBytes = _mesa_get_format_bytes(format);
47	unsigned blockWidth, blockHeight;
48	unsigned stride;
49
50	_mesa_get_format_block_size(format, &blockWidth, &blockHeight);
51
52	/* Count number of blocks required to store the given width.
53	 * And then multiple it with bytes required to store a block.
54	 */
55	stride = (width + blockWidth - 1) / blockWidth * blockBytes;
56
57	/* Round the given minimum stride to the next full blocksize.
58	 * (minStride + blockBytes - 1) / blockBytes * blockBytes
59	 */
60	if ( stride < minStride )
61		stride = (minStride + blockBytes - 1) / blockBytes * blockBytes;
62
63	radeon_print(RADEON_TEXTURE, RADEON_TRACE,
64			"%s width %u, minStride %u, block(bytes %u, width %u):"
65			"stride %u\n",
66			__func__, width, minStride,
67			blockBytes, blockWidth,
68			stride);
69
70	return stride;
71}
72
73unsigned get_texture_image_size(
74		gl_format format,
75		unsigned rowStride,
76		unsigned height,
77		unsigned depth,
78		unsigned tiling)
79{
80	if (_mesa_is_format_compressed(format)) {
81		unsigned blockWidth, blockHeight;
82
83		_mesa_get_format_block_size(format, &blockWidth, &blockHeight);
84
85		return rowStride * ((height + blockHeight - 1) / blockHeight) * depth;
86	} else if (tiling) {
87		/* Need to align height to tile height */
88		unsigned tileWidth, tileHeight;
89
90		get_tile_size(format, &tileWidth, &tileHeight);
91		tileHeight--;
92
93		height = (height + tileHeight) & ~tileHeight;
94	}
95
96	return rowStride * height * depth;
97}
98
99unsigned get_texture_image_row_stride(radeonContextPtr rmesa, gl_format format, unsigned width, unsigned tiling, GLuint target)
100{
101	if (_mesa_is_format_compressed(format)) {
102		return get_aligned_compressed_row_stride(format, width, rmesa->texture_compressed_row_align);
103	} else {
104		unsigned row_align;
105
106		if (!_mesa_is_pow_two(width) || target == GL_TEXTURE_RECTANGLE) {
107			row_align = rmesa->texture_rect_row_align - 1;
108		} else if (tiling) {
109			unsigned tileWidth, tileHeight;
110			get_tile_size(format, &tileWidth, &tileHeight);
111			row_align = tileWidth * _mesa_get_format_bytes(format) - 1;
112		} else {
113			row_align = rmesa->texture_row_align - 1;
114		}
115
116		return (_mesa_format_row_stride(format, width) + row_align) & ~row_align;
117	}
118}
119
120/**
121 * Compute sizes and fill in offset and blit information for the given
122 * image (determined by \p face and \p level).
123 *
124 * \param curOffset points to the offset at which the image is to be stored
125 * and is updated by this function according to the size of the image.
126 */
127static void compute_tex_image_offset(radeonContextPtr rmesa, radeon_mipmap_tree *mt,
128	GLuint face, GLuint level, GLuint* curOffset)
129{
130	radeon_mipmap_level *lvl = &mt->levels[level];
131	GLuint height;
132
133	height = _mesa_next_pow_two_32(lvl->height);
134
135	lvl->rowstride = get_texture_image_row_stride(rmesa, mt->mesaFormat, lvl->width, mt->tilebits, mt->target);
136	lvl->size = get_texture_image_size(mt->mesaFormat, lvl->rowstride, height, lvl->depth, mt->tilebits);
137
138	assert(lvl->size > 0);
139
140	lvl->faces[face].offset = *curOffset;
141	*curOffset += lvl->size;
142
143	radeon_print(RADEON_TEXTURE, RADEON_TRACE,
144			"%s(%p) level %d, face %d: rs:%d %dx%d at %d\n",
145			__func__, rmesa,
146			level, face,
147			lvl->rowstride, lvl->width, height, lvl->faces[face].offset);
148}
149
150static GLuint minify(GLuint size, GLuint levels)
151{
152	size = size >> levels;
153	if (size < 1)
154		size = 1;
155	return size;
156}
157
158
159static void calculate_miptree_layout(radeonContextPtr rmesa, radeon_mipmap_tree *mt)
160{
161	GLuint curOffset, i, face, level;
162
163	assert(mt->numLevels <= rmesa->glCtx->Const.MaxTextureLevels);
164
165	curOffset = 0;
166	for(face = 0; face < mt->faces; face++) {
167
168		for(i = 0, level = mt->baseLevel; i < mt->numLevels; i++, level++) {
169			mt->levels[level].valid = 1;
170			mt->levels[level].width = minify(mt->width0, i);
171			mt->levels[level].height = minify(mt->height0, i);
172			mt->levels[level].depth = minify(mt->depth0, i);
173			compute_tex_image_offset(rmesa, mt, face, level, &curOffset);
174		}
175	}
176
177	/* Note the required size in memory */
178	mt->totalsize = (curOffset + RADEON_OFFSET_MASK) & ~RADEON_OFFSET_MASK;
179
180	radeon_print(RADEON_TEXTURE, RADEON_TRACE,
181			"%s(%p, %p) total size %d\n",
182			__func__, rmesa, mt, mt->totalsize);
183}
184
185/**
186 * Create a new mipmap tree, calculate its layout and allocate memory.
187 */
188radeon_mipmap_tree* radeon_miptree_create(radeonContextPtr rmesa,
189					  GLenum target, gl_format mesaFormat, GLuint baseLevel, GLuint numLevels,
190					  GLuint width0, GLuint height0, GLuint depth0, GLuint tilebits)
191{
192	radeon_mipmap_tree *mt = CALLOC_STRUCT(_radeon_mipmap_tree);
193
194	radeon_print(RADEON_TEXTURE, RADEON_NORMAL,
195		"%s(%p) new tree is %p.\n",
196		__func__, rmesa, mt);
197
198	mt->mesaFormat = mesaFormat;
199	mt->refcount = 1;
200	mt->target = target;
201	mt->faces = _mesa_num_tex_faces(target);
202	mt->baseLevel = baseLevel;
203	mt->numLevels = numLevels;
204	mt->width0 = width0;
205	mt->height0 = height0;
206	mt->depth0 = depth0;
207	mt->tilebits = tilebits;
208
209	calculate_miptree_layout(rmesa, mt);
210
211	mt->bo = radeon_bo_open(rmesa->radeonScreen->bom,
212                            0, mt->totalsize, 1024,
213                            RADEON_GEM_DOMAIN_VRAM,
214                            0);
215
216	return mt;
217}
218
219void radeon_miptree_reference(radeon_mipmap_tree *mt, radeon_mipmap_tree **ptr)
220{
221	assert(!*ptr);
222
223	mt->refcount++;
224	assert(mt->refcount > 0);
225
226	*ptr = mt;
227}
228
229void radeon_miptree_unreference(radeon_mipmap_tree **ptr)
230{
231	radeon_mipmap_tree *mt = *ptr;
232	if (!mt)
233		return;
234
235	assert(mt->refcount > 0);
236
237	mt->refcount--;
238	if (!mt->refcount) {
239		radeon_bo_unref(mt->bo);
240		free(mt);
241	}
242
243	*ptr = 0;
244}
245
246/**
247 * Calculate min and max LOD for the given texture object.
248 * @param[in] tObj texture object whose LOD values to calculate
249 * @param[out] pminLod minimal LOD
250 * @param[out] pmaxLod maximal LOD
251 */
252static void calculate_min_max_lod(struct gl_sampler_object *samp, struct gl_texture_object *tObj,
253				       unsigned *pminLod, unsigned *pmaxLod)
254{
255	int minLod, maxLod;
256	/* Yes, this looks overly complicated, but it's all needed.
257	*/
258	switch (tObj->Target) {
259	case GL_TEXTURE_1D:
260	case GL_TEXTURE_2D:
261	case GL_TEXTURE_3D:
262	case GL_TEXTURE_CUBE_MAP:
263		if (samp->MinFilter == GL_NEAREST || samp->MinFilter == GL_LINEAR) {
264			/* GL_NEAREST and GL_LINEAR only care about GL_TEXTURE_BASE_LEVEL.
265			*/
266			minLod = maxLod = tObj->BaseLevel;
267		} else {
268			minLod = tObj->BaseLevel + (GLint)(samp->MinLod);
269			minLod = MAX2(minLod, tObj->BaseLevel);
270			minLod = MIN2(minLod, tObj->MaxLevel);
271			maxLod = tObj->BaseLevel + (GLint)(samp->MaxLod + 0.5);
272			maxLod = MIN2(maxLod, tObj->MaxLevel);
273			maxLod = MIN2(maxLod, tObj->Image[0][minLod]->MaxNumLevels - 1 + minLod);
274			maxLod = MAX2(maxLod, minLod); /* need at least one level */
275		}
276		break;
277	case GL_TEXTURE_RECTANGLE_NV:
278	case GL_TEXTURE_4D_SGIS:
279		minLod = maxLod = 0;
280		break;
281	default:
282		return;
283	}
284
285	radeon_print(RADEON_TEXTURE, RADEON_TRACE,
286			"%s(%p) target %s, min %d, max %d.\n",
287			__func__, tObj,
288			_mesa_lookup_enum_by_nr(tObj->Target),
289			minLod, maxLod);
290
291	/* save these values */
292	*pminLod = minLod;
293	*pmaxLod = maxLod;
294}
295
296/**
297 * Checks whether the given miptree can hold the given texture image at the
298 * given face and level.
299 */
300GLboolean radeon_miptree_matches_image(radeon_mipmap_tree *mt,
301				       struct gl_texture_image *texImage)
302{
303	radeon_mipmap_level *lvl;
304	GLuint level = texImage->Level;
305	if (texImage->TexFormat != mt->mesaFormat)
306		return GL_FALSE;
307
308	lvl = &mt->levels[level];
309	if (!lvl->valid ||
310	    lvl->width != texImage->Width ||
311	    lvl->height != texImage->Height ||
312	    lvl->depth != texImage->Depth)
313		return GL_FALSE;
314
315	return GL_TRUE;
316}
317
318/**
319 * Checks whether the given miptree has the right format to store the given texture object.
320 */
321static GLboolean radeon_miptree_matches_texture(radeon_mipmap_tree *mt, struct gl_texture_object *texObj)
322{
323	struct gl_texture_image *firstImage;
324	unsigned numLevels;
325	radeon_mipmap_level *mtBaseLevel;
326
327	if (texObj->BaseLevel < mt->baseLevel)
328		return GL_FALSE;
329
330	mtBaseLevel = &mt->levels[texObj->BaseLevel - mt->baseLevel];
331	firstImage = texObj->Image[0][texObj->BaseLevel];
332	numLevels = MIN2(texObj->_MaxLevel - texObj->BaseLevel + 1, firstImage->MaxNumLevels);
333
334	if (radeon_is_debug_enabled(RADEON_TEXTURE,RADEON_TRACE)) {
335		fprintf(stderr, "Checking if miptree %p matches texObj %p\n", mt, texObj);
336		fprintf(stderr, "target %d vs %d\n", mt->target, texObj->Target);
337		fprintf(stderr, "format %d vs %d\n", mt->mesaFormat, firstImage->TexFormat);
338		fprintf(stderr, "numLevels %d vs %d\n", mt->numLevels, numLevels);
339		fprintf(stderr, "width0 %d vs %d\n", mtBaseLevel->width, firstImage->Width);
340		fprintf(stderr, "height0 %d vs %d\n", mtBaseLevel->height, firstImage->Height);
341		fprintf(stderr, "depth0 %d vs %d\n", mtBaseLevel->depth, firstImage->Depth);
342		if (mt->target == texObj->Target &&
343	        mt->mesaFormat == firstImage->TexFormat &&
344	        mt->numLevels >= numLevels &&
345	        mtBaseLevel->width == firstImage->Width &&
346	        mtBaseLevel->height == firstImage->Height &&
347	        mtBaseLevel->depth == firstImage->Depth) {
348			fprintf(stderr, "MATCHED\n");
349		} else {
350			fprintf(stderr, "NOT MATCHED\n");
351		}
352	}
353
354	return (mt->target == texObj->Target &&
355	        mt->mesaFormat == firstImage->TexFormat &&
356	        mt->numLevels >= numLevels &&
357	        mtBaseLevel->width == firstImage->Width &&
358	        mtBaseLevel->height == firstImage->Height &&
359	        mtBaseLevel->depth == firstImage->Depth);
360}
361
362/**
363 * Try to allocate a mipmap tree for the given texture object.
364 * @param[in] rmesa radeon context
365 * @param[in] t radeon texture object
366 */
367void radeon_try_alloc_miptree(radeonContextPtr rmesa, radeonTexObj *t)
368{
369	struct gl_texture_object *texObj = &t->base;
370	struct gl_texture_image *texImg = texObj->Image[0][texObj->BaseLevel];
371	GLuint numLevels;
372	assert(!t->mt);
373
374	if (!texImg) {
375		radeon_warning("%s(%p) No image in given texture object(%p).\n",
376				__func__, rmesa, t);
377		return;
378	}
379
380
381	numLevels = MIN2(texObj->MaxLevel - texObj->BaseLevel + 1, texImg->MaxNumLevels);
382
383	t->mt = radeon_miptree_create(rmesa, t->base.Target,
384		texImg->TexFormat, texObj->BaseLevel,
385		numLevels, texImg->Width, texImg->Height,
386		texImg->Depth, t->tile_bits);
387}
388
389GLuint
390radeon_miptree_image_offset(radeon_mipmap_tree *mt,
391			    GLuint face, GLuint level)
392{
393	if (mt->target == GL_TEXTURE_CUBE_MAP_ARB)
394		return (mt->levels[level].faces[face].offset);
395	else
396		return mt->levels[level].faces[0].offset;
397}
398
399/**
400 * Ensure that the given image is stored in the given miptree from now on.
401 */
402static void migrate_image_to_miptree(radeon_mipmap_tree *mt,
403									 radeon_texture_image *image,
404									 int face, int level)
405{
406	radeon_mipmap_level *dstlvl = &mt->levels[level];
407	unsigned char *dest;
408
409	assert(image->mt != mt);
410	assert(dstlvl->valid);
411	assert(dstlvl->width == image->base.Base.Width);
412	assert(dstlvl->height == image->base.Base.Height);
413	assert(dstlvl->depth == image->base.Base.Depth);
414
415	radeon_print(RADEON_TEXTURE, RADEON_VERBOSE,
416			"%s miptree %p, image %p, face %d, level %d.\n",
417			__func__, mt, image, face, level);
418
419	radeon_bo_map(mt->bo, GL_TRUE);
420	dest = mt->bo->ptr + dstlvl->faces[face].offset;
421
422	if (image->mt) {
423		/* Format etc. should match, so we really just need a memcpy().
424		 * In fact, that memcpy() could be done by the hardware in many
425		 * cases, provided that we have a proper memory manager.
426		 */
427		assert(mt->mesaFormat == image->base.Base.TexFormat);
428
429		radeon_mipmap_level *srclvl = &image->mt->levels[image->base.Base.Level];
430
431		assert(image->base.Base.Level == level);
432		assert(srclvl->size == dstlvl->size);
433		assert(srclvl->rowstride == dstlvl->rowstride);
434
435		radeon_bo_map(image->mt->bo, GL_FALSE);
436
437		memcpy(dest,
438			image->mt->bo->ptr + srclvl->faces[face].offset,
439			dstlvl->size);
440		radeon_bo_unmap(image->mt->bo);
441
442		radeon_miptree_unreference(&image->mt);
443	} else if (image->base.Map) {
444		/* This condition should be removed, it's here to workaround
445		 * a segfault when mapping textures during software fallbacks.
446		 */
447		radeon_print(RADEON_FALLBACKS, RADEON_IMPORTANT,
448				"%s Trying to map texture in software fallback.\n",
449				__func__);
450		const uint32_t srcrowstride = _mesa_format_row_stride(image->base.Base.TexFormat, image->base.Base.Width);
451		uint32_t rows = image->base.Base.Height * image->base.Base.Depth;
452
453		if (_mesa_is_format_compressed(image->base.Base.TexFormat)) {
454			uint32_t blockWidth, blockHeight;
455			_mesa_get_format_block_size(image->base.Base.TexFormat, &blockWidth, &blockHeight);
456			rows = (rows + blockHeight - 1) / blockHeight;
457		}
458
459		copy_rows(dest, dstlvl->rowstride, image->base.Map, srcrowstride,
460				  rows, srcrowstride);
461
462		_mesa_align_free(image->base.Map);
463		image->base.Map = 0;
464	}
465
466	radeon_bo_unmap(mt->bo);
467
468	radeon_miptree_reference(mt, &image->mt);
469}
470
471/**
472 * Filter matching miptrees, and select one with the most of data.
473 * @param[in] texObj radeon texture object
474 * @param[in] firstLevel first texture level to check
475 * @param[in] lastLevel last texture level to check
476 */
477static radeon_mipmap_tree * get_biggest_matching_miptree(radeonTexObj *texObj,
478														 unsigned firstLevel,
479														 unsigned lastLevel)
480{
481	const unsigned numLevels = lastLevel - firstLevel + 1;
482	unsigned *mtSizes = calloc(numLevels, sizeof(unsigned));
483	radeon_mipmap_tree **mts = calloc(numLevels, sizeof(radeon_mipmap_tree *));
484	unsigned mtCount = 0;
485	unsigned maxMtIndex = 0;
486	radeon_mipmap_tree *tmp;
487	unsigned int level;
488	int i;
489
490	for (level = firstLevel; level <= lastLevel; ++level) {
491		radeon_texture_image *img = get_radeon_texture_image(texObj->base.Image[0][level]);
492		unsigned found = 0;
493		// TODO: why this hack??
494		if (!img)
495			break;
496
497		if (!img->mt)
498			continue;
499
500		for (i = 0; i < mtCount; ++i) {
501			if (mts[i] == img->mt) {
502				found = 1;
503				mtSizes[i] += img->mt->levels[img->base.Base.Level].size;
504				break;
505			}
506		}
507
508		if (!found && radeon_miptree_matches_texture(img->mt, &texObj->base)) {
509			mtSizes[mtCount] = img->mt->levels[img->base.Base.Level].size;
510			mts[mtCount] = img->mt;
511			mtCount++;
512		}
513	}
514
515	if (mtCount == 0) {
516		free(mtSizes);
517		free(mts);
518		return NULL;
519	}
520
521	for (i = 1; i < mtCount; ++i) {
522		if (mtSizes[i] > mtSizes[maxMtIndex]) {
523			maxMtIndex = i;
524		}
525	}
526
527	tmp = mts[maxMtIndex];
528	free(mtSizes);
529	free(mts);
530
531	return tmp;
532}
533
534/**
535 * Validate texture mipmap tree.
536 * If individual images are stored in different mipmap trees
537 * use the mipmap tree that has the most of the correct data.
538 */
539int radeon_validate_texture_miptree(struct gl_context * ctx,
540				    struct gl_sampler_object *samp,
541				    struct gl_texture_object *texObj)
542{
543	radeonContextPtr rmesa = RADEON_CONTEXT(ctx);
544	radeonTexObj *t = radeon_tex_obj(texObj);
545	radeon_mipmap_tree *dst_miptree;
546
547	if (samp == &texObj->Sampler && (t->validated || t->image_override)) {
548		return GL_TRUE;
549	}
550
551	calculate_min_max_lod(samp, &t->base, &t->minLod, &t->maxLod);
552
553	radeon_print(RADEON_TEXTURE, RADEON_NORMAL,
554			"%s: Validating texture %p now, minLod = %d, maxLod = %d\n",
555			__FUNCTION__, texObj ,t->minLod, t->maxLod);
556
557	dst_miptree = get_biggest_matching_miptree(t, t->base.BaseLevel, t->base._MaxLevel);
558
559	radeon_miptree_unreference(&t->mt);
560	if (!dst_miptree) {
561		radeon_try_alloc_miptree(rmesa, t);
562		radeon_print(RADEON_TEXTURE, RADEON_NORMAL,
563			"%s: No matching miptree found, allocated new one %p\n",
564			__FUNCTION__, t->mt);
565
566	} else {
567		radeon_miptree_reference(dst_miptree, &t->mt);
568		radeon_print(RADEON_TEXTURE, RADEON_NORMAL,
569			"%s: Using miptree %p\n", __FUNCTION__, t->mt);
570	}
571
572	const unsigned faces = _mesa_num_tex_faces(texObj->Target);
573	unsigned face, level;
574	radeon_texture_image *img;
575	/* Validate only the levels that will actually be used during rendering */
576	for (face = 0; face < faces; ++face) {
577		for (level = t->minLod; level <= t->maxLod; ++level) {
578			img = get_radeon_texture_image(texObj->Image[face][level]);
579
580			radeon_print(RADEON_TEXTURE, RADEON_TRACE,
581				"Checking image level %d, face %d, mt %p ... ",
582				level, face, img->mt);
583
584			if (img->mt != t->mt && !img->used_as_render_target) {
585				radeon_print(RADEON_TEXTURE, RADEON_TRACE,
586					"MIGRATING\n");
587
588				struct radeon_bo *src_bo = (img->mt) ? img->mt->bo : img->bo;
589				if (src_bo && radeon_bo_is_referenced_by_cs(src_bo, rmesa->cmdbuf.cs)) {
590					radeon_firevertices(rmesa);
591				}
592				migrate_image_to_miptree(t->mt, img, face, level);
593			} else
594				radeon_print(RADEON_TEXTURE, RADEON_TRACE, "OK\n");
595		}
596	}
597
598	t->validated = GL_TRUE;
599
600	return GL_TRUE;
601}
602
603uint32_t get_base_teximage_offset(radeonTexObj *texObj)
604{
605	if (!texObj->mt) {
606		return 0;
607	} else {
608		return radeon_miptree_image_offset(texObj->mt, 0, texObj->minLod);
609	}
610}
611