1/* 2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org> 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * on the rights to use, copy, modify, merge, publish, distribute, sub 8 * license, and/or sell copies of the Software, and to permit persons to whom 9 * the Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, 19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 21 * USE OR OTHER DEALINGS IN THE SOFTWARE. 22 * 23 * Authors: 24 * Jerome Glisse 25 */ 26#include "r600_pipe.h" 27#include "evergreend.h" 28#include "util/u_memory.h" 29#include "util/u_math.h" 30 31void evergreen_dma_copy_buffer(struct r600_context *rctx, 32 struct pipe_resource *dst, 33 struct pipe_resource *src, 34 uint64_t dst_offset, 35 uint64_t src_offset, 36 uint64_t size) 37{ 38 struct radeon_winsys_cs *cs = rctx->b.dma.cs; 39 unsigned i, ncopy, csize, sub_cmd, shift; 40 struct r600_resource *rdst = (struct r600_resource*)dst; 41 struct r600_resource *rsrc = (struct r600_resource*)src; 42 43 /* Mark the buffer range of destination as valid (initialized), 44 * so that transfer_map knows it should wait for the GPU when mapping 45 * that range. */ 46 util_range_add(&rdst->valid_buffer_range, dst_offset, 47 dst_offset + size); 48 49 dst_offset += rdst->gpu_address; 50 src_offset += rsrc->gpu_address; 51 52 /* see if we use dword or byte copy */ 53 if (!(dst_offset % 4) && !(src_offset % 4) && !(size % 4)) { 54 size >>= 2; 55 sub_cmd = EG_DMA_COPY_DWORD_ALIGNED; 56 shift = 2; 57 } else { 58 sub_cmd = EG_DMA_COPY_BYTE_ALIGNED; 59 shift = 0; 60 } 61 ncopy = (size / EG_DMA_COPY_MAX_SIZE) + !!(size % EG_DMA_COPY_MAX_SIZE); 62 63 r600_need_dma_space(&rctx->b, ncopy * 5, rdst, rsrc); 64 for (i = 0; i < ncopy; i++) { 65 csize = size < EG_DMA_COPY_MAX_SIZE ? size : EG_DMA_COPY_MAX_SIZE; 66 /* emit reloc before writing cs so that cs is always in consistent state */ 67 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rsrc, RADEON_USAGE_READ, 68 RADEON_PRIO_SDMA_BUFFER); 69 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rdst, RADEON_USAGE_WRITE, 70 RADEON_PRIO_SDMA_BUFFER); 71 radeon_emit(cs, DMA_PACKET(DMA_PACKET_COPY, sub_cmd, csize)); 72 radeon_emit(cs, dst_offset & 0xffffffff); 73 radeon_emit(cs, src_offset & 0xffffffff); 74 radeon_emit(cs, (dst_offset >> 32UL) & 0xff); 75 radeon_emit(cs, (src_offset >> 32UL) & 0xff); 76 dst_offset += csize << shift; 77 src_offset += csize << shift; 78 size -= csize; 79 } 80} 81 82/* The max number of bytes to copy per packet. */ 83#define CP_DMA_MAX_BYTE_COUNT ((1 << 21) - 8) 84 85void evergreen_cp_dma_clear_buffer(struct r600_context *rctx, 86 struct pipe_resource *dst, uint64_t offset, 87 unsigned size, uint32_t clear_value, 88 enum r600_coherency coher) 89{ 90 struct radeon_winsys_cs *cs = rctx->b.gfx.cs; 91 92 assert(size); 93 assert(rctx->screen->b.has_cp_dma); 94 95 /* Mark the buffer range of destination as valid (initialized), 96 * so that transfer_map knows it should wait for the GPU when mapping 97 * that range. */ 98 util_range_add(&r600_resource(dst)->valid_buffer_range, offset, 99 offset + size); 100 101 offset += r600_resource(dst)->gpu_address; 102 103 /* Flush the cache where the resource is bound. */ 104 rctx->b.flags |= r600_get_flush_flags(coher) | 105 R600_CONTEXT_WAIT_3D_IDLE; 106 107 while (size) { 108 unsigned sync = 0; 109 unsigned byte_count = MIN2(size, CP_DMA_MAX_BYTE_COUNT); 110 unsigned reloc; 111 112 r600_need_cs_space(rctx, 113 10 + (rctx->b.flags ? R600_MAX_FLUSH_CS_DWORDS : 0) + 114 R600_MAX_PFP_SYNC_ME_DWORDS, FALSE); 115 116 /* Flush the caches for the first copy only. */ 117 if (rctx->b.flags) { 118 r600_flush_emit(rctx); 119 } 120 121 /* Do the synchronization after the last copy, so that all data is written to memory. */ 122 if (size == byte_count) { 123 sync = PKT3_CP_DMA_CP_SYNC; 124 } 125 126 /* This must be done after r600_need_cs_space. */ 127 reloc = radeon_add_to_buffer_list(&rctx->b, &rctx->b.gfx, 128 (struct r600_resource*)dst, RADEON_USAGE_WRITE, 129 RADEON_PRIO_CP_DMA); 130 131 radeon_emit(cs, PKT3(PKT3_CP_DMA, 4, 0)); 132 radeon_emit(cs, clear_value); /* DATA [31:0] */ 133 radeon_emit(cs, sync | PKT3_CP_DMA_SRC_SEL(2)); /* CP_SYNC [31] | SRC_SEL[30:29] */ 134 radeon_emit(cs, offset); /* DST_ADDR_LO [31:0] */ 135 radeon_emit(cs, (offset >> 32) & 0xff); /* DST_ADDR_HI [7:0] */ 136 radeon_emit(cs, byte_count); /* COMMAND [29:22] | BYTE_COUNT [20:0] */ 137 138 radeon_emit(cs, PKT3(PKT3_NOP, 0, 0)); 139 radeon_emit(cs, reloc); 140 141 size -= byte_count; 142 offset += byte_count; 143 } 144 145 /* CP DMA is executed in ME, but index buffers are read by PFP. 146 * This ensures that ME (CP DMA) is idle before PFP starts fetching 147 * indices. If we wanted to execute CP DMA in PFP, this packet 148 * should precede it. 149 */ 150 if (coher == R600_COHERENCY_SHADER) 151 r600_emit_pfp_sync_me(rctx); 152} 153