radeonsi_pm4.c revision 862df0885aa04ef68319382fc2df27a7b68fc0dd
1/*
2 * Copyright 2012 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
22 *
23 * Authors:
24 *      Christian König <christian.koenig@amd.com>
25 */
26
27#include "util/u_memory.h"
28#include "radeonsi_pipe.h"
29#include "radeonsi_pm4.h"
30#include "sid.h"
31#include "r600_hw_context_priv.h"
32
33#define NUMBER_OF_STATES (sizeof(union si_state) / sizeof(struct si_pm4_state *))
34
35void si_pm4_cmd_begin(struct si_pm4_state *state, unsigned opcode)
36{
37	state->last_opcode = opcode;
38	state->last_pm4 = state->ndw++;
39}
40
41void si_pm4_cmd_add(struct si_pm4_state *state, uint32_t dw)
42{
43	state->pm4[state->ndw++] = dw;
44}
45
46void si_pm4_cmd_end(struct si_pm4_state *state, bool predicate)
47{
48	unsigned count;
49	count = state->ndw - state->last_pm4 - 2;
50	state->pm4[state->last_pm4] = PKT3(state->last_opcode,
51					   count, predicate);
52
53	assert(state->ndw <= SI_PM4_MAX_DW);
54}
55
56void si_pm4_set_reg(struct si_pm4_state *state, unsigned reg, uint32_t val)
57{
58	unsigned opcode;
59
60	if (reg >= SI_CONFIG_REG_OFFSET && reg <= SI_CONFIG_REG_END) {
61		opcode = PKT3_SET_CONFIG_REG;
62		reg -= SI_CONFIG_REG_OFFSET;
63
64	} else if (reg >= SI_SH_REG_OFFSET && reg <= SI_SH_REG_END) {
65		opcode = PKT3_SET_SH_REG;
66		reg -= SI_SH_REG_OFFSET;
67
68	} else if (reg >= SI_CONTEXT_REG_OFFSET && reg <= SI_CONTEXT_REG_END) {
69		opcode = PKT3_SET_CONTEXT_REG;
70		reg -= SI_CONTEXT_REG_OFFSET;
71	} else {
72		R600_ERR("Invalid register offset %08x!\n", reg);
73		return;
74	}
75
76	reg >>= 2;
77
78	if (opcode != state->last_opcode || reg != (state->last_reg + 1)) {
79		si_pm4_cmd_begin(state, opcode);
80		si_pm4_cmd_add(state, reg);
81	}
82
83	state->last_reg = reg;
84	si_pm4_cmd_add(state, val);
85	si_pm4_cmd_end(state, false);
86}
87
88void si_pm4_add_bo(struct si_pm4_state *state,
89                   struct si_resource *bo,
90                   enum radeon_bo_usage usage)
91{
92	unsigned idx = state->nbo++;
93	assert(idx < SI_PM4_MAX_BO);
94
95	si_resource_reference(&state->bo[idx], bo);
96	state->bo_usage[idx] = usage;
97}
98
99void si_pm4_inval_shader_cache(struct si_pm4_state *state)
100{
101	state->cp_coher_cntl |= S_0085F0_SH_ICACHE_ACTION_ENA(1);
102	state->cp_coher_cntl |= S_0085F0_SH_KCACHE_ACTION_ENA(1);
103}
104
105void si_pm4_inval_texture_cache(struct si_pm4_state *state)
106{
107	state->cp_coher_cntl |= S_0085F0_TC_ACTION_ENA(1);
108}
109
110void si_pm4_inval_vertex_cache(struct si_pm4_state *state)
111{
112        /* Some GPUs don't have the vertex cache and must use the texture cache instead. */
113	state->cp_coher_cntl |= S_0085F0_TC_ACTION_ENA(1);
114}
115
116void si_pm4_inval_fb_cache(struct si_pm4_state *state, unsigned nr_cbufs)
117{
118	state->cp_coher_cntl |= S_0085F0_CB_ACTION_ENA(1);
119	state->cp_coher_cntl |= ((1 << nr_cbufs) - 1) << S_0085F0_CB0_DEST_BASE_ENA_SHIFT;
120}
121
122void si_pm4_inval_zsbuf_cache(struct si_pm4_state *state)
123{
124	state->cp_coher_cntl |= S_0085F0_DB_ACTION_ENA(1) | S_0085F0_DB_DEST_BASE_ENA(1);
125}
126
127void si_pm4_free_state(struct r600_context *rctx,
128		       struct si_pm4_state *state,
129		       unsigned idx)
130{
131	if (state == NULL)
132		return;
133
134	if (rctx->emitted.array[idx] == state) {
135		rctx->emitted.array[idx] = NULL;
136	}
137
138	for (int i = 0; i < state->nbo; ++i) {
139		si_resource_reference(&state->bo[i], NULL);
140	}
141	FREE(state);
142}
143
144unsigned si_pm4_dirty_dw(struct r600_context *rctx)
145{
146	unsigned count = 0;
147	uint32_t cp_coher_cntl = 0;
148
149	for (int i = 0; i < NUMBER_OF_STATES; ++i) {
150		struct si_pm4_state *state = rctx->queued.array[i];
151
152		if (!state || rctx->emitted.array[i] == state)
153			continue;
154
155		count += state->ndw;
156		cp_coher_cntl |= state->cp_coher_cntl;
157	}
158
159	//TODO
160	rctx->atom_surface_sync.flush_flags |= cp_coher_cntl;
161	r600_atom_dirty(rctx, &rctx->atom_surface_sync.atom);
162	return count;
163}
164
165void si_pm4_emit_dirty(struct r600_context *rctx)
166{
167	struct radeon_winsys_cs *cs = rctx->cs;
168
169	for (int i = 0; i < NUMBER_OF_STATES; ++i) {
170		struct si_pm4_state *state = rctx->queued.array[i];
171
172		if (!state || rctx->emitted.array[i] == state)
173			continue;
174
175		for (int j = 0; j < state->nbo; ++j) {
176			r600_context_bo_reloc(rctx, state->bo[j],
177					      state->bo_usage[j]);
178		}
179
180		memcpy(&cs->buf[cs->cdw], state->pm4, state->ndw * 4);
181		cs->cdw += state->ndw;
182
183		rctx->emitted.array[i] = state;
184	}
185}
186
187void si_pm4_reset_emitted(struct r600_context *rctx)
188{
189	memset(&rctx->emitted, 0, sizeof(rctx->emitted));
190}
191