1 2#ifndef COMMON_CONTEXT_H 3#define COMMON_CONTEXT_H 4 5#include "main/mm.h" 6#include "math/m_vector.h" 7#include "tnl/t_context.h" 8#include "main/colormac.h" 9 10#include "radeon_screen.h" 11#include "radeon_debug.h" 12#include "radeon_drm.h" 13#include "dri_util.h" 14#include "tnl/t_vertex.h" 15#include "swrast/s_context.h" 16 17struct radeon_context; 18 19#include "radeon_bo_gem.h" 20#include "radeon_cs_gem.h" 21 22/* This union is used to avoid warnings/miscompilation 23 with float to uint32_t casts due to strict-aliasing */ 24typedef union { GLfloat f; uint32_t ui32; } float_ui32_type; 25 26struct radeon_context; 27typedef struct radeon_context radeonContextRec; 28typedef struct radeon_context *radeonContextPtr; 29 30 31#define TEX_0 0x1 32#define TEX_1 0x2 33#define TEX_2 0x4 34#define TEX_3 0x8 35#define TEX_4 0x10 36#define TEX_5 0x20 37 38/* Rasterizing fallbacks */ 39/* See correponding strings in r200_swtcl.c */ 40#define RADEON_FALLBACK_TEXTURE 0x0001 41#define RADEON_FALLBACK_DRAW_BUFFER 0x0002 42#define RADEON_FALLBACK_STENCIL 0x0004 43#define RADEON_FALLBACK_RENDER_MODE 0x0008 44#define RADEON_FALLBACK_BLEND_EQ 0x0010 45#define RADEON_FALLBACK_BLEND_FUNC 0x0020 46#define RADEON_FALLBACK_DISABLE 0x0040 47#define RADEON_FALLBACK_BORDER_MODE 0x0080 48#define RADEON_FALLBACK_DEPTH_BUFFER 0x0100 49#define RADEON_FALLBACK_STENCIL_BUFFER 0x0200 50 51#define R200_FALLBACK_TEXTURE 0x01 52#define R200_FALLBACK_DRAW_BUFFER 0x02 53#define R200_FALLBACK_STENCIL 0x04 54#define R200_FALLBACK_RENDER_MODE 0x08 55#define R200_FALLBACK_DISABLE 0x10 56#define R200_FALLBACK_BORDER_MODE 0x20 57 58#define RADEON_TCL_FALLBACK_RASTER 0x1 /* rasterization */ 59#define RADEON_TCL_FALLBACK_UNFILLED 0x2 /* unfilled tris */ 60#define RADEON_TCL_FALLBACK_LIGHT_TWOSIDE 0x4 /* twoside tris */ 61#define RADEON_TCL_FALLBACK_MATERIAL 0x8 /* material in vb */ 62#define RADEON_TCL_FALLBACK_TEXGEN_0 0x10 /* texgen, unit 0 */ 63#define RADEON_TCL_FALLBACK_TEXGEN_1 0x20 /* texgen, unit 1 */ 64#define RADEON_TCL_FALLBACK_TEXGEN_2 0x40 /* texgen, unit 2 */ 65#define RADEON_TCL_FALLBACK_TCL_DISABLE 0x80 /* user disable */ 66#define RADEON_TCL_FALLBACK_FOGCOORDSPEC 0x100 /* fogcoord, sep. spec light */ 67 68/* The blit width for texture uploads 69 */ 70#define BLIT_WIDTH_BYTES 1024 71 72/* Use the templated vertex format: 73 */ 74#define COLOR_IS_RGBA 75#define TAG(x) radeon##x 76#include "tnl_dd/t_dd_vertex.h" 77#undef TAG 78 79#define RADEON_RB_CLASS 0xdeadbeef 80 81struct radeon_renderbuffer 82{ 83 struct swrast_renderbuffer base; 84 85 struct radeon_bo *bo; 86 unsigned int cpp; 87 /* unsigned int offset; */ 88 unsigned int pitch; 89 90 struct radeon_bo *map_bo; 91 GLbitfield map_mode; 92 int map_x, map_y, map_w, map_h; 93 int map_pitch; 94 void *map_buffer; 95 96 uint32_t draw_offset; /* FBO */ 97 /* boo Xorg 6.8.2 compat */ 98 int has_surface; 99 100 GLuint pf_pending; /**< sequence number of pending flip */ 101 __DRIdrawable *dPriv; 102}; 103 104struct radeon_framebuffer 105{ 106 struct gl_framebuffer base; 107 108 struct radeon_renderbuffer *color_rb[2]; 109}; 110 111 112struct radeon_colorbuffer_state { 113 int roundEnable; 114 struct gl_renderbuffer *rb; 115 uint32_t draw_offset; /* offset into color renderbuffer - FBOs */ 116}; 117 118struct radeon_depthbuffer_state { 119 struct gl_renderbuffer *rb; 120}; 121 122struct radeon_scissor_state { 123 drm_clip_rect_t rect; 124 GLboolean enabled; 125}; 126 127struct radeon_state_atom { 128 struct radeon_state_atom *next, *prev; 129 const char *name; /* for debug */ 130 int cmd_size; /* size in bytes */ 131 GLuint idx; 132 GLuint is_tcl; 133 GLuint *cmd; /* one or more cmd's */ 134 GLuint *lastcmd; /* one or more cmd's */ 135 GLboolean dirty; /* dirty-mark in emit_state_list */ 136 int (*check) (struct gl_context *, struct radeon_state_atom *atom); /* is this state active? */ 137 void (*emit) (struct gl_context *, struct radeon_state_atom *atom); 138}; 139 140struct radeon_hw_state { 141 /* Head of the linked list of state atoms. */ 142 struct radeon_state_atom atomlist; 143 int max_state_size; /* Number of bytes necessary for a full state emit. */ 144 int max_post_flush_size; /* Number of bytes necessary for post flushing emits */ 145 GLboolean is_dirty, all_dirty; 146}; 147 148 149/* Texture related */ 150typedef struct _radeon_texture_image radeon_texture_image; 151 152 153/** 154 * This is a subclass of swrast_texture_image since we use swrast 155 * for software fallback rendering. 156 */ 157struct _radeon_texture_image { 158 struct swrast_texture_image base; 159 160 /** 161 * If mt != 0, the image is stored in hardware format in the 162 * given mipmap tree. In this case, base.Data may point into the 163 * mapping of the buffer object that contains the mipmap tree. 164 * 165 * If mt == 0, the image is stored in normal memory pointed to 166 * by base.Data. 167 */ 168 struct _radeon_mipmap_tree *mt; 169 struct radeon_bo *bo; 170 GLboolean used_as_render_target; 171}; 172 173 174static inline radeon_texture_image *get_radeon_texture_image(struct gl_texture_image *image) 175{ 176 return (radeon_texture_image*)image; 177} 178 179 180typedef struct radeon_tex_obj radeonTexObj, *radeonTexObjPtr; 181 182#define RADEON_TXO_MICRO_TILE (1 << 3) 183 184/* Texture object in locally shared texture space. 185 */ 186struct radeon_tex_obj { 187 struct gl_texture_object base; 188 struct _radeon_mipmap_tree *mt; 189 190 /** 191 * This is true if we've verified that the mipmap tree above is complete 192 * and so on. 193 */ 194 GLboolean validated; 195 /* Minimum LOD to be used during rendering */ 196 unsigned minLod; 197 /* Miximum LOD to be used during rendering */ 198 unsigned maxLod; 199 200 GLuint override_offset; 201 GLboolean image_override; /* Image overridden by GLX_EXT_tfp */ 202 GLuint tile_bits; /* hw texture tile bits used on this texture */ 203 struct radeon_bo *bo; 204 205 GLuint pp_txfilter; /* hardware register values */ 206 GLuint pp_txformat; 207 GLuint pp_txformat_x; 208 GLuint pp_txsize; /* npot only */ 209 GLuint pp_txpitch; /* npot only */ 210 GLuint pp_border_color; 211 GLuint pp_cubic_faces; /* cube face 1,2,3,4 log2 sizes */ 212 213 GLboolean border_fallback; 214}; 215 216static inline radeonTexObj* radeon_tex_obj(struct gl_texture_object *texObj) 217{ 218 return (radeonTexObj*)texObj; 219} 220 221/* occlusion query */ 222struct radeon_query_object { 223 struct gl_query_object Base; 224 struct radeon_bo *bo; 225 int curr_offset; 226 GLboolean emitted_begin; 227 228 /* Double linked list of not flushed query objects */ 229 struct radeon_query_object *prev, *next; 230}; 231 232/* Need refcounting on dma buffers: 233 */ 234struct radeon_dma_buffer { 235 int refcount; /* the number of retained regions in buf */ 236 drmBufPtr buf; 237}; 238 239struct radeon_aos { 240 struct radeon_bo *bo; /** Buffer object where vertex data is stored */ 241 int offset; /** Offset into buffer object, in bytes */ 242 int components; /** Number of components per vertex */ 243 int stride; /** Stride in dwords (may be 0 for repeating) */ 244 int count; /** Number of vertices */ 245}; 246 247#define DMA_BO_FREE_TIME 100 248 249struct radeon_dma_bo { 250 struct radeon_dma_bo *next, *prev; 251 struct radeon_bo *bo; 252 int expire_counter; 253}; 254 255struct radeon_dma { 256 /* Active dma region. Allocations for vertices and retained 257 * regions come from here. Also used for emitting random vertices, 258 * these may be flushed by calling flush_current(); 259 */ 260 struct radeon_dma_bo free; 261 struct radeon_dma_bo wait; 262 struct radeon_dma_bo reserved; 263 size_t current_used; /** Number of bytes allocated and forgotten about */ 264 size_t current_vertexptr; /** End of active vertex region */ 265 size_t minimum_size; 266 267 /** 268 * If current_vertexptr != current_used then flush must be non-zero. 269 * flush must be called before non-active vertex allocations can be 270 * performed. 271 */ 272 void (*flush) (struct gl_context *); 273}; 274 275/* radeon_swtcl.c 276 */ 277struct radeon_swtcl_info { 278 279 GLuint RenderIndex; 280 GLuint vertex_size; 281 GLubyte *verts; 282 283 /* Fallback rasterization functions 284 */ 285 GLuint hw_primitive; 286 GLenum render_primitive; 287 GLuint numverts; 288 289 struct tnl_attr_map vertex_attrs[VERT_ATTRIB_MAX]; 290 GLuint vertex_attr_count; 291 292 GLuint emit_prediction; 293 struct radeon_bo *bo; 294}; 295 296#define RADEON_MAX_AOS_ARRAYS 16 297struct radeon_tcl_info { 298 struct radeon_aos aos[RADEON_MAX_AOS_ARRAYS]; 299 GLuint aos_count; 300 struct radeon_bo *elt_dma_bo; /** Buffer object that contains element indices */ 301 int elt_dma_offset; /** Offset into this buffer object, in bytes */ 302}; 303 304struct radeon_ioctl { 305 GLuint vertex_offset; 306 GLuint vertex_max; 307 struct radeon_bo *bo; 308 GLuint vertex_size; 309}; 310 311#define RADEON_MAX_PRIMS 64 312 313struct radeon_prim { 314 GLuint start; 315 GLuint end; 316 GLuint prim; 317}; 318 319static inline GLuint radeonPackColor(GLuint cpp, 320 GLubyte r, GLubyte g, 321 GLubyte b, GLubyte a) 322{ 323 switch (cpp) { 324 case 2: 325 return PACK_COLOR_565(r, g, b); 326 case 4: 327 return PACK_COLOR_8888(a, r, g, b); 328 default: 329 return 0; 330 } 331} 332 333#define MAX_CMD_BUF_SZ (16*1024) 334 335#define MAX_DMA_BUF_SZ (64*1024) 336 337struct radeon_store { 338 GLuint statenr; 339 GLuint primnr; 340 char cmd_buf[MAX_CMD_BUF_SZ]; 341 int cmd_used; 342 int elts_start; 343}; 344 345typedef void (*radeon_tri_func) (radeonContextPtr, 346 radeonVertex *, 347 radeonVertex *, radeonVertex *); 348 349typedef void (*radeon_line_func) (radeonContextPtr, 350 radeonVertex *, radeonVertex *); 351 352typedef void (*radeon_point_func) (radeonContextPtr, radeonVertex *); 353 354#define RADEON_MAX_BOS 32 355struct radeon_state { 356 struct radeon_colorbuffer_state color; 357 struct radeon_depthbuffer_state depth; 358 struct radeon_scissor_state scissor; 359}; 360 361/** 362 * This structure holds the command buffer while it is being constructed. 363 * 364 * The first batch of commands in the buffer is always the state that needs 365 * to be re-emitted when the context is lost. This batch can be skipped 366 * otherwise. 367 */ 368struct radeon_cmdbuf { 369 struct radeon_cs_manager *csm; 370 struct radeon_cs *cs; 371 int size; /** # of dwords total */ 372 unsigned int flushing:1; /** whether we're currently in FlushCmdBufLocked */ 373}; 374 375struct radeon_context { 376 struct gl_context glCtx; /**< base class, must be first */ 377 __DRIcontext *driContext; /* DRI context */ 378 radeonScreenPtr radeonScreen; /* Screen private DRI data */ 379 380 /* Texture object bookkeeping 381 */ 382 int texture_depth; 383 float initialMaxAnisotropy; 384 uint32_t texture_row_align; 385 uint32_t texture_rect_row_align; 386 uint32_t texture_compressed_row_align; 387 388 struct radeon_dma dma; 389 struct radeon_hw_state hw; 390 /* Rasterization and vertex state: 391 */ 392 GLuint TclFallback; 393 GLuint Fallback; 394 GLuint NewGLState; 395 GLbitfield64 tnl_index_bitset; /* index of bits for last tnl_install_attrs */ 396 397 /* Drawable information */ 398 unsigned int lastStamp; 399 400 /* Busy waiting */ 401 GLuint do_usleeps; 402 GLuint do_irqs; 403 GLuint irqsEmitted; 404 drm_radeon_irq_wait_t iw; 405 406 /* Derived state - for r300 only */ 407 struct radeon_state state; 408 409 struct radeon_swtcl_info swtcl; 410 struct radeon_tcl_info tcl; 411 /* Configuration cache 412 */ 413 driOptionCache optionCache; 414 415 struct radeon_cmdbuf cmdbuf; 416 417 struct radeon_debug debug; 418 419 drm_clip_rect_t fboRect; 420 GLboolean front_cliprects; 421 422 /** 423 * Set if rendering has occurred to the drawable's front buffer. 424 * 425 * This is used in the DRI2 case to detect that glFlush should also copy 426 * the contents of the fake front buffer to the real front buffer. 427 */ 428 GLboolean front_buffer_dirty; 429 430 /** 431 * Track whether front-buffer rendering is currently enabled 432 * 433 * A separate flag is used to track this in order to support MRT more 434 * easily. 435 */ 436 GLboolean is_front_buffer_rendering; 437 438 /** 439 * Track whether front-buffer is the current read target. 440 * 441 * This is closely associated with is_front_buffer_rendering, but may 442 * be set separately. The DRI2 fake front buffer must be referenced 443 * either way. 444 */ 445 GLboolean is_front_buffer_reading; 446 447 struct { 448 struct radeon_query_object *current; 449 struct radeon_state_atom queryobj; 450 } query; 451 452 struct { 453 void (*swtcl_flush)(struct gl_context *ctx, uint32_t offset); 454 void (*pre_emit_state)(radeonContextPtr rmesa); 455 void (*fallback)(struct gl_context *ctx, GLuint bit, GLboolean mode); 456 void (*free_context)(struct gl_context *ctx); 457 void (*emit_query_finish)(radeonContextPtr radeon); 458 void (*update_scissor)(struct gl_context *ctx); 459 unsigned (*check_blit)(mesa_format mesa_format, uint32_t dst_pitch); 460 unsigned (*blit)(struct gl_context *ctx, 461 struct radeon_bo *src_bo, 462 intptr_t src_offset, 463 mesa_format src_mesaformat, 464 unsigned src_pitch, 465 unsigned src_width, 466 unsigned src_height, 467 unsigned src_x_offset, 468 unsigned src_y_offset, 469 struct radeon_bo *dst_bo, 470 intptr_t dst_offset, 471 mesa_format dst_mesaformat, 472 unsigned dst_pitch, 473 unsigned dst_width, 474 unsigned dst_height, 475 unsigned dst_x_offset, 476 unsigned dst_y_offset, 477 unsigned reg_width, 478 unsigned reg_height, 479 unsigned flip_y); 480 unsigned (*is_format_renderable)(mesa_format mesa_format); 481 GLboolean (*revalidate_all_buffers)(struct gl_context *ctx); 482 } vtbl; 483}; 484 485static inline radeonContextPtr RADEON_CONTEXT(struct gl_context *ctx) 486{ 487 return (radeonContextPtr) ctx; 488} 489 490static inline __DRIdrawable* radeon_get_drawable(radeonContextPtr radeon) 491{ 492 return radeon->driContext->driDrawablePriv; 493} 494 495static inline __DRIdrawable* radeon_get_readable(radeonContextPtr radeon) 496{ 497 return radeon->driContext->driReadablePriv; 498} 499 500extern const char const *radeonVendorString; 501 502const char *radeonGetRendererString(radeonScreenPtr radeonScreen); 503 504GLboolean radeonInitContext(radeonContextPtr radeon, 505 gl_api api, 506 struct dd_function_table* functions, 507 const struct gl_config * glVisual, 508 __DRIcontext * driContextPriv, 509 void *sharedContextPrivate); 510 511void radeonCleanupContext(radeonContextPtr radeon); 512GLboolean radeonUnbindContext(__DRIcontext * driContextPriv); 513void radeon_update_renderbuffers(__DRIcontext *context, __DRIdrawable *drawable, 514 GLboolean front_only); 515GLboolean radeonMakeCurrent(__DRIcontext * driContextPriv, 516 __DRIdrawable * driDrawPriv, 517 __DRIdrawable * driReadPriv); 518extern void radeonDestroyContext(__DRIcontext * driContextPriv); 519void radeon_prepare_render(radeonContextPtr radeon); 520 521#endif 522