1/*
2 * Copyright 2003 Tungsten Graphics, inc.
3 * All Rights Reserved.
4 *
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the "Software"),
7 * to deal in the Software without restriction, including without limitation
8 * on the rights to use, copy, modify, merge, publish, distribute, sub
9 * license, and/or sell copies of the Software, and to permit persons to whom
10 * the Software is furnished to do so, subject to the following conditions:
11 *
12 * The above copyright notice and this permission notice (including the next
13 * paragraph) shall be included in all copies or substantial portions of the
14 * Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.  IN NO EVENT SHALL
19 * TUNGSTEN GRAPHICS AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
20 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
21 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
22 * USE OR OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors:
25 *    Keith Whitwell <keithw@tungstengraphics.com>
26 */
27
28#include "main/glheader.h"
29#include "main/context.h"
30#include "main/colormac.h"
31#include "swrast/s_chan.h"
32#include "t_context.h"
33#include "t_vertex.h"
34
35#define DBG 0
36
37/* Build and manage clipspace/ndc/window vertices.
38 */
39
40static GLboolean match_fastpath( struct tnl_clipspace *vtx,
41				 const struct tnl_clipspace_fastpath *fp)
42{
43   GLuint j;
44
45   if (vtx->attr_count != fp->attr_count)
46      return GL_FALSE;
47
48   for (j = 0; j < vtx->attr_count; j++)
49      if (vtx->attr[j].format != fp->attr[j].format ||
50	  vtx->attr[j].inputsize != fp->attr[j].size ||
51	  vtx->attr[j].vertoffset != fp->attr[j].offset)
52	 return GL_FALSE;
53
54   if (fp->match_strides) {
55      if (vtx->vertex_size != fp->vertex_size)
56	 return GL_FALSE;
57
58      for (j = 0; j < vtx->attr_count; j++)
59	 if (vtx->attr[j].inputstride != fp->attr[j].stride)
60	    return GL_FALSE;
61   }
62
63   return GL_TRUE;
64}
65
66static GLboolean search_fastpath_emit( struct tnl_clipspace *vtx )
67{
68   struct tnl_clipspace_fastpath *fp = vtx->fastpath;
69
70   for ( ; fp ; fp = fp->next) {
71      if (match_fastpath(vtx, fp)) {
72         vtx->emit = fp->func;
73	 return GL_TRUE;
74      }
75   }
76
77   return GL_FALSE;
78}
79
80void _tnl_register_fastpath( struct tnl_clipspace *vtx,
81			     GLboolean match_strides )
82{
83   struct tnl_clipspace_fastpath *fastpath = CALLOC_STRUCT(tnl_clipspace_fastpath);
84   GLuint i;
85
86   fastpath->vertex_size = vtx->vertex_size;
87   fastpath->attr_count = vtx->attr_count;
88   fastpath->match_strides = match_strides;
89   fastpath->func = vtx->emit;
90   fastpath->attr = (struct tnl_attr_type *)
91      malloc(vtx->attr_count * sizeof(fastpath->attr[0]));
92
93   for (i = 0; i < vtx->attr_count; i++) {
94      fastpath->attr[i].format = vtx->attr[i].format;
95      fastpath->attr[i].stride = vtx->attr[i].inputstride;
96      fastpath->attr[i].size = vtx->attr[i].inputsize;
97      fastpath->attr[i].offset = vtx->attr[i].vertoffset;
98   }
99
100   fastpath->next = vtx->fastpath;
101   vtx->fastpath = fastpath;
102}
103
104
105
106/***********************************************************************
107 * Build codegen functions or return generic ones:
108 */
109static void choose_emit_func( struct gl_context *ctx, GLuint count, GLubyte *dest)
110{
111   struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
112   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
113   struct tnl_clipspace_attr *a = vtx->attr;
114   const GLuint attr_count = vtx->attr_count;
115   GLuint j;
116
117   for (j = 0; j < attr_count; j++) {
118      GLvector4f *vptr = VB->AttribPtr[a[j].attrib];
119      a[j].inputstride = vptr->stride;
120      a[j].inputsize = vptr->size;
121      a[j].emit = a[j].insert[vptr->size - 1]; /* not always used */
122   }
123
124   vtx->emit = NULL;
125
126   /* Does this match an existing (hardwired, codegen or known-bad)
127    * fastpath?
128    */
129   if (search_fastpath_emit(vtx)) {
130      /* Use this result.  If it is null, then it is already known
131       * that the current state will fail for codegen and there is no
132       * point trying again.
133       */
134   }
135   else if (vtx->codegen_emit) {
136      vtx->codegen_emit(ctx);
137   }
138
139   if (!vtx->emit) {
140      _tnl_generate_hardwired_emit(ctx);
141   }
142
143   /* Otherwise use the generic version:
144    */
145   if (!vtx->emit)
146      vtx->emit = _tnl_generic_emit;
147
148   vtx->emit( ctx, count, dest );
149}
150
151
152
153static void choose_interp_func( struct gl_context *ctx,
154				GLfloat t,
155				GLuint edst, GLuint eout, GLuint ein,
156				GLboolean force_boundary )
157{
158   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
159
160   if (vtx->need_extras &&
161       (ctx->_TriangleCaps & (DD_TRI_LIGHT_TWOSIDE|DD_TRI_UNFILLED))) {
162      vtx->interp = _tnl_generic_interp_extras;
163   } else {
164      vtx->interp = _tnl_generic_interp;
165   }
166
167   vtx->interp( ctx, t, edst, eout, ein, force_boundary );
168}
169
170
171static void choose_copy_pv_func(  struct gl_context *ctx, GLuint edst, GLuint esrc )
172{
173   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
174
175   if (vtx->need_extras &&
176       (ctx->_TriangleCaps & (DD_TRI_LIGHT_TWOSIDE|DD_TRI_UNFILLED))) {
177      vtx->copy_pv = _tnl_generic_copy_pv_extras;
178   } else {
179      vtx->copy_pv = _tnl_generic_copy_pv;
180   }
181
182   vtx->copy_pv( ctx, edst, esrc );
183}
184
185
186/***********************************************************************
187 * Public entrypoints, mostly dispatch to the above:
188 */
189
190
191/* Interpolate between two vertices to produce a third:
192 */
193void _tnl_interp( struct gl_context *ctx,
194		  GLfloat t,
195		  GLuint edst, GLuint eout, GLuint ein,
196		  GLboolean force_boundary )
197{
198   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
199   vtx->interp( ctx, t, edst, eout, ein, force_boundary );
200}
201
202/* Copy colors from one vertex to another:
203 */
204void _tnl_copy_pv(  struct gl_context *ctx, GLuint edst, GLuint esrc )
205{
206   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
207   vtx->copy_pv( ctx, edst, esrc );
208}
209
210
211/* Extract a named attribute from a hardware vertex.  Will have to
212 * reverse any viewport transformation, swizzling or other conversions
213 * which may have been applied:
214 */
215void _tnl_get_attr( struct gl_context *ctx, const void *vin,
216			      GLenum attr, GLfloat *dest )
217{
218   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
219   const struct tnl_clipspace_attr *a = vtx->attr;
220   const GLuint attr_count = vtx->attr_count;
221   GLuint j;
222
223   for (j = 0; j < attr_count; j++) {
224      if (a[j].attrib == attr) {
225	 a[j].extract( &a[j], dest, (GLubyte *)vin + a[j].vertoffset );
226	 return;
227      }
228   }
229
230   /* Else return the value from ctx->Current.
231    */
232   if (attr == _TNL_ATTRIB_POINTSIZE) {
233      /* If the hardware vertex doesn't have point size then use size from
234       * struct gl_context.  XXX this will be wrong if drawing attenuated points!
235       */
236      dest[0] = ctx->Point.Size;
237   }
238   else {
239      memcpy( dest, ctx->Current.Attrib[attr], 4*sizeof(GLfloat));
240   }
241}
242
243
244/* Complementary operation to the above.
245 */
246void _tnl_set_attr( struct gl_context *ctx, void *vout,
247		    GLenum attr, const GLfloat *src )
248{
249   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
250   const struct tnl_clipspace_attr *a = vtx->attr;
251   const GLuint attr_count = vtx->attr_count;
252   GLuint j;
253
254   for (j = 0; j < attr_count; j++) {
255      if (a[j].attrib == attr) {
256	 a[j].insert[4-1]( &a[j], (GLubyte *)vout + a[j].vertoffset, src );
257	 return;
258      }
259   }
260}
261
262
263void *_tnl_get_vertex( struct gl_context *ctx, GLuint nr )
264{
265   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
266
267   return vtx->vertex_buf + nr * vtx->vertex_size;
268}
269
270void _tnl_invalidate_vertex_state( struct gl_context *ctx, GLuint new_state )
271{
272   /* if two-sided lighting changes or filled/unfilled polygon state changes */
273   if (new_state & (_NEW_LIGHT | _NEW_POLYGON) ) {
274      struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
275      vtx->new_inputs = ~0;
276      vtx->interp = choose_interp_func;
277      vtx->copy_pv = choose_copy_pv_func;
278   }
279}
280
281static void invalidate_funcs( struct tnl_clipspace *vtx )
282{
283   vtx->emit = choose_emit_func;
284   vtx->interp = choose_interp_func;
285   vtx->copy_pv = choose_copy_pv_func;
286   vtx->new_inputs = ~0;
287}
288
289GLuint _tnl_install_attrs( struct gl_context *ctx, const struct tnl_attr_map *map,
290			   GLuint nr, const GLfloat *vp,
291			   GLuint unpacked_size )
292{
293   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
294   GLuint offset = 0;
295   GLuint i, j;
296
297   assert(nr < _TNL_ATTRIB_MAX);
298   assert(nr == 0 || map[0].attrib == VERT_ATTRIB_POS);
299
300   vtx->new_inputs = ~0;
301   vtx->need_viewport = GL_FALSE;
302
303   if (vp) {
304      vtx->need_viewport = GL_TRUE;
305   }
306
307   for (j = 0, i = 0; i < nr; i++) {
308      const GLuint format = map[i].format;
309      if (format == EMIT_PAD) {
310	 if (DBG)
311	    printf("%d: pad %d, offset %d\n", i,
312		   map[i].offset, offset);
313
314	 offset += map[i].offset;
315
316      }
317      else {
318	 GLuint tmpoffset;
319
320	 if (unpacked_size)
321	    tmpoffset = map[i].offset;
322	 else
323	    tmpoffset = offset;
324
325	 if (vtx->attr_count != j ||
326	     vtx->attr[j].attrib != map[i].attrib ||
327	     vtx->attr[j].format != format ||
328	     vtx->attr[j].vertoffset != tmpoffset) {
329	    invalidate_funcs(vtx);
330
331	    vtx->attr[j].attrib = map[i].attrib;
332	    vtx->attr[j].format = format;
333	    vtx->attr[j].vp = vp;
334	    vtx->attr[j].insert = _tnl_format_info[format].insert;
335	    vtx->attr[j].extract = _tnl_format_info[format].extract;
336	    vtx->attr[j].vertattrsize = _tnl_format_info[format].attrsize;
337	    vtx->attr[j].vertoffset = tmpoffset;
338	 }
339
340
341	 if (DBG)
342	    printf("%d: %s, vp %p, offset %d\n", i,
343		   _tnl_format_info[format].name, (void *)vp,
344		   vtx->attr[j].vertoffset);
345
346	 offset += _tnl_format_info[format].attrsize;
347	 j++;
348      }
349   }
350
351   vtx->attr_count = j;
352
353   if (unpacked_size)
354      vtx->vertex_size = unpacked_size;
355   else
356      vtx->vertex_size = offset;
357
358   assert(vtx->vertex_size <= vtx->max_vertex_size);
359   return vtx->vertex_size;
360}
361
362
363
364void _tnl_invalidate_vertices( struct gl_context *ctx, GLuint newinputs )
365{
366   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
367   vtx->new_inputs |= newinputs;
368}
369
370
371/* This event has broader use beyond this file - will move elsewhere
372 * and probably invoke a driver callback.
373 */
374void _tnl_notify_pipeline_output_change( struct gl_context *ctx )
375{
376   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
377   invalidate_funcs(vtx);
378}
379
380
381static void adjust_input_ptrs( struct gl_context *ctx, GLint diff)
382{
383   struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
384   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
385   struct tnl_clipspace_attr *a = vtx->attr;
386   const GLuint count = vtx->attr_count;
387   GLuint j;
388
389   diff -= 1;
390   for (j=0; j<count; ++j) {
391           register GLvector4f *vptr = VB->AttribPtr[a->attrib];
392	   (a++)->inputptr += diff*vptr->stride;
393   }
394}
395
396static void update_input_ptrs( struct gl_context *ctx, GLuint start )
397{
398   struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
399   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
400   struct tnl_clipspace_attr *a = vtx->attr;
401   const GLuint count = vtx->attr_count;
402   GLuint j;
403
404   for (j = 0; j < count; j++) {
405      GLvector4f *vptr = VB->AttribPtr[a[j].attrib];
406
407      if (vtx->emit != choose_emit_func) {
408	 assert(a[j].inputstride == vptr->stride);
409	 assert(a[j].inputsize == vptr->size);
410      }
411
412      a[j].inputptr = ((GLubyte *)vptr->data) + start * vptr->stride;
413   }
414
415   if (a->vp) {
416      vtx->vp_scale[0] = a->vp[MAT_SX];
417      vtx->vp_scale[1] = a->vp[MAT_SY];
418      vtx->vp_scale[2] = a->vp[MAT_SZ];
419      vtx->vp_scale[3] = 1.0;
420      vtx->vp_xlate[0] = a->vp[MAT_TX];
421      vtx->vp_xlate[1] = a->vp[MAT_TY];
422      vtx->vp_xlate[2] = a->vp[MAT_TZ];
423      vtx->vp_xlate[3] = 0.0;
424   }
425}
426
427
428void _tnl_build_vertices( struct gl_context *ctx,
429			  GLuint start,
430			  GLuint end,
431			  GLuint newinputs )
432{
433   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
434   update_input_ptrs( ctx, start );
435   vtx->emit( ctx, end - start,
436	      (GLubyte *)(vtx->vertex_buf +
437			  start * vtx->vertex_size));
438}
439
440/* Emit VB vertices start..end to dest.  Note that VB vertex at
441 * postion start will be emitted to dest at position zero.
442 */
443void *_tnl_emit_vertices_to_buffer( struct gl_context *ctx,
444				    GLuint start,
445				    GLuint end,
446				    void *dest )
447{
448   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
449
450   update_input_ptrs(ctx, start);
451   /* Note: dest should not be adjusted for non-zero 'start' values:
452    */
453   vtx->emit( ctx, end - start, (GLubyte*) dest );
454   return (void *)((GLubyte *)dest + vtx->vertex_size * (end - start));
455}
456
457/* Emit indexed VB vertices start..end to dest.  Note that VB vertex at
458 * postion start will be emitted to dest at position zero.
459 */
460
461void *_tnl_emit_indexed_vertices_to_buffer( struct gl_context *ctx,
462					    const GLuint *elts,
463					    GLuint start,
464					    GLuint end,
465					    void *dest )
466{
467   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
468   GLuint oldIndex;
469   GLubyte *cdest = dest;
470
471   update_input_ptrs(ctx, oldIndex = elts[start++]);
472   vtx->emit( ctx, 1, cdest );
473   cdest += vtx->vertex_size;
474
475   for (; start < end; ++start) {
476      adjust_input_ptrs(ctx, elts[start] - oldIndex);
477      oldIndex = elts[start];
478      vtx->emit( ctx, 1, cdest);
479      cdest += vtx->vertex_size;
480   }
481
482   return (void *) cdest;
483}
484
485
486void _tnl_init_vertices( struct gl_context *ctx,
487			GLuint vb_size,
488			GLuint max_vertex_size )
489{
490   struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
491
492   _tnl_install_attrs( ctx, NULL, 0, NULL, 0 );
493
494   vtx->need_extras = GL_TRUE;
495   if (max_vertex_size > vtx->max_vertex_size) {
496      _tnl_free_vertices( ctx );
497      vtx->max_vertex_size = max_vertex_size;
498      vtx->vertex_buf = (GLubyte *)_mesa_align_calloc(vb_size * max_vertex_size, 32 );
499      invalidate_funcs(vtx);
500   }
501
502   switch(CHAN_TYPE) {
503   case GL_UNSIGNED_BYTE:
504      vtx->chan_scale[0] = 255.0;
505      vtx->chan_scale[1] = 255.0;
506      vtx->chan_scale[2] = 255.0;
507      vtx->chan_scale[3] = 255.0;
508      break;
509   case GL_UNSIGNED_SHORT:
510      vtx->chan_scale[0] = 65535.0;
511      vtx->chan_scale[1] = 65535.0;
512      vtx->chan_scale[2] = 65535.0;
513      vtx->chan_scale[3] = 65535.0;
514      break;
515   default:
516      vtx->chan_scale[0] = 1.0;
517      vtx->chan_scale[1] = 1.0;
518      vtx->chan_scale[2] = 1.0;
519      vtx->chan_scale[3] = 1.0;
520      break;
521   }
522
523   vtx->identity[0] = 0.0;
524   vtx->identity[1] = 0.0;
525   vtx->identity[2] = 0.0;
526   vtx->identity[3] = 1.0;
527
528   vtx->codegen_emit = NULL;
529
530#ifdef USE_SSE_ASM
531   if (!_mesa_getenv("MESA_NO_CODEGEN"))
532      vtx->codegen_emit = _tnl_generate_sse_emit;
533#endif
534}
535
536
537void _tnl_free_vertices( struct gl_context *ctx )
538{
539   TNLcontext *tnl = TNL_CONTEXT(ctx);
540   if (tnl) {
541      struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
542      struct tnl_clipspace_fastpath *fp, *tmp;
543
544      if (vtx->vertex_buf) {
545         _mesa_align_free(vtx->vertex_buf);
546         vtx->vertex_buf = NULL;
547      }
548
549      for (fp = vtx->fastpath ; fp ; fp = tmp) {
550         tmp = fp->next;
551         FREE(fp->attr);
552
553         /* KW: At the moment, fp->func is constrained to be allocated by
554          * _mesa_exec_alloc(), as the hardwired fastpaths in
555          * t_vertex_generic.c are handled specially.  It would be nice
556          * to unify them, but this probably won't change until this
557          * module gets another overhaul.
558          */
559         _mesa_exec_free((void *) fp->func);
560         FREE(fp);
561      }
562
563      vtx->fastpath = NULL;
564   }
565}
566