1 /*
2  * Copyright 2003 Tungsten Graphics, inc.
3  * All Rights Reserved.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a
6  * copy of this software and associated documentation files (the "Software"),
7  * to deal in the Software without restriction, including without limitation
8  * on the rights to use, copy, modify, merge, publish, distribute, sub
9  * license, and/or sell copies of the Software, and to permit persons to whom
10  * the Software is furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice (including the next
13  * paragraph) shall be included in all copies or substantial portions of the
14  * Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.  IN NO EVENT SHALL
19  * TUNGSTEN GRAPHICS AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
20  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
21  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
22  * USE OR OTHER DEALINGS IN THE SOFTWARE.
23  *
24  * Authors:
25  *    Keith Whitwell <keithw@tungstengraphics.com>
26  */
27 
28 #include "main/glheader.h"
29 #include "main/context.h"
30 #include "main/colormac.h"
31 #include "swrast/s_chan.h"
32 #include "t_context.h"
33 #include "t_vertex.h"
34 
35 #define DBG 0
36 
37 /* Build and manage clipspace/ndc/window vertices.
38  */
39 
match_fastpath(struct tnl_clipspace * vtx,const struct tnl_clipspace_fastpath * fp)40 static GLboolean match_fastpath( struct tnl_clipspace *vtx,
41 				 const struct tnl_clipspace_fastpath *fp)
42 {
43    GLuint j;
44 
45    if (vtx->attr_count != fp->attr_count)
46       return GL_FALSE;
47 
48    for (j = 0; j < vtx->attr_count; j++)
49       if (vtx->attr[j].format != fp->attr[j].format ||
50 	  vtx->attr[j].inputsize != fp->attr[j].size ||
51 	  vtx->attr[j].vertoffset != fp->attr[j].offset)
52 	 return GL_FALSE;
53 
54    if (fp->match_strides) {
55       if (vtx->vertex_size != fp->vertex_size)
56 	 return GL_FALSE;
57 
58       for (j = 0; j < vtx->attr_count; j++)
59 	 if (vtx->attr[j].inputstride != fp->attr[j].stride)
60 	    return GL_FALSE;
61    }
62 
63    return GL_TRUE;
64 }
65 
search_fastpath_emit(struct tnl_clipspace * vtx)66 static GLboolean search_fastpath_emit( struct tnl_clipspace *vtx )
67 {
68    struct tnl_clipspace_fastpath *fp = vtx->fastpath;
69 
70    for ( ; fp ; fp = fp->next) {
71       if (match_fastpath(vtx, fp)) {
72          vtx->emit = fp->func;
73 	 return GL_TRUE;
74       }
75    }
76 
77    return GL_FALSE;
78 }
79 
_tnl_register_fastpath(struct tnl_clipspace * vtx,GLboolean match_strides)80 void _tnl_register_fastpath( struct tnl_clipspace *vtx,
81 			     GLboolean match_strides )
82 {
83    struct tnl_clipspace_fastpath *fastpath = CALLOC_STRUCT(tnl_clipspace_fastpath);
84    GLuint i;
85 
86    fastpath->vertex_size = vtx->vertex_size;
87    fastpath->attr_count = vtx->attr_count;
88    fastpath->match_strides = match_strides;
89    fastpath->func = vtx->emit;
90    fastpath->attr = (struct tnl_attr_type *)
91       malloc(vtx->attr_count * sizeof(fastpath->attr[0]));
92 
93    for (i = 0; i < vtx->attr_count; i++) {
94       fastpath->attr[i].format = vtx->attr[i].format;
95       fastpath->attr[i].stride = vtx->attr[i].inputstride;
96       fastpath->attr[i].size = vtx->attr[i].inputsize;
97       fastpath->attr[i].offset = vtx->attr[i].vertoffset;
98    }
99 
100    fastpath->next = vtx->fastpath;
101    vtx->fastpath = fastpath;
102 }
103 
104 
105 
106 /***********************************************************************
107  * Build codegen functions or return generic ones:
108  */
choose_emit_func(struct gl_context * ctx,GLuint count,GLubyte * dest)109 static void choose_emit_func( struct gl_context *ctx, GLuint count, GLubyte *dest)
110 {
111    struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
112    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
113    struct tnl_clipspace_attr *a = vtx->attr;
114    const GLuint attr_count = vtx->attr_count;
115    GLuint j;
116 
117    for (j = 0; j < attr_count; j++) {
118       GLvector4f *vptr = VB->AttribPtr[a[j].attrib];
119       a[j].inputstride = vptr->stride;
120       a[j].inputsize = vptr->size;
121       a[j].emit = a[j].insert[vptr->size - 1]; /* not always used */
122    }
123 
124    vtx->emit = NULL;
125 
126    /* Does this match an existing (hardwired, codegen or known-bad)
127     * fastpath?
128     */
129    if (search_fastpath_emit(vtx)) {
130       /* Use this result.  If it is null, then it is already known
131        * that the current state will fail for codegen and there is no
132        * point trying again.
133        */
134    }
135    else if (vtx->codegen_emit) {
136       vtx->codegen_emit(ctx);
137    }
138 
139    if (!vtx->emit) {
140       _tnl_generate_hardwired_emit(ctx);
141    }
142 
143    /* Otherwise use the generic version:
144     */
145    if (!vtx->emit)
146       vtx->emit = _tnl_generic_emit;
147 
148    vtx->emit( ctx, count, dest );
149 }
150 
151 
152 
choose_interp_func(struct gl_context * ctx,GLfloat t,GLuint edst,GLuint eout,GLuint ein,GLboolean force_boundary)153 static void choose_interp_func( struct gl_context *ctx,
154 				GLfloat t,
155 				GLuint edst, GLuint eout, GLuint ein,
156 				GLboolean force_boundary )
157 {
158    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
159 
160    if (vtx->need_extras &&
161        (ctx->_TriangleCaps & (DD_TRI_LIGHT_TWOSIDE|DD_TRI_UNFILLED))) {
162       vtx->interp = _tnl_generic_interp_extras;
163    } else {
164       vtx->interp = _tnl_generic_interp;
165    }
166 
167    vtx->interp( ctx, t, edst, eout, ein, force_boundary );
168 }
169 
170 
choose_copy_pv_func(struct gl_context * ctx,GLuint edst,GLuint esrc)171 static void choose_copy_pv_func(  struct gl_context *ctx, GLuint edst, GLuint esrc )
172 {
173    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
174 
175    if (vtx->need_extras &&
176        (ctx->_TriangleCaps & (DD_TRI_LIGHT_TWOSIDE|DD_TRI_UNFILLED))) {
177       vtx->copy_pv = _tnl_generic_copy_pv_extras;
178    } else {
179       vtx->copy_pv = _tnl_generic_copy_pv;
180    }
181 
182    vtx->copy_pv( ctx, edst, esrc );
183 }
184 
185 
186 /***********************************************************************
187  * Public entrypoints, mostly dispatch to the above:
188  */
189 
190 
191 /* Interpolate between two vertices to produce a third:
192  */
_tnl_interp(struct gl_context * ctx,GLfloat t,GLuint edst,GLuint eout,GLuint ein,GLboolean force_boundary)193 void _tnl_interp( struct gl_context *ctx,
194 		  GLfloat t,
195 		  GLuint edst, GLuint eout, GLuint ein,
196 		  GLboolean force_boundary )
197 {
198    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
199    vtx->interp( ctx, t, edst, eout, ein, force_boundary );
200 }
201 
202 /* Copy colors from one vertex to another:
203  */
_tnl_copy_pv(struct gl_context * ctx,GLuint edst,GLuint esrc)204 void _tnl_copy_pv(  struct gl_context *ctx, GLuint edst, GLuint esrc )
205 {
206    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
207    vtx->copy_pv( ctx, edst, esrc );
208 }
209 
210 
211 /* Extract a named attribute from a hardware vertex.  Will have to
212  * reverse any viewport transformation, swizzling or other conversions
213  * which may have been applied:
214  */
_tnl_get_attr(struct gl_context * ctx,const void * vin,GLenum attr,GLfloat * dest)215 void _tnl_get_attr( struct gl_context *ctx, const void *vin,
216 			      GLenum attr, GLfloat *dest )
217 {
218    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
219    const struct tnl_clipspace_attr *a = vtx->attr;
220    const GLuint attr_count = vtx->attr_count;
221    GLuint j;
222 
223    for (j = 0; j < attr_count; j++) {
224       if (a[j].attrib == attr) {
225 	 a[j].extract( &a[j], dest, (GLubyte *)vin + a[j].vertoffset );
226 	 return;
227       }
228    }
229 
230    /* Else return the value from ctx->Current.
231     */
232    if (attr == _TNL_ATTRIB_POINTSIZE) {
233       /* If the hardware vertex doesn't have point size then use size from
234        * struct gl_context.  XXX this will be wrong if drawing attenuated points!
235        */
236       dest[0] = ctx->Point.Size;
237    }
238    else {
239       memcpy( dest, ctx->Current.Attrib[attr], 4*sizeof(GLfloat));
240    }
241 }
242 
243 
244 /* Complementary operation to the above.
245  */
_tnl_set_attr(struct gl_context * ctx,void * vout,GLenum attr,const GLfloat * src)246 void _tnl_set_attr( struct gl_context *ctx, void *vout,
247 		    GLenum attr, const GLfloat *src )
248 {
249    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
250    const struct tnl_clipspace_attr *a = vtx->attr;
251    const GLuint attr_count = vtx->attr_count;
252    GLuint j;
253 
254    for (j = 0; j < attr_count; j++) {
255       if (a[j].attrib == attr) {
256 	 a[j].insert[4-1]( &a[j], (GLubyte *)vout + a[j].vertoffset, src );
257 	 return;
258       }
259    }
260 }
261 
262 
_tnl_get_vertex(struct gl_context * ctx,GLuint nr)263 void *_tnl_get_vertex( struct gl_context *ctx, GLuint nr )
264 {
265    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
266 
267    return vtx->vertex_buf + nr * vtx->vertex_size;
268 }
269 
_tnl_invalidate_vertex_state(struct gl_context * ctx,GLuint new_state)270 void _tnl_invalidate_vertex_state( struct gl_context *ctx, GLuint new_state )
271 {
272    /* if two-sided lighting changes or filled/unfilled polygon state changes */
273    if (new_state & (_NEW_LIGHT | _NEW_POLYGON) ) {
274       struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
275       vtx->new_inputs = ~0;
276       vtx->interp = choose_interp_func;
277       vtx->copy_pv = choose_copy_pv_func;
278    }
279 }
280 
invalidate_funcs(struct tnl_clipspace * vtx)281 static void invalidate_funcs( struct tnl_clipspace *vtx )
282 {
283    vtx->emit = choose_emit_func;
284    vtx->interp = choose_interp_func;
285    vtx->copy_pv = choose_copy_pv_func;
286    vtx->new_inputs = ~0;
287 }
288 
_tnl_install_attrs(struct gl_context * ctx,const struct tnl_attr_map * map,GLuint nr,const GLfloat * vp,GLuint unpacked_size)289 GLuint _tnl_install_attrs( struct gl_context *ctx, const struct tnl_attr_map *map,
290 			   GLuint nr, const GLfloat *vp,
291 			   GLuint unpacked_size )
292 {
293    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
294    GLuint offset = 0;
295    GLuint i, j;
296 
297    assert(nr < _TNL_ATTRIB_MAX);
298    assert(nr == 0 || map[0].attrib == VERT_ATTRIB_POS);
299 
300    vtx->new_inputs = ~0;
301    vtx->need_viewport = GL_FALSE;
302 
303    if (vp) {
304       vtx->need_viewport = GL_TRUE;
305    }
306 
307    for (j = 0, i = 0; i < nr; i++) {
308       const GLuint format = map[i].format;
309       if (format == EMIT_PAD) {
310 	 if (DBG)
311 	    printf("%d: pad %d, offset %d\n", i,
312 		   map[i].offset, offset);
313 
314 	 offset += map[i].offset;
315 
316       }
317       else {
318 	 GLuint tmpoffset;
319 
320 	 if (unpacked_size)
321 	    tmpoffset = map[i].offset;
322 	 else
323 	    tmpoffset = offset;
324 
325 	 if (vtx->attr_count != j ||
326 	     vtx->attr[j].attrib != map[i].attrib ||
327 	     vtx->attr[j].format != format ||
328 	     vtx->attr[j].vertoffset != tmpoffset) {
329 	    invalidate_funcs(vtx);
330 
331 	    vtx->attr[j].attrib = map[i].attrib;
332 	    vtx->attr[j].format = format;
333 	    vtx->attr[j].vp = vp;
334 	    vtx->attr[j].insert = _tnl_format_info[format].insert;
335 	    vtx->attr[j].extract = _tnl_format_info[format].extract;
336 	    vtx->attr[j].vertattrsize = _tnl_format_info[format].attrsize;
337 	    vtx->attr[j].vertoffset = tmpoffset;
338 	 }
339 
340 
341 	 if (DBG)
342 	    printf("%d: %s, vp %p, offset %d\n", i,
343 		   _tnl_format_info[format].name, (void *)vp,
344 		   vtx->attr[j].vertoffset);
345 
346 	 offset += _tnl_format_info[format].attrsize;
347 	 j++;
348       }
349    }
350 
351    vtx->attr_count = j;
352 
353    if (unpacked_size)
354       vtx->vertex_size = unpacked_size;
355    else
356       vtx->vertex_size = offset;
357 
358    assert(vtx->vertex_size <= vtx->max_vertex_size);
359    return vtx->vertex_size;
360 }
361 
362 
363 
_tnl_invalidate_vertices(struct gl_context * ctx,GLuint newinputs)364 void _tnl_invalidate_vertices( struct gl_context *ctx, GLuint newinputs )
365 {
366    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
367    vtx->new_inputs |= newinputs;
368 }
369 
370 
371 /* This event has broader use beyond this file - will move elsewhere
372  * and probably invoke a driver callback.
373  */
_tnl_notify_pipeline_output_change(struct gl_context * ctx)374 void _tnl_notify_pipeline_output_change( struct gl_context *ctx )
375 {
376    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
377    invalidate_funcs(vtx);
378 }
379 
380 
adjust_input_ptrs(struct gl_context * ctx,GLint diff)381 static void adjust_input_ptrs( struct gl_context *ctx, GLint diff)
382 {
383    struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
384    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
385    struct tnl_clipspace_attr *a = vtx->attr;
386    const GLuint count = vtx->attr_count;
387    GLuint j;
388 
389    diff -= 1;
390    for (j=0; j<count; ++j) {
391            register GLvector4f *vptr = VB->AttribPtr[a->attrib];
392 	   (a++)->inputptr += diff*vptr->stride;
393    }
394 }
395 
update_input_ptrs(struct gl_context * ctx,GLuint start)396 static void update_input_ptrs( struct gl_context *ctx, GLuint start )
397 {
398    struct vertex_buffer *VB = &TNL_CONTEXT(ctx)->vb;
399    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
400    struct tnl_clipspace_attr *a = vtx->attr;
401    const GLuint count = vtx->attr_count;
402    GLuint j;
403 
404    for (j = 0; j < count; j++) {
405       GLvector4f *vptr = VB->AttribPtr[a[j].attrib];
406 
407       if (vtx->emit != choose_emit_func) {
408 	 assert(a[j].inputstride == vptr->stride);
409 	 assert(a[j].inputsize == vptr->size);
410       }
411 
412       a[j].inputptr = ((GLubyte *)vptr->data) + start * vptr->stride;
413    }
414 
415    if (a->vp) {
416       vtx->vp_scale[0] = a->vp[MAT_SX];
417       vtx->vp_scale[1] = a->vp[MAT_SY];
418       vtx->vp_scale[2] = a->vp[MAT_SZ];
419       vtx->vp_scale[3] = 1.0;
420       vtx->vp_xlate[0] = a->vp[MAT_TX];
421       vtx->vp_xlate[1] = a->vp[MAT_TY];
422       vtx->vp_xlate[2] = a->vp[MAT_TZ];
423       vtx->vp_xlate[3] = 0.0;
424    }
425 }
426 
427 
_tnl_build_vertices(struct gl_context * ctx,GLuint start,GLuint end,GLuint newinputs)428 void _tnl_build_vertices( struct gl_context *ctx,
429 			  GLuint start,
430 			  GLuint end,
431 			  GLuint newinputs )
432 {
433    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
434    update_input_ptrs( ctx, start );
435    vtx->emit( ctx, end - start,
436 	      (GLubyte *)(vtx->vertex_buf +
437 			  start * vtx->vertex_size));
438 }
439 
440 /* Emit VB vertices start..end to dest.  Note that VB vertex at
441  * postion start will be emitted to dest at position zero.
442  */
_tnl_emit_vertices_to_buffer(struct gl_context * ctx,GLuint start,GLuint end,void * dest)443 void *_tnl_emit_vertices_to_buffer( struct gl_context *ctx,
444 				    GLuint start,
445 				    GLuint end,
446 				    void *dest )
447 {
448    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
449 
450    update_input_ptrs(ctx, start);
451    /* Note: dest should not be adjusted for non-zero 'start' values:
452     */
453    vtx->emit( ctx, end - start, (GLubyte*) dest );
454    return (void *)((GLubyte *)dest + vtx->vertex_size * (end - start));
455 }
456 
457 /* Emit indexed VB vertices start..end to dest.  Note that VB vertex at
458  * postion start will be emitted to dest at position zero.
459  */
460 
_tnl_emit_indexed_vertices_to_buffer(struct gl_context * ctx,const GLuint * elts,GLuint start,GLuint end,void * dest)461 void *_tnl_emit_indexed_vertices_to_buffer( struct gl_context *ctx,
462 					    const GLuint *elts,
463 					    GLuint start,
464 					    GLuint end,
465 					    void *dest )
466 {
467    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
468    GLuint oldIndex;
469    GLubyte *cdest = dest;
470 
471    update_input_ptrs(ctx, oldIndex = elts[start++]);
472    vtx->emit( ctx, 1, cdest );
473    cdest += vtx->vertex_size;
474 
475    for (; start < end; ++start) {
476       adjust_input_ptrs(ctx, elts[start] - oldIndex);
477       oldIndex = elts[start];
478       vtx->emit( ctx, 1, cdest);
479       cdest += vtx->vertex_size;
480    }
481 
482    return (void *) cdest;
483 }
484 
485 
_tnl_init_vertices(struct gl_context * ctx,GLuint vb_size,GLuint max_vertex_size)486 void _tnl_init_vertices( struct gl_context *ctx,
487 			GLuint vb_size,
488 			GLuint max_vertex_size )
489 {
490    struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
491 
492    _tnl_install_attrs( ctx, NULL, 0, NULL, 0 );
493 
494    vtx->need_extras = GL_TRUE;
495    if (max_vertex_size > vtx->max_vertex_size) {
496       _tnl_free_vertices( ctx );
497       vtx->max_vertex_size = max_vertex_size;
498       vtx->vertex_buf = (GLubyte *)_mesa_align_calloc(vb_size * max_vertex_size, 32 );
499       invalidate_funcs(vtx);
500    }
501 
502    switch(CHAN_TYPE) {
503    case GL_UNSIGNED_BYTE:
504       vtx->chan_scale[0] = 255.0;
505       vtx->chan_scale[1] = 255.0;
506       vtx->chan_scale[2] = 255.0;
507       vtx->chan_scale[3] = 255.0;
508       break;
509    case GL_UNSIGNED_SHORT:
510       vtx->chan_scale[0] = 65535.0;
511       vtx->chan_scale[1] = 65535.0;
512       vtx->chan_scale[2] = 65535.0;
513       vtx->chan_scale[3] = 65535.0;
514       break;
515    default:
516       vtx->chan_scale[0] = 1.0;
517       vtx->chan_scale[1] = 1.0;
518       vtx->chan_scale[2] = 1.0;
519       vtx->chan_scale[3] = 1.0;
520       break;
521    }
522 
523    vtx->identity[0] = 0.0;
524    vtx->identity[1] = 0.0;
525    vtx->identity[2] = 0.0;
526    vtx->identity[3] = 1.0;
527 
528    vtx->codegen_emit = NULL;
529 
530 #ifdef USE_SSE_ASM
531    if (!_mesa_getenv("MESA_NO_CODEGEN"))
532       vtx->codegen_emit = _tnl_generate_sse_emit;
533 #endif
534 }
535 
536 
_tnl_free_vertices(struct gl_context * ctx)537 void _tnl_free_vertices( struct gl_context *ctx )
538 {
539    TNLcontext *tnl = TNL_CONTEXT(ctx);
540    if (tnl) {
541       struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
542       struct tnl_clipspace_fastpath *fp, *tmp;
543 
544       if (vtx->vertex_buf) {
545          _mesa_align_free(vtx->vertex_buf);
546          vtx->vertex_buf = NULL;
547       }
548 
549       for (fp = vtx->fastpath ; fp ; fp = tmp) {
550          tmp = fp->next;
551          FREE(fp->attr);
552 
553          /* KW: At the moment, fp->func is constrained to be allocated by
554           * _mesa_exec_alloc(), as the hardwired fastpaths in
555           * t_vertex_generic.c are handled specially.  It would be nice
556           * to unify them, but this probably won't change until this
557           * module gets another overhaul.
558           */
559          _mesa_exec_free((void *) fp->func);
560          FREE(fp);
561       }
562 
563       vtx->fastpath = NULL;
564    }
565 }
566