1 /**************************************************************************
3 Copyright 2002-2008 VMware, Inc.
7 Permission is hereby granted, free of charge, to any person obtaining a
8 copy of this software and associated documentation files (the "Software"),
9 to deal in the Software without restriction, including without limitation
10 on the rights to use, copy, modify, merge, publish, distribute, sub
11 license, and/or sell copies of the Software, and to permit persons to whom
12 the Software is furnished to do so, subject to the following conditions:
14 The above copyright notice and this permission notice (including the next
15 paragraph) shall be included in all copies or substantial portions of the
18 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21 VMWARE AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
22 DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23 OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24 USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Keith Whitwell <keithw@vmware.com>
33 #include "main/glheader.h"
34 #include "main/bufferobj.h"
35 #include "main/context.h"
36 #include "main/macros.h"
37 #include "main/vtxfmt.h"
38 #include "main/dlist.h"
39 #include "main/eval.h"
40 #include "main/state.h"
41 #include "main/light.h"
42 #include "main/api_arrayelt.h"
43 #include "main/draw_validate.h"
44 #include "main/dispatch.h"
45 #include "util/bitscan.h"
48 #include "vbo_private.h"
51 /** ID/name for immediate-mode VBO */
52 #define IMM_BUFFER_NAME 0xaabbccdd
56 vbo_reset_all_attr(struct vbo_exec_context
*exec
);
60 * Close off the last primitive, execute the buffer, restart the
61 * primitive. This is called when we fill a vertex buffer before
65 vbo_exec_wrap_buffers(struct vbo_exec_context
*exec
)
67 if (exec
->vtx
.prim_count
== 0) {
68 exec
->vtx
.copied
.nr
= 0;
69 exec
->vtx
.vert_count
= 0;
70 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
73 struct _mesa_prim
*last_prim
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
74 const GLuint last_begin
= last_prim
->begin
;
77 if (_mesa_inside_begin_end(exec
->ctx
)) {
78 last_prim
->count
= exec
->vtx
.vert_count
- last_prim
->start
;
81 last_count
= last_prim
->count
;
83 /* Special handling for wrapping GL_LINE_LOOP */
84 if (last_prim
->mode
== GL_LINE_LOOP
&&
87 /* draw this section of the incomplete line loop as a line strip */
88 last_prim
->mode
= GL_LINE_STRIP
;
89 if (!last_prim
->begin
) {
90 /* This is not the first section of the line loop, so don't
91 * draw the 0th vertex. We're saving it until we draw the
92 * very last section of the loop.
99 /* Execute the buffer and save copied vertices.
101 if (exec
->vtx
.vert_count
)
102 vbo_exec_vtx_flush(exec
);
104 exec
->vtx
.prim_count
= 0;
105 exec
->vtx
.copied
.nr
= 0;
108 /* Emit a glBegin to start the new list.
110 assert(exec
->vtx
.prim_count
== 0);
112 if (_mesa_inside_begin_end(exec
->ctx
)) {
113 exec
->vtx
.prim
[0].mode
= exec
->ctx
->Driver
.CurrentExecPrimitive
;
114 exec
->vtx
.prim
[0].begin
= 0;
115 exec
->vtx
.prim
[0].end
= 0;
116 exec
->vtx
.prim
[0].start
= 0;
117 exec
->vtx
.prim
[0].count
= 0;
118 exec
->vtx
.prim_count
++;
120 if (exec
->vtx
.copied
.nr
== last_count
)
121 exec
->vtx
.prim
[0].begin
= last_begin
;
128 * Deal with buffer wrapping where provoked by the vertex buffer
129 * filling up, as opposed to upgrade_vertex().
132 vbo_exec_vtx_wrap(struct vbo_exec_context
*exec
)
134 unsigned numComponents
;
136 /* Run pipeline on current vertices, copy wrapped vertices
137 * to exec->vtx.copied.
139 vbo_exec_wrap_buffers(exec
);
141 if (!exec
->vtx
.buffer_ptr
) {
142 /* probably ran out of memory earlier when allocating the VBO */
146 /* Copy stored stored vertices to start of new list.
148 assert(exec
->vtx
.max_vert
- exec
->vtx
.vert_count
> exec
->vtx
.copied
.nr
);
150 numComponents
= exec
->vtx
.copied
.nr
* exec
->vtx
.vertex_size
;
151 memcpy(exec
->vtx
.buffer_ptr
,
152 exec
->vtx
.copied
.buffer
,
153 numComponents
* sizeof(fi_type
));
154 exec
->vtx
.buffer_ptr
+= numComponents
;
155 exec
->vtx
.vert_count
+= exec
->vtx
.copied
.nr
;
157 exec
->vtx
.copied
.nr
= 0;
162 * Copy the active vertex's values to the ctx->Current fields.
165 vbo_exec_copy_to_current(struct vbo_exec_context
*exec
)
167 struct gl_context
*ctx
= exec
->ctx
;
168 struct vbo_context
*vbo
= vbo_context(ctx
);
169 GLbitfield64 enabled
= exec
->vtx
.enabled
& (~BITFIELD64_BIT(VBO_ATTRIB_POS
));
172 const int i
= u_bit_scan64(&enabled
);
174 /* Note: the exec->vtx.current[i] pointers point into the
175 * ctx->Current.Attrib and ctx->Light.Material.Attrib arrays.
177 GLfloat
*current
= (GLfloat
*)vbo
->current
[i
].Ptr
;
178 fi_type tmp
[8]; /* space for doubles */
181 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
182 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
)
185 assert(exec
->vtx
.attr
[i
].size
);
187 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
188 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
) {
189 memset(tmp
, 0, sizeof(tmp
));
190 memcpy(tmp
, exec
->vtx
.attrptr
[i
], exec
->vtx
.attr
[i
].size
* sizeof(GLfloat
));
192 COPY_CLEAN_4V_TYPE_AS_UNION(tmp
,
193 exec
->vtx
.attr
[i
].size
,
194 exec
->vtx
.attrptr
[i
],
195 exec
->vtx
.attr
[i
].type
);
198 if (exec
->vtx
.attr
[i
].type
!= vbo
->current
[i
].Format
.Type
||
199 memcmp(current
, tmp
, 4 * sizeof(GLfloat
) * dmul
) != 0) {
200 memcpy(current
, tmp
, 4 * sizeof(GLfloat
) * dmul
);
202 /* Given that we explicitly state size here, there is no need
203 * for the COPY_CLEAN above, could just copy 16 bytes and be
204 * done. The only problem is when Mesa accesses ctx->Current
207 /* Size here is in components - not bytes */
208 vbo_set_vertex_format(&vbo
->current
[i
].Format
,
209 exec
->vtx
.attr
[i
].size
/ dmul
,
210 exec
->vtx
.attr
[i
].type
);
212 /* This triggers rather too much recalculation of Mesa state
213 * that doesn't get used (eg light positions).
215 if (i
>= VBO_ATTRIB_MAT_FRONT_AMBIENT
&&
216 i
<= VBO_ATTRIB_MAT_BACK_INDEXES
)
217 ctx
->NewState
|= _NEW_LIGHT
;
219 ctx
->NewState
|= _NEW_CURRENT_ATTRIB
;
223 /* Colormaterial -- this kindof sucks.
225 if (ctx
->Light
.ColorMaterialEnabled
&&
226 exec
->vtx
.attr
[VBO_ATTRIB_COLOR0
].size
) {
227 _mesa_update_color_material(ctx
,
228 ctx
->Current
.Attrib
[VBO_ATTRIB_COLOR0
]);
234 * Copy current vertex attribute values into the current vertex.
237 vbo_exec_copy_from_current(struct vbo_exec_context
*exec
)
239 struct gl_context
*ctx
= exec
->ctx
;
240 struct vbo_context
*vbo
= vbo_context(ctx
);
243 for (i
= VBO_ATTRIB_POS
+ 1; i
< VBO_ATTRIB_MAX
; i
++) {
244 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
245 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
) {
246 memcpy(exec
->vtx
.attrptr
[i
], vbo
->current
[i
].Ptr
,
247 exec
->vtx
.attr
[i
].size
* sizeof(GLfloat
));
249 const fi_type
*current
= (fi_type
*) vbo
->current
[i
].Ptr
;
250 switch (exec
->vtx
.attr
[i
].size
) {
251 case 4: exec
->vtx
.attrptr
[i
][3] = current
[3];
252 case 3: exec
->vtx
.attrptr
[i
][2] = current
[2];
253 case 2: exec
->vtx
.attrptr
[i
][1] = current
[1];
254 case 1: exec
->vtx
.attrptr
[i
][0] = current
[0];
263 * Flush existing data, set new attrib size, replay copied vertices.
264 * This is called when we transition from a small vertex attribute size
265 * to a larger one. Ex: glTexCoord2f -> glTexCoord4f.
266 * We need to go back over the previous 2-component texcoords and insert
267 * zero and one values.
268 * \param attr VBO_ATTRIB_x vertex attribute value
271 vbo_exec_wrap_upgrade_vertex(struct vbo_exec_context
*exec
,
272 GLuint attr
, GLuint newSize
)
274 struct gl_context
*ctx
= exec
->ctx
;
275 struct vbo_context
*vbo
= vbo_context(ctx
);
276 const GLint lastcount
= exec
->vtx
.vert_count
;
277 fi_type
*old_attrptr
[VBO_ATTRIB_MAX
];
278 const GLuint old_vtx_size
= exec
->vtx
.vertex_size
; /* floats per vertex */
279 const GLuint oldSize
= exec
->vtx
.attr
[attr
].size
;
282 assert(attr
< VBO_ATTRIB_MAX
);
284 /* Run pipeline on current vertices, copy wrapped vertices
285 * to exec->vtx.copied.
287 vbo_exec_wrap_buffers(exec
);
289 if (unlikely(exec
->vtx
.copied
.nr
)) {
290 /* We're in the middle of a primitive, keep the old vertex
291 * format around to be able to translate the copied vertices to
294 memcpy(old_attrptr
, exec
->vtx
.attrptr
, sizeof(old_attrptr
));
297 if (unlikely(oldSize
)) {
298 /* Do a COPY_TO_CURRENT to ensure back-copying works for the
299 * case when the attribute already exists in the vertex and is
300 * having its size increased.
302 vbo_exec_copy_to_current(exec
);
305 /* Heuristic: Attempt to isolate attributes received outside
306 * begin/end so that they don't bloat the vertices.
308 if (!_mesa_inside_begin_end(ctx
) &&
309 !oldSize
&& lastcount
> 8 && exec
->vtx
.vertex_size
) {
310 vbo_exec_copy_to_current(exec
);
311 vbo_reset_all_attr(exec
);
316 exec
->vtx
.attr
[attr
].size
= newSize
;
317 exec
->vtx
.vertex_size
+= newSize
- oldSize
;
318 exec
->vtx
.max_vert
= vbo_compute_max_verts(exec
);
319 exec
->vtx
.vert_count
= 0;
320 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
321 exec
->vtx
.enabled
|= BITFIELD64_BIT(attr
);
323 if (unlikely(oldSize
)) {
324 /* Size changed, recalculate all the attrptr[] values
326 fi_type
*tmp
= exec
->vtx
.vertex
;
328 for (i
= 0 ; i
< VBO_ATTRIB_MAX
; i
++) {
329 if (exec
->vtx
.attr
[i
].size
) {
330 exec
->vtx
.attrptr
[i
] = tmp
;
331 tmp
+= exec
->vtx
.attr
[i
].size
;
334 exec
->vtx
.attrptr
[i
] = NULL
; /* will not be dereferenced */
337 /* Copy from current to repopulate the vertex with correct
340 vbo_exec_copy_from_current(exec
);
343 /* Just have to append the new attribute at the end */
344 exec
->vtx
.attrptr
[attr
] = exec
->vtx
.vertex
+
345 exec
->vtx
.vertex_size
- newSize
;
348 /* Replay stored vertices to translate them
349 * to new format here.
351 * -- No need to replay - just copy piecewise
353 if (unlikely(exec
->vtx
.copied
.nr
)) {
354 fi_type
*data
= exec
->vtx
.copied
.buffer
;
355 fi_type
*dest
= exec
->vtx
.buffer_ptr
;
357 assert(exec
->vtx
.buffer_ptr
== exec
->vtx
.buffer_map
);
359 for (i
= 0 ; i
< exec
->vtx
.copied
.nr
; i
++) {
360 GLbitfield64 enabled
= exec
->vtx
.enabled
;
362 const int j
= u_bit_scan64(&enabled
);
363 GLuint sz
= exec
->vtx
.attr
[j
].size
;
364 GLint old_offset
= old_attrptr
[j
] - exec
->vtx
.vertex
;
365 GLint new_offset
= exec
->vtx
.attrptr
[j
] - exec
->vtx
.vertex
;
372 COPY_CLEAN_4V_TYPE_AS_UNION(tmp
, oldSize
,
374 exec
->vtx
.attr
[j
].type
);
375 COPY_SZ_4V(dest
+ new_offset
, newSize
, tmp
);
377 fi_type
*current
= (fi_type
*)vbo
->current
[j
].Ptr
;
378 COPY_SZ_4V(dest
+ new_offset
, sz
, current
);
382 COPY_SZ_4V(dest
+ new_offset
, sz
, data
+ old_offset
);
386 data
+= old_vtx_size
;
387 dest
+= exec
->vtx
.vertex_size
;
390 exec
->vtx
.buffer_ptr
= dest
;
391 exec
->vtx
.vert_count
+= exec
->vtx
.copied
.nr
;
392 exec
->vtx
.copied
.nr
= 0;
398 * This is when a vertex attribute transitions to a different size.
399 * For example, we saw a bunch of glTexCoord2f() calls and now we got a
400 * glTexCoord4f() call. We promote the array from size=2 to size=4.
401 * \param newSize size of new vertex (number of 32-bit words).
402 * \param attr VBO_ATTRIB_x vertex attribute value
405 vbo_exec_fixup_vertex(struct gl_context
*ctx
, GLuint attr
,
406 GLuint newSize
, GLenum newType
)
408 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
410 assert(attr
< VBO_ATTRIB_MAX
);
412 if (newSize
> exec
->vtx
.attr
[attr
].size
||
413 newType
!= exec
->vtx
.attr
[attr
].type
) {
414 /* New size is larger. Need to flush existing vertices and get
415 * an enlarged vertex format.
417 vbo_exec_wrap_upgrade_vertex(exec
, attr
, newSize
);
419 else if (newSize
< exec
->vtx
.attr
[attr
].active_size
) {
422 vbo_get_default_vals_as_union(exec
->vtx
.attr
[attr
].type
);
424 /* New size is smaller - just need to fill in some
425 * zeros. Don't need to flush or wrap.
427 for (i
= newSize
; i
<= exec
->vtx
.attr
[attr
].size
; i
++)
428 exec
->vtx
.attrptr
[attr
][i
-1] = id
[i
-1];
431 exec
->vtx
.attr
[attr
].active_size
= newSize
;
432 exec
->vtx
.attr
[attr
].type
= newType
;
434 /* Does setting NeedFlush belong here? Necessitates resetting
435 * vtxfmt on each flush (otherwise flags won't get reset
439 ctx
->Driver
.NeedFlush
|= FLUSH_STORED_VERTICES
;
444 * If index=0, does glVertexAttrib*() alias glVertex() to emit a vertex?
445 * It depends on a few things, including whether we're inside or outside
449 is_vertex_position(const struct gl_context
*ctx
, GLuint index
)
451 return (index
== 0 &&
452 _mesa_attr_zero_aliases_vertex(ctx
) &&
453 _mesa_inside_begin_end(ctx
));
458 * This macro is used to implement all the glVertex, glColor, glTexCoord,
459 * glVertexAttrib, etc functions.
460 * \param A VBO_ATTRIB_x attribute index
461 * \param N attribute size (1..4)
462 * \param T type (GL_FLOAT, GL_DOUBLE, GL_INT, GL_UNSIGNED_INT)
463 * \param C cast type (uint32_t or uint64_t)
464 * \param V0, V1, v2, V3 attribute value
466 #define ATTR_UNION(A, N, T, C, V0, V1, V2, V3) \
468 struct vbo_exec_context *exec = &vbo_context(ctx)->exec; \
469 int sz = (sizeof(C) / sizeof(GLfloat)); \
471 assert(sz == 1 || sz == 2); \
473 /* check if attribute size or type is changing */ \
474 if (unlikely(exec->vtx.attr[A].active_size != N * sz) || \
475 unlikely(exec->vtx.attr[A].type != T)) { \
476 vbo_exec_fixup_vertex(ctx, A, N * sz, T); \
479 /* store vertex attribute in vertex buffer */ \
481 C *dest = (C *)exec->vtx.attrptr[A]; \
482 if (N>0) dest[0] = V0; \
483 if (N>1) dest[1] = V1; \
484 if (N>2) dest[2] = V2; \
485 if (N>3) dest[3] = V3; \
486 assert(exec->vtx.attr[A].type == T); \
490 /* This is a glVertex call */ \
493 /* copy 32-bit words */ \
494 for (i = 0; i < exec->vtx.vertex_size; i++) \
495 exec->vtx.buffer_ptr[i] = exec->vtx.vertex[i]; \
497 exec->vtx.buffer_ptr += exec->vtx.vertex_size; \
499 /* Set FLUSH_STORED_VERTICES to indicate that there's now */ \
500 /* something to draw (not just updating a color or texcoord).*/ \
501 /* Don't set FLUSH_UPDATE_CURRENT because */ \
502 /* Current.Attrib[VBO_ATTRIB_POS] is never used. */ \
503 ctx->Driver.NeedFlush |= FLUSH_STORED_VERTICES; \
505 if (++exec->vtx.vert_count >= exec->vtx.max_vert) \
506 vbo_exec_vtx_wrap(exec); \
508 /* we now have accumulated per-vertex attributes */ \
509 ctx->Driver.NeedFlush |= FLUSH_UPDATE_CURRENT; \
515 #define ERROR(err) _mesa_error(ctx, err, __func__)
516 #define TAG(x) vbo_exec_##x
518 #include "vbo_attrib_tmp.h"
523 * Execute a glMaterial call. Note that if GL_COLOR_MATERIAL is enabled,
524 * this may be a (partial) no-op.
526 static void GLAPIENTRY
527 vbo_exec_Materialfv(GLenum face
, GLenum pname
, const GLfloat
*params
)
529 GLbitfield updateMats
;
530 GET_CURRENT_CONTEXT(ctx
);
532 /* This function should be a no-op when it tries to update material
533 * attributes which are currently tracking glColor via glColorMaterial.
534 * The updateMats var will be a mask of the MAT_BIT_FRONT/BACK_x bits
535 * indicating which material attributes can actually be updated below.
537 if (ctx
->Light
.ColorMaterialEnabled
) {
538 updateMats
= ~ctx
->Light
._ColorMaterialBitmask
;
541 /* GL_COLOR_MATERIAL is disabled so don't skip any material updates */
542 updateMats
= ALL_MATERIAL_BITS
;
545 if (ctx
->API
== API_OPENGL_COMPAT
&& face
== GL_FRONT
) {
546 updateMats
&= FRONT_MATERIAL_BITS
;
548 else if (ctx
->API
== API_OPENGL_COMPAT
&& face
== GL_BACK
) {
549 updateMats
&= BACK_MATERIAL_BITS
;
551 else if (face
!= GL_FRONT_AND_BACK
) {
552 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterial(invalid face)");
558 if (updateMats
& MAT_BIT_FRONT_EMISSION
)
559 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_EMISSION
, 4, params
);
560 if (updateMats
& MAT_BIT_BACK_EMISSION
)
561 MAT_ATTR(VBO_ATTRIB_MAT_BACK_EMISSION
, 4, params
);
564 if (updateMats
& MAT_BIT_FRONT_AMBIENT
)
565 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT
, 4, params
);
566 if (updateMats
& MAT_BIT_BACK_AMBIENT
)
567 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT
, 4, params
);
570 if (updateMats
& MAT_BIT_FRONT_DIFFUSE
)
571 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE
, 4, params
);
572 if (updateMats
& MAT_BIT_BACK_DIFFUSE
)
573 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE
, 4, params
);
576 if (updateMats
& MAT_BIT_FRONT_SPECULAR
)
577 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SPECULAR
, 4, params
);
578 if (updateMats
& MAT_BIT_BACK_SPECULAR
)
579 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SPECULAR
, 4, params
);
582 if (*params
< 0 || *params
> ctx
->Const
.MaxShininess
) {
583 _mesa_error(ctx
, GL_INVALID_VALUE
,
584 "glMaterial(invalid shininess: %f out range [0, %f])",
585 *params
, ctx
->Const
.MaxShininess
);
588 if (updateMats
& MAT_BIT_FRONT_SHININESS
)
589 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SHININESS
, 1, params
);
590 if (updateMats
& MAT_BIT_BACK_SHININESS
)
591 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SHININESS
, 1, params
);
593 case GL_COLOR_INDEXES
:
594 if (ctx
->API
!= API_OPENGL_COMPAT
) {
595 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterialfv(pname)");
598 if (updateMats
& MAT_BIT_FRONT_INDEXES
)
599 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_INDEXES
, 3, params
);
600 if (updateMats
& MAT_BIT_BACK_INDEXES
)
601 MAT_ATTR(VBO_ATTRIB_MAT_BACK_INDEXES
, 3, params
);
603 case GL_AMBIENT_AND_DIFFUSE
:
604 if (updateMats
& MAT_BIT_FRONT_AMBIENT
)
605 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT
, 4, params
);
606 if (updateMats
& MAT_BIT_FRONT_DIFFUSE
)
607 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE
, 4, params
);
608 if (updateMats
& MAT_BIT_BACK_AMBIENT
)
609 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT
, 4, params
);
610 if (updateMats
& MAT_BIT_BACK_DIFFUSE
)
611 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE
, 4, params
);
614 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterialfv(pname)");
621 * Flush (draw) vertices.
624 vbo_exec_FlushVertices_internal(struct vbo_exec_context
*exec
)
626 if (exec
->vtx
.vert_count
) {
627 vbo_exec_vtx_flush(exec
);
630 if (exec
->vtx
.vertex_size
) {
631 vbo_exec_copy_to_current(exec
);
632 vbo_reset_all_attr(exec
);
637 static void GLAPIENTRY
638 vbo_exec_EvalCoord1f(GLfloat u
)
640 GET_CURRENT_CONTEXT(ctx
);
641 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
645 if (exec
->eval
.recalculate_maps
)
646 vbo_exec_eval_update(exec
);
648 for (i
= 0; i
<= VBO_ATTRIB_TEX7
; i
++) {
649 if (exec
->eval
.map1
[i
].map
)
650 if (exec
->vtx
.attr
[i
].active_size
!= exec
->eval
.map1
[i
].sz
)
651 vbo_exec_fixup_vertex(ctx
, i
, exec
->eval
.map1
[i
].sz
, GL_FLOAT
);
655 memcpy(exec
->vtx
.copied
.buffer
, exec
->vtx
.vertex
,
656 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
658 vbo_exec_do_EvalCoord1f(exec
, u
);
660 memcpy(exec
->vtx
.vertex
, exec
->vtx
.copied
.buffer
,
661 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
665 static void GLAPIENTRY
666 vbo_exec_EvalCoord2f(GLfloat u
, GLfloat v
)
668 GET_CURRENT_CONTEXT(ctx
);
669 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
673 if (exec
->eval
.recalculate_maps
)
674 vbo_exec_eval_update(exec
);
676 for (i
= 0; i
<= VBO_ATTRIB_TEX7
; i
++) {
677 if (exec
->eval
.map2
[i
].map
)
678 if (exec
->vtx
.attr
[i
].active_size
!= exec
->eval
.map2
[i
].sz
)
679 vbo_exec_fixup_vertex(ctx
, i
, exec
->eval
.map2
[i
].sz
, GL_FLOAT
);
682 if (ctx
->Eval
.AutoNormal
)
683 if (exec
->vtx
.attr
[VBO_ATTRIB_NORMAL
].active_size
!= 3)
684 vbo_exec_fixup_vertex(ctx
, VBO_ATTRIB_NORMAL
, 3, GL_FLOAT
);
687 memcpy(exec
->vtx
.copied
.buffer
, exec
->vtx
.vertex
,
688 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
690 vbo_exec_do_EvalCoord2f(exec
, u
, v
);
692 memcpy(exec
->vtx
.vertex
, exec
->vtx
.copied
.buffer
,
693 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
697 static void GLAPIENTRY
698 vbo_exec_EvalCoord1fv(const GLfloat
*u
)
700 vbo_exec_EvalCoord1f(u
[0]);
704 static void GLAPIENTRY
705 vbo_exec_EvalCoord2fv(const GLfloat
*u
)
707 vbo_exec_EvalCoord2f(u
[0], u
[1]);
711 static void GLAPIENTRY
712 vbo_exec_EvalPoint1(GLint i
)
714 GET_CURRENT_CONTEXT(ctx
);
715 GLfloat du
= ((ctx
->Eval
.MapGrid1u2
- ctx
->Eval
.MapGrid1u1
) /
716 (GLfloat
) ctx
->Eval
.MapGrid1un
);
717 GLfloat u
= i
* du
+ ctx
->Eval
.MapGrid1u1
;
719 vbo_exec_EvalCoord1f(u
);
723 static void GLAPIENTRY
724 vbo_exec_EvalPoint2(GLint i
, GLint j
)
726 GET_CURRENT_CONTEXT(ctx
);
727 GLfloat du
= ((ctx
->Eval
.MapGrid2u2
- ctx
->Eval
.MapGrid2u1
) /
728 (GLfloat
) ctx
->Eval
.MapGrid2un
);
729 GLfloat dv
= ((ctx
->Eval
.MapGrid2v2
- ctx
->Eval
.MapGrid2v1
) /
730 (GLfloat
) ctx
->Eval
.MapGrid2vn
);
731 GLfloat u
= i
* du
+ ctx
->Eval
.MapGrid2u1
;
732 GLfloat v
= j
* dv
+ ctx
->Eval
.MapGrid2v1
;
734 vbo_exec_EvalCoord2f(u
, v
);
739 * Called via glBegin.
741 static void GLAPIENTRY
742 vbo_exec_Begin(GLenum mode
)
744 GET_CURRENT_CONTEXT(ctx
);
745 struct vbo_context
*vbo
= vbo_context(ctx
);
746 struct vbo_exec_context
*exec
= &vbo
->exec
;
749 if (_mesa_inside_begin_end(ctx
)) {
750 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glBegin");
754 if (!_mesa_valid_prim_mode(ctx
, mode
, "glBegin")) {
758 if (!_mesa_valid_to_render(ctx
, "glBegin")) {
762 /* Heuristic: attempt to isolate attributes occurring outside
765 if (exec
->vtx
.vertex_size
&& !exec
->vtx
.attr
[VBO_ATTRIB_POS
].size
)
766 vbo_exec_FlushVertices_internal(exec
);
768 i
= exec
->vtx
.prim_count
++;
769 exec
->vtx
.prim
[i
].mode
= mode
;
770 exec
->vtx
.prim
[i
].begin
= 1;
771 exec
->vtx
.prim
[i
].end
= 0;
772 exec
->vtx
.prim
[i
].indexed
= 0;
773 exec
->vtx
.prim
[i
].pad
= 0;
774 exec
->vtx
.prim
[i
].start
= exec
->vtx
.vert_count
;
775 exec
->vtx
.prim
[i
].count
= 0;
776 exec
->vtx
.prim
[i
].num_instances
= 1;
777 exec
->vtx
.prim
[i
].base_instance
= 0;
778 exec
->vtx
.prim
[i
].is_indirect
= 0;
780 ctx
->Driver
.CurrentExecPrimitive
= mode
;
782 ctx
->Exec
= ctx
->BeginEnd
;
784 /* We may have been called from a display list, in which case we should
785 * leave dlist.c's dispatch table in place.
787 if (ctx
->CurrentClientDispatch
== ctx
->MarshalExec
) {
788 ctx
->CurrentServerDispatch
= ctx
->Exec
;
789 } else if (ctx
->CurrentClientDispatch
== ctx
->OutsideBeginEnd
) {
790 ctx
->CurrentClientDispatch
= ctx
->Exec
;
791 _glapi_set_dispatch(ctx
->CurrentClientDispatch
);
793 assert(ctx
->CurrentClientDispatch
== ctx
->Save
);
799 * Try to merge / concatenate the two most recent VBO primitives.
802 try_vbo_merge(struct vbo_exec_context
*exec
)
804 struct _mesa_prim
*cur
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
806 assert(exec
->vtx
.prim_count
>= 1);
808 vbo_try_prim_conversion(cur
);
810 if (exec
->vtx
.prim_count
>= 2) {
811 struct _mesa_prim
*prev
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 2];
812 assert(prev
== cur
- 1);
814 if (vbo_can_merge_prims(prev
, cur
)) {
819 vbo_merge_prims(prev
, cur
);
820 exec
->vtx
.prim_count
--; /* drop the last primitive */
829 static void GLAPIENTRY
832 GET_CURRENT_CONTEXT(ctx
);
833 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
835 if (!_mesa_inside_begin_end(ctx
)) {
836 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glEnd");
840 ctx
->Exec
= ctx
->OutsideBeginEnd
;
842 if (ctx
->CurrentClientDispatch
== ctx
->MarshalExec
) {
843 ctx
->CurrentServerDispatch
= ctx
->Exec
;
844 } else if (ctx
->CurrentClientDispatch
== ctx
->BeginEnd
) {
845 ctx
->CurrentClientDispatch
= ctx
->Exec
;
846 _glapi_set_dispatch(ctx
->CurrentClientDispatch
);
849 if (exec
->vtx
.prim_count
> 0) {
850 /* close off current primitive */
851 struct _mesa_prim
*last_prim
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
854 last_prim
->count
= exec
->vtx
.vert_count
- last_prim
->start
;
856 /* Special handling for GL_LINE_LOOP */
857 if (last_prim
->mode
== GL_LINE_LOOP
&& last_prim
->begin
== 0) {
858 /* We're finishing drawing a line loop. Append 0th vertex onto
859 * end of vertex buffer so we can draw it as a line strip.
861 const fi_type
*src
= exec
->vtx
.buffer_map
+
862 last_prim
->start
* exec
->vtx
.vertex_size
;
863 fi_type
*dst
= exec
->vtx
.buffer_map
+
864 exec
->vtx
.vert_count
* exec
->vtx
.vertex_size
;
866 /* copy 0th vertex to end of buffer */
867 memcpy(dst
, src
, exec
->vtx
.vertex_size
* sizeof(fi_type
));
869 last_prim
->start
++; /* skip vertex0 */
870 /* note that last_prim->count stays unchanged */
871 last_prim
->mode
= GL_LINE_STRIP
;
873 /* Increment the vertex count so the next primitive doesn't
874 * overwrite the last vertex which we just added.
876 exec
->vtx
.vert_count
++;
877 exec
->vtx
.buffer_ptr
+= exec
->vtx
.vertex_size
;
883 ctx
->Driver
.CurrentExecPrimitive
= PRIM_OUTSIDE_BEGIN_END
;
885 if (exec
->vtx
.prim_count
== VBO_MAX_PRIM
)
886 vbo_exec_vtx_flush(exec
);
888 if (MESA_DEBUG_FLAGS
& DEBUG_ALWAYS_FLUSH
) {
895 * Called via glPrimitiveRestartNV()
897 static void GLAPIENTRY
898 vbo_exec_PrimitiveRestartNV(void)
901 GET_CURRENT_CONTEXT(ctx
);
903 curPrim
= ctx
->Driver
.CurrentExecPrimitive
;
905 if (curPrim
== PRIM_OUTSIDE_BEGIN_END
) {
906 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glPrimitiveRestartNV");
910 vbo_exec_Begin(curPrim
);
916 vbo_exec_vtxfmt_init(struct vbo_exec_context
*exec
)
918 struct gl_context
*ctx
= exec
->ctx
;
919 GLvertexformat
*vfmt
= &exec
->vtxfmt
;
921 #define NAME_AE(x) _ae_##x
922 #define NAME_CALLLIST(x) _mesa_##x
923 #define NAME(x) vbo_exec_##x
924 #define NAME_ES(x) _es_##x
926 #include "vbo_init_tmp.h"
931 * Tell the VBO module to use a real OpenGL vertex buffer object to
932 * store accumulated immediate-mode vertex data.
933 * This replaces the malloced buffer which was created in
934 * vb_exec_vtx_init() below.
937 vbo_use_buffer_objects(struct gl_context
*ctx
)
939 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
940 /* Any buffer name but 0 can be used here since this bufferobj won't
941 * go into the bufferobj hashtable.
943 GLuint bufName
= IMM_BUFFER_NAME
;
945 /* Make sure this func is only used once */
946 assert(exec
->vtx
.bufferobj
== ctx
->Shared
->NullBufferObj
);
948 _mesa_align_free(exec
->vtx
.buffer_map
);
949 exec
->vtx
.buffer_map
= NULL
;
950 exec
->vtx
.buffer_ptr
= NULL
;
952 /* Allocate a real buffer object now */
953 _mesa_reference_buffer_object(ctx
, &exec
->vtx
.bufferobj
, NULL
);
954 exec
->vtx
.bufferobj
= ctx
->Driver
.NewBufferObject(ctx
, bufName
);
956 /* Map the buffer. */
957 vbo_exec_vtx_map(exec
);
958 assert(exec
->vtx
.buffer_ptr
);
963 vbo_exec_vtx_init(struct vbo_exec_context
*exec
)
965 struct gl_context
*ctx
= exec
->ctx
;
968 /* Allocate a buffer object. Will just reuse this object
969 * continuously, unless vbo_use_buffer_objects() is called to enable
972 _mesa_reference_buffer_object(ctx
,
973 &exec
->vtx
.bufferobj
,
974 ctx
->Shared
->NullBufferObj
);
976 assert(!exec
->vtx
.buffer_map
);
977 exec
->vtx
.buffer_map
= _mesa_align_malloc(VBO_VERT_BUFFER_SIZE
, 64);
978 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
980 vbo_exec_vtxfmt_init(exec
);
981 _mesa_noop_vtxfmt_init(ctx
, &exec
->vtxfmt_noop
);
983 exec
->vtx
.enabled
= 0;
984 for (i
= 0 ; i
< ARRAY_SIZE(exec
->vtx
.attr
); i
++) {
985 exec
->vtx
.attr
[i
].size
= 0;
986 exec
->vtx
.attr
[i
].type
= GL_FLOAT
;
987 exec
->vtx
.attr
[i
].active_size
= 0;
990 exec
->vtx
.vertex_size
= 0;
995 vbo_exec_vtx_destroy(struct vbo_exec_context
*exec
)
997 /* using a real VBO for vertex data */
998 struct gl_context
*ctx
= exec
->ctx
;
1000 /* True VBOs should already be unmapped
1002 if (exec
->vtx
.buffer_map
) {
1003 assert(exec
->vtx
.bufferobj
->Name
== 0 ||
1004 exec
->vtx
.bufferobj
->Name
== IMM_BUFFER_NAME
);
1005 if (exec
->vtx
.bufferobj
->Name
== 0) {
1006 _mesa_align_free(exec
->vtx
.buffer_map
);
1007 exec
->vtx
.buffer_map
= NULL
;
1008 exec
->vtx
.buffer_ptr
= NULL
;
1012 /* Free the vertex buffer. Unmap first if needed.
1014 if (_mesa_bufferobj_mapped(exec
->vtx
.bufferobj
, MAP_INTERNAL
)) {
1015 ctx
->Driver
.UnmapBuffer(ctx
, exec
->vtx
.bufferobj
, MAP_INTERNAL
);
1017 _mesa_reference_buffer_object(ctx
, &exec
->vtx
.bufferobj
, NULL
);
1022 * If inside glBegin()/glEnd(), it should assert(0). Otherwise, if
1023 * FLUSH_STORED_VERTICES bit in \p flags is set flushes any buffered
1024 * vertices, if FLUSH_UPDATE_CURRENT bit is set updates
1025 * __struct gl_contextRec::Current and gl_light_attrib::Material
1027 * Note that the default T&L engine never clears the
1028 * FLUSH_UPDATE_CURRENT bit, even after performing the update.
1030 * \param flags bitmask of FLUSH_STORED_VERTICES, FLUSH_UPDATE_CURRENT
1033 vbo_exec_FlushVertices(struct gl_context
*ctx
, GLuint flags
)
1035 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
1038 /* debug check: make sure we don't get called recursively */
1039 exec
->flush_call_depth
++;
1040 assert(exec
->flush_call_depth
== 1);
1043 if (_mesa_inside_begin_end(ctx
)) {
1044 /* We've had glBegin but not glEnd! */
1046 exec
->flush_call_depth
--;
1047 assert(exec
->flush_call_depth
== 0);
1053 vbo_exec_FlushVertices_internal(exec
);
1055 /* Clear the dirty flush flags, because the flush is finished. */
1056 ctx
->Driver
.NeedFlush
&= ~(FLUSH_UPDATE_CURRENT
| flags
);
1059 exec
->flush_call_depth
--;
1060 assert(exec
->flush_call_depth
== 0);
1066 * Reset the vertex attribute by setting its size to zero.
1069 vbo_reset_attr(struct vbo_exec_context
*exec
, GLuint attr
)
1071 exec
->vtx
.attr
[attr
].size
= 0;
1072 exec
->vtx
.attr
[attr
].type
= GL_FLOAT
;
1073 exec
->vtx
.attr
[attr
].active_size
= 0;
1078 vbo_reset_all_attr(struct vbo_exec_context
*exec
)
1080 while (exec
->vtx
.enabled
) {
1081 const int i
= u_bit_scan64(&exec
->vtx
.enabled
);
1082 vbo_reset_attr(exec
, i
);
1085 exec
->vtx
.vertex_size
= 0;
1090 _es_Color4f(GLfloat r
, GLfloat g
, GLfloat b
, GLfloat a
)
1092 vbo_exec_Color4f(r
, g
, b
, a
);
1097 _es_Normal3f(GLfloat x
, GLfloat y
, GLfloat z
)
1099 vbo_exec_Normal3f(x
, y
, z
);
1104 _es_MultiTexCoord4f(GLenum target
, GLfloat s
, GLfloat t
, GLfloat r
, GLfloat q
)
1106 vbo_exec_MultiTexCoord4f(target
, s
, t
, r
, q
);
1111 _es_Materialfv(GLenum face
, GLenum pname
, const GLfloat
*params
)
1113 vbo_exec_Materialfv(face
, pname
, params
);
1118 _es_Materialf(GLenum face
, GLenum pname
, GLfloat param
)
1122 p
[1] = p
[2] = p
[3] = 0.0F
;
1123 vbo_exec_Materialfv(face
, pname
, p
);
1128 * A special version of glVertexAttrib4f that does not treat index 0 as
1132 VertexAttrib4f_nopos(GLuint index
, GLfloat x
, GLfloat y
, GLfloat z
, GLfloat w
)
1134 GET_CURRENT_CONTEXT(ctx
);
1135 if (index
< MAX_VERTEX_GENERIC_ATTRIBS
)
1136 ATTRF(VBO_ATTRIB_GENERIC0
+ index
, 4, x
, y
, z
, w
);
1138 ERROR(GL_INVALID_VALUE
);
1142 _es_VertexAttrib4f(GLuint index
, GLfloat x
, GLfloat y
, GLfloat z
, GLfloat w
)
1144 VertexAttrib4f_nopos(index
, x
, y
, z
, w
);
1149 _es_VertexAttrib1f(GLuint indx
, GLfloat x
)
1151 VertexAttrib4f_nopos(indx
, x
, 0.0f
, 0.0f
, 1.0f
);
1156 _es_VertexAttrib1fv(GLuint indx
, const GLfloat
* values
)
1158 VertexAttrib4f_nopos(indx
, values
[0], 0.0f
, 0.0f
, 1.0f
);
1163 _es_VertexAttrib2f(GLuint indx
, GLfloat x
, GLfloat y
)
1165 VertexAttrib4f_nopos(indx
, x
, y
, 0.0f
, 1.0f
);
1170 _es_VertexAttrib2fv(GLuint indx
, const GLfloat
* values
)
1172 VertexAttrib4f_nopos(indx
, values
[0], values
[1], 0.0f
, 1.0f
);
1177 _es_VertexAttrib3f(GLuint indx
, GLfloat x
, GLfloat y
, GLfloat z
)
1179 VertexAttrib4f_nopos(indx
, x
, y
, z
, 1.0f
);
1184 _es_VertexAttrib3fv(GLuint indx
, const GLfloat
* values
)
1186 VertexAttrib4f_nopos(indx
, values
[0], values
[1], values
[2], 1.0f
);
1191 _es_VertexAttrib4fv(GLuint indx
, const GLfloat
* values
)
1193 VertexAttrib4f_nopos(indx
, values
[0], values
[1], values
[2], values
[3]);