1 /**************************************************************************
3 Copyright 2002-2008 VMware, Inc.
7 Permission is hereby granted, free of charge, to any person obtaining a
8 copy of this software and associated documentation files (the "Software"),
9 to deal in the Software without restriction, including without limitation
10 on the rights to use, copy, modify, merge, publish, distribute, sub
11 license, and/or sell copies of the Software, and to permit persons to whom
12 the Software is furnished to do so, subject to the following conditions:
14 The above copyright notice and this permission notice (including the next
15 paragraph) shall be included in all copies or substantial portions of the
18 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21 VMWARE AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
22 DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23 OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24 USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Keith Whitwell <keithw@vmware.com>
33 #include "main/glheader.h"
34 #include "main/bufferobj.h"
35 #include "main/context.h"
36 #include "main/macros.h"
37 #include "main/vtxfmt.h"
38 #include "main/dlist.h"
39 #include "main/eval.h"
40 #include "main/state.h"
41 #include "main/light.h"
42 #include "main/api_arrayelt.h"
43 #include "main/draw_validate.h"
44 #include "main/dispatch.h"
45 #include "util/bitscan.h"
48 #include "vbo_private.h"
51 /** ID/name for immediate-mode VBO */
52 #define IMM_BUFFER_NAME 0xaabbccdd
56 vbo_reset_all_attr(struct vbo_exec_context
*exec
);
60 * Close off the last primitive, execute the buffer, restart the
61 * primitive. This is called when we fill a vertex buffer before
65 vbo_exec_wrap_buffers(struct vbo_exec_context
*exec
)
67 if (exec
->vtx
.prim_count
== 0) {
68 exec
->vtx
.copied
.nr
= 0;
69 exec
->vtx
.vert_count
= 0;
70 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
73 struct _mesa_prim
*last_prim
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
74 const GLuint last_begin
= last_prim
->begin
;
77 if (_mesa_inside_begin_end(exec
->ctx
)) {
78 last_prim
->count
= exec
->vtx
.vert_count
- last_prim
->start
;
81 last_count
= last_prim
->count
;
83 /* Special handling for wrapping GL_LINE_LOOP */
84 if (last_prim
->mode
== GL_LINE_LOOP
&&
87 /* draw this section of the incomplete line loop as a line strip */
88 last_prim
->mode
= GL_LINE_STRIP
;
89 if (!last_prim
->begin
) {
90 /* This is not the first section of the line loop, so don't
91 * draw the 0th vertex. We're saving it until we draw the
92 * very last section of the loop.
99 /* Execute the buffer and save copied vertices.
101 if (exec
->vtx
.vert_count
)
102 vbo_exec_vtx_flush(exec
, GL_FALSE
);
104 exec
->vtx
.prim_count
= 0;
105 exec
->vtx
.copied
.nr
= 0;
108 /* Emit a glBegin to start the new list.
110 assert(exec
->vtx
.prim_count
== 0);
112 if (_mesa_inside_begin_end(exec
->ctx
)) {
113 exec
->vtx
.prim
[0].mode
= exec
->ctx
->Driver
.CurrentExecPrimitive
;
114 exec
->vtx
.prim
[0].begin
= 0;
115 exec
->vtx
.prim
[0].end
= 0;
116 exec
->vtx
.prim
[0].start
= 0;
117 exec
->vtx
.prim
[0].count
= 0;
118 exec
->vtx
.prim_count
++;
120 if (exec
->vtx
.copied
.nr
== last_count
)
121 exec
->vtx
.prim
[0].begin
= last_begin
;
128 * Deal with buffer wrapping where provoked by the vertex buffer
129 * filling up, as opposed to upgrade_vertex().
132 vbo_exec_vtx_wrap(struct vbo_exec_context
*exec
)
134 unsigned numComponents
;
136 /* Run pipeline on current vertices, copy wrapped vertices
137 * to exec->vtx.copied.
139 vbo_exec_wrap_buffers(exec
);
141 if (!exec
->vtx
.buffer_ptr
) {
142 /* probably ran out of memory earlier when allocating the VBO */
146 /* Copy stored stored vertices to start of new list.
148 assert(exec
->vtx
.max_vert
- exec
->vtx
.vert_count
> exec
->vtx
.copied
.nr
);
150 numComponents
= exec
->vtx
.copied
.nr
* exec
->vtx
.vertex_size
;
151 memcpy(exec
->vtx
.buffer_ptr
,
152 exec
->vtx
.copied
.buffer
,
153 numComponents
* sizeof(fi_type
));
154 exec
->vtx
.buffer_ptr
+= numComponents
;
155 exec
->vtx
.vert_count
+= exec
->vtx
.copied
.nr
;
157 exec
->vtx
.copied
.nr
= 0;
162 * Copy the active vertex's values to the ctx->Current fields.
165 vbo_exec_copy_to_current(struct vbo_exec_context
*exec
)
167 struct gl_context
*ctx
= exec
->ctx
;
168 struct vbo_context
*vbo
= vbo_context(ctx
);
169 GLbitfield64 enabled
= exec
->vtx
.enabled
& (~BITFIELD64_BIT(VBO_ATTRIB_POS
));
172 const int i
= u_bit_scan64(&enabled
);
174 /* Note: the exec->vtx.current[i] pointers point into the
175 * ctx->Current.Attrib and ctx->Light.Material.Attrib arrays.
177 GLfloat
*current
= (GLfloat
*)vbo
->current
[i
].Ptr
;
178 fi_type tmp
[8]; /* space for doubles */
181 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
182 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
)
185 assert(exec
->vtx
.attr
[i
].size
);
187 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
188 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
) {
189 memset(tmp
, 0, sizeof(tmp
));
190 memcpy(tmp
, exec
->vtx
.attrptr
[i
], exec
->vtx
.attr
[i
].size
* sizeof(GLfloat
));
192 COPY_CLEAN_4V_TYPE_AS_UNION(tmp
,
193 exec
->vtx
.attr
[i
].size
,
194 exec
->vtx
.attrptr
[i
],
195 exec
->vtx
.attr
[i
].type
);
198 if (exec
->vtx
.attr
[i
].type
!= vbo
->current
[i
].Format
.Type
||
199 memcmp(current
, tmp
, 4 * sizeof(GLfloat
) * dmul
) != 0) {
200 memcpy(current
, tmp
, 4 * sizeof(GLfloat
) * dmul
);
202 /* Given that we explicitly state size here, there is no need
203 * for the COPY_CLEAN above, could just copy 16 bytes and be
204 * done. The only problem is when Mesa accesses ctx->Current
207 /* Size here is in components - not bytes */
208 vbo_set_vertex_format(&vbo
->current
[i
].Format
,
209 exec
->vtx
.attr
[i
].size
/ dmul
,
210 exec
->vtx
.attr
[i
].type
);
212 /* This triggers rather too much recalculation of Mesa state
213 * that doesn't get used (eg light positions).
215 if (i
>= VBO_ATTRIB_MAT_FRONT_AMBIENT
&&
216 i
<= VBO_ATTRIB_MAT_BACK_INDEXES
)
217 ctx
->NewState
|= _NEW_LIGHT
;
219 ctx
->NewState
|= _NEW_CURRENT_ATTRIB
;
223 /* Colormaterial -- this kindof sucks.
225 if (ctx
->Light
.ColorMaterialEnabled
&&
226 exec
->vtx
.attr
[VBO_ATTRIB_COLOR0
].size
) {
227 _mesa_update_color_material(ctx
,
228 ctx
->Current
.Attrib
[VBO_ATTRIB_COLOR0
]);
234 * Copy current vertex attribute values into the current vertex.
237 vbo_exec_copy_from_current(struct vbo_exec_context
*exec
)
239 struct gl_context
*ctx
= exec
->ctx
;
240 struct vbo_context
*vbo
= vbo_context(ctx
);
243 for (i
= VBO_ATTRIB_POS
+ 1; i
< VBO_ATTRIB_MAX
; i
++) {
244 if (exec
->vtx
.attr
[i
].type
== GL_DOUBLE
||
245 exec
->vtx
.attr
[i
].type
== GL_UNSIGNED_INT64_ARB
) {
246 memcpy(exec
->vtx
.attrptr
[i
], vbo
->current
[i
].Ptr
,
247 exec
->vtx
.attr
[i
].size
* sizeof(GLfloat
));
249 const fi_type
*current
= (fi_type
*) vbo
->current
[i
].Ptr
;
250 switch (exec
->vtx
.attr
[i
].size
) {
251 case 4: exec
->vtx
.attrptr
[i
][3] = current
[3];
252 case 3: exec
->vtx
.attrptr
[i
][2] = current
[2];
253 case 2: exec
->vtx
.attrptr
[i
][1] = current
[1];
254 case 1: exec
->vtx
.attrptr
[i
][0] = current
[0];
263 * Flush existing data, set new attrib size, replay copied vertices.
264 * This is called when we transition from a small vertex attribute size
265 * to a larger one. Ex: glTexCoord2f -> glTexCoord4f.
266 * We need to go back over the previous 2-component texcoords and insert
267 * zero and one values.
268 * \param attr VBO_ATTRIB_x vertex attribute value
271 vbo_exec_wrap_upgrade_vertex(struct vbo_exec_context
*exec
,
272 GLuint attr
, GLuint newSize
)
274 struct gl_context
*ctx
= exec
->ctx
;
275 struct vbo_context
*vbo
= vbo_context(ctx
);
276 const GLint lastcount
= exec
->vtx
.vert_count
;
277 fi_type
*old_attrptr
[VBO_ATTRIB_MAX
];
278 const GLuint old_vtx_size
= exec
->vtx
.vertex_size
; /* floats per vertex */
279 const GLuint oldSize
= exec
->vtx
.attr
[attr
].size
;
282 assert(attr
< VBO_ATTRIB_MAX
);
284 /* Run pipeline on current vertices, copy wrapped vertices
285 * to exec->vtx.copied.
287 vbo_exec_wrap_buffers(exec
);
289 if (unlikely(exec
->vtx
.copied
.nr
)) {
290 /* We're in the middle of a primitive, keep the old vertex
291 * format around to be able to translate the copied vertices to
294 memcpy(old_attrptr
, exec
->vtx
.attrptr
, sizeof(old_attrptr
));
297 if (unlikely(oldSize
)) {
298 /* Do a COPY_TO_CURRENT to ensure back-copying works for the
299 * case when the attribute already exists in the vertex and is
300 * having its size increased.
302 vbo_exec_copy_to_current(exec
);
305 /* Heuristic: Attempt to isolate attributes received outside
306 * begin/end so that they don't bloat the vertices.
308 if (!_mesa_inside_begin_end(ctx
) &&
309 !oldSize
&& lastcount
> 8 && exec
->vtx
.vertex_size
) {
310 vbo_exec_copy_to_current(exec
);
311 vbo_reset_all_attr(exec
);
316 exec
->vtx
.attr
[attr
].size
= newSize
;
317 exec
->vtx
.vertex_size
+= newSize
- oldSize
;
318 exec
->vtx
.max_vert
= vbo_compute_max_verts(exec
);
319 exec
->vtx
.vert_count
= 0;
320 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
321 exec
->vtx
.enabled
|= BITFIELD64_BIT(attr
);
323 if (unlikely(oldSize
)) {
324 /* Size changed, recalculate all the attrptr[] values
326 fi_type
*tmp
= exec
->vtx
.vertex
;
328 for (i
= 0 ; i
< VBO_ATTRIB_MAX
; i
++) {
329 if (exec
->vtx
.attr
[i
].size
) {
330 exec
->vtx
.attrptr
[i
] = tmp
;
331 tmp
+= exec
->vtx
.attr
[i
].size
;
334 exec
->vtx
.attrptr
[i
] = NULL
; /* will not be dereferenced */
337 /* Copy from current to repopulate the vertex with correct
340 vbo_exec_copy_from_current(exec
);
343 /* Just have to append the new attribute at the end */
344 exec
->vtx
.attrptr
[attr
] = exec
->vtx
.vertex
+
345 exec
->vtx
.vertex_size
- newSize
;
348 /* Replay stored vertices to translate them
349 * to new format here.
351 * -- No need to replay - just copy piecewise
353 if (unlikely(exec
->vtx
.copied
.nr
)) {
354 fi_type
*data
= exec
->vtx
.copied
.buffer
;
355 fi_type
*dest
= exec
->vtx
.buffer_ptr
;
357 assert(exec
->vtx
.buffer_ptr
== exec
->vtx
.buffer_map
);
359 for (i
= 0 ; i
< exec
->vtx
.copied
.nr
; i
++) {
360 GLbitfield64 enabled
= exec
->vtx
.enabled
;
362 const int j
= u_bit_scan64(&enabled
);
363 GLuint sz
= exec
->vtx
.attr
[j
].size
;
364 GLint old_offset
= old_attrptr
[j
] - exec
->vtx
.vertex
;
365 GLint new_offset
= exec
->vtx
.attrptr
[j
] - exec
->vtx
.vertex
;
372 COPY_CLEAN_4V_TYPE_AS_UNION(tmp
, oldSize
,
374 exec
->vtx
.attr
[j
].type
);
375 COPY_SZ_4V(dest
+ new_offset
, newSize
, tmp
);
377 fi_type
*current
= (fi_type
*)vbo
->current
[j
].Ptr
;
378 COPY_SZ_4V(dest
+ new_offset
, sz
, current
);
382 COPY_SZ_4V(dest
+ new_offset
, sz
, data
+ old_offset
);
386 data
+= old_vtx_size
;
387 dest
+= exec
->vtx
.vertex_size
;
390 exec
->vtx
.buffer_ptr
= dest
;
391 exec
->vtx
.vert_count
+= exec
->vtx
.copied
.nr
;
392 exec
->vtx
.copied
.nr
= 0;
398 * This is when a vertex attribute transitions to a different size.
399 * For example, we saw a bunch of glTexCoord2f() calls and now we got a
400 * glTexCoord4f() call. We promote the array from size=2 to size=4.
401 * \param newSize size of new vertex (number of 32-bit words).
402 * \param attr VBO_ATTRIB_x vertex attribute value
405 vbo_exec_fixup_vertex(struct gl_context
*ctx
, GLuint attr
,
406 GLuint newSize
, GLenum newType
)
408 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
410 assert(attr
< VBO_ATTRIB_MAX
);
412 if (newSize
> exec
->vtx
.attr
[attr
].size
||
413 newType
!= exec
->vtx
.attr
[attr
].type
) {
414 /* New size is larger. Need to flush existing vertices and get
415 * an enlarged vertex format.
417 vbo_exec_wrap_upgrade_vertex(exec
, attr
, newSize
);
419 else if (newSize
< exec
->vtx
.attr
[attr
].active_size
) {
422 vbo_get_default_vals_as_union(exec
->vtx
.attr
[attr
].type
);
424 /* New size is smaller - just need to fill in some
425 * zeros. Don't need to flush or wrap.
427 for (i
= newSize
; i
<= exec
->vtx
.attr
[attr
].size
; i
++)
428 exec
->vtx
.attrptr
[attr
][i
-1] = id
[i
-1];
431 exec
->vtx
.attr
[attr
].active_size
= newSize
;
432 exec
->vtx
.attr
[attr
].type
= newType
;
434 /* Does setting NeedFlush belong here? Necessitates resetting
435 * vtxfmt on each flush (otherwise flags won't get reset
439 ctx
->Driver
.NeedFlush
|= FLUSH_STORED_VERTICES
;
444 * If index=0, does glVertexAttrib*() alias glVertex() to emit a vertex?
445 * It depends on a few things, including whether we're inside or outside
449 is_vertex_position(const struct gl_context
*ctx
, GLuint index
)
451 return (index
== 0 &&
452 _mesa_attr_zero_aliases_vertex(ctx
) &&
453 _mesa_inside_begin_end(ctx
));
458 * This macro is used to implement all the glVertex, glColor, glTexCoord,
459 * glVertexAttrib, etc functions.
460 * \param A VBO_ATTRIB_x attribute index
461 * \param N attribute size (1..4)
462 * \param T type (GL_FLOAT, GL_DOUBLE, GL_INT, GL_UNSIGNED_INT)
463 * \param C cast type (fi_type or double)
464 * \param V0, V1, v2, V3 attribute value
466 #define ATTR_UNION(A, N, T, C, V0, V1, V2, V3) \
468 struct vbo_exec_context *exec = &vbo_context(ctx)->exec; \
469 int sz = (sizeof(C) / sizeof(GLfloat)); \
471 assert(sz == 1 || sz == 2); \
473 /* check if attribute size or type is changing */ \
474 if (unlikely(exec->vtx.attr[A].active_size != N * sz) || \
475 unlikely(exec->vtx.attr[A].type != T)) { \
476 vbo_exec_fixup_vertex(ctx, A, N * sz, T); \
479 /* store vertex attribute in vertex buffer */ \
481 C *dest = (C *)exec->vtx.attrptr[A]; \
482 if (N>0) dest[0] = V0; \
483 if (N>1) dest[1] = V1; \
484 if (N>2) dest[2] = V2; \
485 if (N>3) dest[3] = V3; \
486 assert(exec->vtx.attr[A].type == T); \
490 /* This is a glVertex call */ \
493 if (unlikely(!exec->vtx.buffer_ptr)) { \
494 vbo_exec_vtx_map(exec); \
496 assert(exec->vtx.buffer_ptr); \
498 /* copy 32-bit words */ \
499 for (i = 0; i < exec->vtx.vertex_size; i++) \
500 exec->vtx.buffer_ptr[i] = exec->vtx.vertex[i]; \
502 exec->vtx.buffer_ptr += exec->vtx.vertex_size; \
504 /* Set FLUSH_STORED_VERTICES to indicate that there's now */ \
505 /* something to draw (not just updating a color or texcoord).*/ \
506 ctx->Driver.NeedFlush |= FLUSH_UPDATE_CURRENT | \
507 FLUSH_STORED_VERTICES; \
509 if (++exec->vtx.vert_count >= exec->vtx.max_vert) \
510 vbo_exec_vtx_wrap(exec); \
512 /* we now have accumulated per-vertex attributes */ \
513 ctx->Driver.NeedFlush |= FLUSH_UPDATE_CURRENT; \
519 #define ERROR(err) _mesa_error(ctx, err, __func__)
520 #define TAG(x) vbo_exec_##x
522 #include "vbo_attrib_tmp.h"
527 * Execute a glMaterial call. Note that if GL_COLOR_MATERIAL is enabled,
528 * this may be a (partial) no-op.
530 static void GLAPIENTRY
531 vbo_exec_Materialfv(GLenum face
, GLenum pname
, const GLfloat
*params
)
533 GLbitfield updateMats
;
534 GET_CURRENT_CONTEXT(ctx
);
536 /* This function should be a no-op when it tries to update material
537 * attributes which are currently tracking glColor via glColorMaterial.
538 * The updateMats var will be a mask of the MAT_BIT_FRONT/BACK_x bits
539 * indicating which material attributes can actually be updated below.
541 if (ctx
->Light
.ColorMaterialEnabled
) {
542 updateMats
= ~ctx
->Light
._ColorMaterialBitmask
;
545 /* GL_COLOR_MATERIAL is disabled so don't skip any material updates */
546 updateMats
= ALL_MATERIAL_BITS
;
549 if (ctx
->API
== API_OPENGL_COMPAT
&& face
== GL_FRONT
) {
550 updateMats
&= FRONT_MATERIAL_BITS
;
552 else if (ctx
->API
== API_OPENGL_COMPAT
&& face
== GL_BACK
) {
553 updateMats
&= BACK_MATERIAL_BITS
;
555 else if (face
!= GL_FRONT_AND_BACK
) {
556 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterial(invalid face)");
562 if (updateMats
& MAT_BIT_FRONT_EMISSION
)
563 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_EMISSION
, 4, params
);
564 if (updateMats
& MAT_BIT_BACK_EMISSION
)
565 MAT_ATTR(VBO_ATTRIB_MAT_BACK_EMISSION
, 4, params
);
568 if (updateMats
& MAT_BIT_FRONT_AMBIENT
)
569 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT
, 4, params
);
570 if (updateMats
& MAT_BIT_BACK_AMBIENT
)
571 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT
, 4, params
);
574 if (updateMats
& MAT_BIT_FRONT_DIFFUSE
)
575 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE
, 4, params
);
576 if (updateMats
& MAT_BIT_BACK_DIFFUSE
)
577 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE
, 4, params
);
580 if (updateMats
& MAT_BIT_FRONT_SPECULAR
)
581 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SPECULAR
, 4, params
);
582 if (updateMats
& MAT_BIT_BACK_SPECULAR
)
583 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SPECULAR
, 4, params
);
586 if (*params
< 0 || *params
> ctx
->Const
.MaxShininess
) {
587 _mesa_error(ctx
, GL_INVALID_VALUE
,
588 "glMaterial(invalid shininess: %f out range [0, %f])",
589 *params
, ctx
->Const
.MaxShininess
);
592 if (updateMats
& MAT_BIT_FRONT_SHININESS
)
593 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SHININESS
, 1, params
);
594 if (updateMats
& MAT_BIT_BACK_SHININESS
)
595 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SHININESS
, 1, params
);
597 case GL_COLOR_INDEXES
:
598 if (ctx
->API
!= API_OPENGL_COMPAT
) {
599 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterialfv(pname)");
602 if (updateMats
& MAT_BIT_FRONT_INDEXES
)
603 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_INDEXES
, 3, params
);
604 if (updateMats
& MAT_BIT_BACK_INDEXES
)
605 MAT_ATTR(VBO_ATTRIB_MAT_BACK_INDEXES
, 3, params
);
607 case GL_AMBIENT_AND_DIFFUSE
:
608 if (updateMats
& MAT_BIT_FRONT_AMBIENT
)
609 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT
, 4, params
);
610 if (updateMats
& MAT_BIT_FRONT_DIFFUSE
)
611 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE
, 4, params
);
612 if (updateMats
& MAT_BIT_BACK_AMBIENT
)
613 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT
, 4, params
);
614 if (updateMats
& MAT_BIT_BACK_DIFFUSE
)
615 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE
, 4, params
);
618 _mesa_error(ctx
, GL_INVALID_ENUM
, "glMaterialfv(pname)");
625 * Flush (draw) vertices.
626 * \param unmap - leave VBO unmapped after flushing?
629 vbo_exec_FlushVertices_internal(struct vbo_exec_context
*exec
, GLboolean unmap
)
631 if (exec
->vtx
.vert_count
|| unmap
) {
632 vbo_exec_vtx_flush(exec
, unmap
);
635 if (exec
->vtx
.vertex_size
) {
636 vbo_exec_copy_to_current(exec
);
637 vbo_reset_all_attr(exec
);
642 static void GLAPIENTRY
643 vbo_exec_EvalCoord1f(GLfloat u
)
645 GET_CURRENT_CONTEXT(ctx
);
646 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
650 if (exec
->eval
.recalculate_maps
)
651 vbo_exec_eval_update(exec
);
653 for (i
= 0; i
<= VBO_ATTRIB_TEX7
; i
++) {
654 if (exec
->eval
.map1
[i
].map
)
655 if (exec
->vtx
.attr
[i
].active_size
!= exec
->eval
.map1
[i
].sz
)
656 vbo_exec_fixup_vertex(ctx
, i
, exec
->eval
.map1
[i
].sz
, GL_FLOAT
);
660 memcpy(exec
->vtx
.copied
.buffer
, exec
->vtx
.vertex
,
661 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
663 vbo_exec_do_EvalCoord1f(exec
, u
);
665 memcpy(exec
->vtx
.vertex
, exec
->vtx
.copied
.buffer
,
666 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
670 static void GLAPIENTRY
671 vbo_exec_EvalCoord2f(GLfloat u
, GLfloat v
)
673 GET_CURRENT_CONTEXT(ctx
);
674 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
678 if (exec
->eval
.recalculate_maps
)
679 vbo_exec_eval_update(exec
);
681 for (i
= 0; i
<= VBO_ATTRIB_TEX7
; i
++) {
682 if (exec
->eval
.map2
[i
].map
)
683 if (exec
->vtx
.attr
[i
].active_size
!= exec
->eval
.map2
[i
].sz
)
684 vbo_exec_fixup_vertex(ctx
, i
, exec
->eval
.map2
[i
].sz
, GL_FLOAT
);
687 if (ctx
->Eval
.AutoNormal
)
688 if (exec
->vtx
.attr
[VBO_ATTRIB_NORMAL
].active_size
!= 3)
689 vbo_exec_fixup_vertex(ctx
, VBO_ATTRIB_NORMAL
, 3, GL_FLOAT
);
692 memcpy(exec
->vtx
.copied
.buffer
, exec
->vtx
.vertex
,
693 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
695 vbo_exec_do_EvalCoord2f(exec
, u
, v
);
697 memcpy(exec
->vtx
.vertex
, exec
->vtx
.copied
.buffer
,
698 exec
->vtx
.vertex_size
* sizeof(GLfloat
));
702 static void GLAPIENTRY
703 vbo_exec_EvalCoord1fv(const GLfloat
*u
)
705 vbo_exec_EvalCoord1f(u
[0]);
709 static void GLAPIENTRY
710 vbo_exec_EvalCoord2fv(const GLfloat
*u
)
712 vbo_exec_EvalCoord2f(u
[0], u
[1]);
716 static void GLAPIENTRY
717 vbo_exec_EvalPoint1(GLint i
)
719 GET_CURRENT_CONTEXT(ctx
);
720 GLfloat du
= ((ctx
->Eval
.MapGrid1u2
- ctx
->Eval
.MapGrid1u1
) /
721 (GLfloat
) ctx
->Eval
.MapGrid1un
);
722 GLfloat u
= i
* du
+ ctx
->Eval
.MapGrid1u1
;
724 vbo_exec_EvalCoord1f(u
);
728 static void GLAPIENTRY
729 vbo_exec_EvalPoint2(GLint i
, GLint j
)
731 GET_CURRENT_CONTEXT(ctx
);
732 GLfloat du
= ((ctx
->Eval
.MapGrid2u2
- ctx
->Eval
.MapGrid2u1
) /
733 (GLfloat
) ctx
->Eval
.MapGrid2un
);
734 GLfloat dv
= ((ctx
->Eval
.MapGrid2v2
- ctx
->Eval
.MapGrid2v1
) /
735 (GLfloat
) ctx
->Eval
.MapGrid2vn
);
736 GLfloat u
= i
* du
+ ctx
->Eval
.MapGrid2u1
;
737 GLfloat v
= j
* dv
+ ctx
->Eval
.MapGrid2v1
;
739 vbo_exec_EvalCoord2f(u
, v
);
744 * Called via glBegin.
746 static void GLAPIENTRY
747 vbo_exec_Begin(GLenum mode
)
749 GET_CURRENT_CONTEXT(ctx
);
750 struct vbo_context
*vbo
= vbo_context(ctx
);
751 struct vbo_exec_context
*exec
= &vbo
->exec
;
754 if (_mesa_inside_begin_end(ctx
)) {
755 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glBegin");
759 if (!_mesa_valid_prim_mode(ctx
, mode
, "glBegin")) {
764 _mesa_update_state(ctx
);
766 if (!_mesa_valid_to_render(ctx
, "glBegin")) {
770 /* Heuristic: attempt to isolate attributes occurring outside
773 if (exec
->vtx
.vertex_size
&& !exec
->vtx
.attr
[VBO_ATTRIB_POS
].size
)
774 vbo_exec_FlushVertices_internal(exec
, GL_FALSE
);
776 i
= exec
->vtx
.prim_count
++;
777 exec
->vtx
.prim
[i
].mode
= mode
;
778 exec
->vtx
.prim
[i
].begin
= 1;
779 exec
->vtx
.prim
[i
].end
= 0;
780 exec
->vtx
.prim
[i
].indexed
= 0;
781 exec
->vtx
.prim
[i
].pad
= 0;
782 exec
->vtx
.prim
[i
].start
= exec
->vtx
.vert_count
;
783 exec
->vtx
.prim
[i
].count
= 0;
784 exec
->vtx
.prim
[i
].num_instances
= 1;
785 exec
->vtx
.prim
[i
].base_instance
= 0;
786 exec
->vtx
.prim
[i
].is_indirect
= 0;
788 ctx
->Driver
.CurrentExecPrimitive
= mode
;
790 ctx
->Exec
= ctx
->BeginEnd
;
792 /* We may have been called from a display list, in which case we should
793 * leave dlist.c's dispatch table in place.
795 if (ctx
->CurrentClientDispatch
== ctx
->MarshalExec
) {
796 ctx
->CurrentServerDispatch
= ctx
->Exec
;
797 } else if (ctx
->CurrentClientDispatch
== ctx
->OutsideBeginEnd
) {
798 ctx
->CurrentClientDispatch
= ctx
->Exec
;
799 _glapi_set_dispatch(ctx
->CurrentClientDispatch
);
801 assert(ctx
->CurrentClientDispatch
== ctx
->Save
);
807 * Try to merge / concatenate the two most recent VBO primitives.
810 try_vbo_merge(struct vbo_exec_context
*exec
)
812 struct _mesa_prim
*cur
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
814 assert(exec
->vtx
.prim_count
>= 1);
816 vbo_try_prim_conversion(cur
);
818 if (exec
->vtx
.prim_count
>= 2) {
819 struct _mesa_prim
*prev
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 2];
820 assert(prev
== cur
- 1);
822 if (vbo_can_merge_prims(prev
, cur
)) {
827 vbo_merge_prims(prev
, cur
);
828 exec
->vtx
.prim_count
--; /* drop the last primitive */
837 static void GLAPIENTRY
840 GET_CURRENT_CONTEXT(ctx
);
841 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
843 if (!_mesa_inside_begin_end(ctx
)) {
844 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glEnd");
848 ctx
->Exec
= ctx
->OutsideBeginEnd
;
850 if (ctx
->CurrentClientDispatch
== ctx
->MarshalExec
) {
851 ctx
->CurrentServerDispatch
= ctx
->Exec
;
852 } else if (ctx
->CurrentClientDispatch
== ctx
->BeginEnd
) {
853 ctx
->CurrentClientDispatch
= ctx
->Exec
;
854 _glapi_set_dispatch(ctx
->CurrentClientDispatch
);
857 if (exec
->vtx
.prim_count
> 0) {
858 /* close off current primitive */
859 struct _mesa_prim
*last_prim
= &exec
->vtx
.prim
[exec
->vtx
.prim_count
- 1];
862 last_prim
->count
= exec
->vtx
.vert_count
- last_prim
->start
;
864 /* Special handling for GL_LINE_LOOP */
865 if (last_prim
->mode
== GL_LINE_LOOP
&& last_prim
->begin
== 0) {
866 /* We're finishing drawing a line loop. Append 0th vertex onto
867 * end of vertex buffer so we can draw it as a line strip.
869 const fi_type
*src
= exec
->vtx
.buffer_map
+
870 last_prim
->start
* exec
->vtx
.vertex_size
;
871 fi_type
*dst
= exec
->vtx
.buffer_map
+
872 exec
->vtx
.vert_count
* exec
->vtx
.vertex_size
;
874 /* copy 0th vertex to end of buffer */
875 memcpy(dst
, src
, exec
->vtx
.vertex_size
* sizeof(fi_type
));
877 last_prim
->start
++; /* skip vertex0 */
878 /* note that last_prim->count stays unchanged */
879 last_prim
->mode
= GL_LINE_STRIP
;
881 /* Increment the vertex count so the next primitive doesn't
882 * overwrite the last vertex which we just added.
884 exec
->vtx
.vert_count
++;
885 exec
->vtx
.buffer_ptr
+= exec
->vtx
.vertex_size
;
891 ctx
->Driver
.CurrentExecPrimitive
= PRIM_OUTSIDE_BEGIN_END
;
893 if (exec
->vtx
.prim_count
== VBO_MAX_PRIM
)
894 vbo_exec_vtx_flush(exec
, GL_FALSE
);
896 if (MESA_DEBUG_FLAGS
& DEBUG_ALWAYS_FLUSH
) {
903 * Called via glPrimitiveRestartNV()
905 static void GLAPIENTRY
906 vbo_exec_PrimitiveRestartNV(void)
909 GET_CURRENT_CONTEXT(ctx
);
911 curPrim
= ctx
->Driver
.CurrentExecPrimitive
;
913 if (curPrim
== PRIM_OUTSIDE_BEGIN_END
) {
914 _mesa_error(ctx
, GL_INVALID_OPERATION
, "glPrimitiveRestartNV");
918 vbo_exec_Begin(curPrim
);
924 vbo_exec_vtxfmt_init(struct vbo_exec_context
*exec
)
926 struct gl_context
*ctx
= exec
->ctx
;
927 GLvertexformat
*vfmt
= &exec
->vtxfmt
;
929 #define NAME_AE(x) _ae_##x
930 #define NAME_CALLLIST(x) _mesa_##x
931 #define NAME(x) vbo_exec_##x
932 #define NAME_ES(x) _es_##x
934 #include "vbo_init_tmp.h"
939 * Tell the VBO module to use a real OpenGL vertex buffer object to
940 * store accumulated immediate-mode vertex data.
941 * This replaces the malloced buffer which was created in
942 * vb_exec_vtx_init() below.
945 vbo_use_buffer_objects(struct gl_context
*ctx
)
947 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
948 /* Any buffer name but 0 can be used here since this bufferobj won't
949 * go into the bufferobj hashtable.
951 GLuint bufName
= IMM_BUFFER_NAME
;
953 /* Make sure this func is only used once */
954 assert(exec
->vtx
.bufferobj
== ctx
->Shared
->NullBufferObj
);
956 _mesa_align_free(exec
->vtx
.buffer_map
);
957 exec
->vtx
.buffer_map
= NULL
;
958 exec
->vtx
.buffer_ptr
= NULL
;
960 /* Allocate a real buffer object now */
961 _mesa_reference_buffer_object(ctx
, &exec
->vtx
.bufferobj
, NULL
);
962 exec
->vtx
.bufferobj
= ctx
->Driver
.NewBufferObject(ctx
, bufName
);
967 vbo_exec_vtx_init(struct vbo_exec_context
*exec
)
969 struct gl_context
*ctx
= exec
->ctx
;
972 /* Allocate a buffer object. Will just reuse this object
973 * continuously, unless vbo_use_buffer_objects() is called to enable
976 _mesa_reference_buffer_object(ctx
,
977 &exec
->vtx
.bufferobj
,
978 ctx
->Shared
->NullBufferObj
);
980 assert(!exec
->vtx
.buffer_map
);
981 exec
->vtx
.buffer_map
= _mesa_align_malloc(VBO_VERT_BUFFER_SIZE
, 64);
982 exec
->vtx
.buffer_ptr
= exec
->vtx
.buffer_map
;
984 vbo_exec_vtxfmt_init(exec
);
985 _mesa_noop_vtxfmt_init(ctx
, &exec
->vtxfmt_noop
);
987 exec
->vtx
.enabled
= 0;
988 for (i
= 0 ; i
< ARRAY_SIZE(exec
->vtx
.attr
); i
++) {
989 exec
->vtx
.attr
[i
].size
= 0;
990 exec
->vtx
.attr
[i
].type
= GL_FLOAT
;
991 exec
->vtx
.attr
[i
].active_size
= 0;
994 exec
->vtx
.vertex_size
= 0;
999 vbo_exec_vtx_destroy(struct vbo_exec_context
*exec
)
1001 /* using a real VBO for vertex data */
1002 struct gl_context
*ctx
= exec
->ctx
;
1004 /* True VBOs should already be unmapped
1006 if (exec
->vtx
.buffer_map
) {
1007 assert(exec
->vtx
.bufferobj
->Name
== 0 ||
1008 exec
->vtx
.bufferobj
->Name
== IMM_BUFFER_NAME
);
1009 if (exec
->vtx
.bufferobj
->Name
== 0) {
1010 _mesa_align_free(exec
->vtx
.buffer_map
);
1011 exec
->vtx
.buffer_map
= NULL
;
1012 exec
->vtx
.buffer_ptr
= NULL
;
1016 /* Free the vertex buffer. Unmap first if needed.
1018 if (_mesa_bufferobj_mapped(exec
->vtx
.bufferobj
, MAP_INTERNAL
)) {
1019 ctx
->Driver
.UnmapBuffer(ctx
, exec
->vtx
.bufferobj
, MAP_INTERNAL
);
1021 _mesa_reference_buffer_object(ctx
, &exec
->vtx
.bufferobj
, NULL
);
1026 * If inside glBegin()/glEnd(), it should assert(0). Otherwise, if
1027 * FLUSH_STORED_VERTICES bit in \p flags is set flushes any buffered
1028 * vertices, if FLUSH_UPDATE_CURRENT bit is set updates
1029 * __struct gl_contextRec::Current and gl_light_attrib::Material
1031 * Note that the default T&L engine never clears the
1032 * FLUSH_UPDATE_CURRENT bit, even after performing the update.
1034 * \param flags bitmask of FLUSH_STORED_VERTICES, FLUSH_UPDATE_CURRENT
1037 vbo_exec_FlushVertices(struct gl_context
*ctx
, GLuint flags
)
1039 struct vbo_exec_context
*exec
= &vbo_context(ctx
)->exec
;
1042 /* debug check: make sure we don't get called recursively */
1043 exec
->flush_call_depth
++;
1044 assert(exec
->flush_call_depth
== 1);
1047 if (_mesa_inside_begin_end(ctx
)) {
1048 /* We've had glBegin but not glEnd! */
1050 exec
->flush_call_depth
--;
1051 assert(exec
->flush_call_depth
== 0);
1056 /* Flush (draw), and make sure VBO is left unmapped when done */
1057 vbo_exec_FlushVertices_internal(exec
, GL_TRUE
);
1059 /* Clear the dirty flush flags, because the flush is finished. */
1060 ctx
->Driver
.NeedFlush
&= ~(FLUSH_UPDATE_CURRENT
| flags
);
1063 exec
->flush_call_depth
--;
1064 assert(exec
->flush_call_depth
== 0);
1070 * Reset the vertex attribute by setting its size to zero.
1073 vbo_reset_attr(struct vbo_exec_context
*exec
, GLuint attr
)
1075 exec
->vtx
.attr
[attr
].size
= 0;
1076 exec
->vtx
.attr
[attr
].type
= GL_FLOAT
;
1077 exec
->vtx
.attr
[attr
].active_size
= 0;
1082 vbo_reset_all_attr(struct vbo_exec_context
*exec
)
1084 while (exec
->vtx
.enabled
) {
1085 const int i
= u_bit_scan64(&exec
->vtx
.enabled
);
1086 vbo_reset_attr(exec
, i
);
1089 exec
->vtx
.vertex_size
= 0;
1094 _es_Color4f(GLfloat r
, GLfloat g
, GLfloat b
, GLfloat a
)
1096 vbo_exec_Color4f(r
, g
, b
, a
);
1101 _es_Normal3f(GLfloat x
, GLfloat y
, GLfloat z
)
1103 vbo_exec_Normal3f(x
, y
, z
);
1108 _es_MultiTexCoord4f(GLenum target
, GLfloat s
, GLfloat t
, GLfloat r
, GLfloat q
)
1110 vbo_exec_MultiTexCoord4f(target
, s
, t
, r
, q
);
1115 _es_Materialfv(GLenum face
, GLenum pname
, const GLfloat
*params
)
1117 vbo_exec_Materialfv(face
, pname
, params
);
1122 _es_Materialf(GLenum face
, GLenum pname
, GLfloat param
)
1126 p
[1] = p
[2] = p
[3] = 0.0F
;
1127 vbo_exec_Materialfv(face
, pname
, p
);
1132 * A special version of glVertexAttrib4f that does not treat index 0 as
1136 VertexAttrib4f_nopos(GLuint index
, GLfloat x
, GLfloat y
, GLfloat z
, GLfloat w
)
1138 GET_CURRENT_CONTEXT(ctx
);
1139 if (index
< MAX_VERTEX_GENERIC_ATTRIBS
)
1140 ATTRF(VBO_ATTRIB_GENERIC0
+ index
, 4, x
, y
, z
, w
);
1142 ERROR(GL_INVALID_VALUE
);
1146 _es_VertexAttrib4f(GLuint index
, GLfloat x
, GLfloat y
, GLfloat z
, GLfloat w
)
1148 VertexAttrib4f_nopos(index
, x
, y
, z
, w
);
1153 _es_VertexAttrib1f(GLuint indx
, GLfloat x
)
1155 VertexAttrib4f_nopos(indx
, x
, 0.0f
, 0.0f
, 1.0f
);
1160 _es_VertexAttrib1fv(GLuint indx
, const GLfloat
* values
)
1162 VertexAttrib4f_nopos(indx
, values
[0], 0.0f
, 0.0f
, 1.0f
);
1167 _es_VertexAttrib2f(GLuint indx
, GLfloat x
, GLfloat y
)
1169 VertexAttrib4f_nopos(indx
, x
, y
, 0.0f
, 1.0f
);
1174 _es_VertexAttrib2fv(GLuint indx
, const GLfloat
* values
)
1176 VertexAttrib4f_nopos(indx
, values
[0], values
[1], 0.0f
, 1.0f
);
1181 _es_VertexAttrib3f(GLuint indx
, GLfloat x
, GLfloat y
, GLfloat z
)
1183 VertexAttrib4f_nopos(indx
, x
, y
, z
, 1.0f
);
1188 _es_VertexAttrib3fv(GLuint indx
, const GLfloat
* values
)
1190 VertexAttrib4f_nopos(indx
, values
[0], values
[1], values
[2], 1.0f
);
1195 _es_VertexAttrib4fv(GLuint indx
, const GLfloat
* values
)
1197 VertexAttrib4f_nopos(indx
, values
[0], values
[1], values
[2], values
[3]);