mesa: don't ever set NullBufferObj in gl_vertex_array_binding
[mesa.git] / src / mesa / vbo / vbo_exec_api.c
1 /**************************************************************************
2
3 Copyright 2002-2008 VMware, Inc.
4
5 All Rights Reserved.
6
7 Permission is hereby granted, free of charge, to any person obtaining a
8 copy of this software and associated documentation files (the "Software"),
9 to deal in the Software without restriction, including without limitation
10 on the rights to use, copy, modify, merge, publish, distribute, sub
11 license, and/or sell copies of the Software, and to permit persons to whom
12 the Software is furnished to do so, subject to the following conditions:
13
14 The above copyright notice and this permission notice (including the next
15 paragraph) shall be included in all copies or substantial portions of the
16 Software.
17
18 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21 VMWARE AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
22 DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23 OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24 USE OR OTHER DEALINGS IN THE SOFTWARE.
25
26 **************************************************************************/
27
28 /*
29 * Authors:
30 * Keith Whitwell <keithw@vmware.com>
31 */
32
33 #include "main/glheader.h"
34 #include "main/bufferobj.h"
35 #include "main/context.h"
36 #include "main/macros.h"
37 #include "main/vtxfmt.h"
38 #include "main/dlist.h"
39 #include "main/eval.h"
40 #include "main/state.h"
41 #include "main/light.h"
42 #include "main/api_arrayelt.h"
43 #include "main/draw_validate.h"
44 #include "main/dispatch.h"
45 #include "util/bitscan.h"
46
47 #include "vbo_noop.h"
48 #include "vbo_private.h"
49
50
51 /** ID/name for immediate-mode VBO */
52 #define IMM_BUFFER_NAME 0xaabbccdd
53
54
55 static void
56 vbo_reset_all_attr(struct vbo_exec_context *exec);
57
58
59 /**
60 * Close off the last primitive, execute the buffer, restart the
61 * primitive. This is called when we fill a vertex buffer before
62 * hitting glEnd.
63 */
64 static void
65 vbo_exec_wrap_buffers(struct vbo_exec_context *exec)
66 {
67 if (exec->vtx.prim_count == 0) {
68 exec->vtx.copied.nr = 0;
69 exec->vtx.vert_count = 0;
70 exec->vtx.buffer_ptr = exec->vtx.buffer_map;
71 }
72 else {
73 struct _mesa_prim *last_prim = &exec->vtx.prim[exec->vtx.prim_count - 1];
74 const GLuint last_begin = last_prim->begin;
75 GLuint last_count;
76
77 if (_mesa_inside_begin_end(exec->ctx)) {
78 last_prim->count = exec->vtx.vert_count - last_prim->start;
79 }
80
81 last_count = last_prim->count;
82
83 /* Special handling for wrapping GL_LINE_LOOP */
84 if (last_prim->mode == GL_LINE_LOOP &&
85 last_count > 0 &&
86 !last_prim->end) {
87 /* draw this section of the incomplete line loop as a line strip */
88 last_prim->mode = GL_LINE_STRIP;
89 if (!last_prim->begin) {
90 /* This is not the first section of the line loop, so don't
91 * draw the 0th vertex. We're saving it until we draw the
92 * very last section of the loop.
93 */
94 last_prim->start++;
95 last_prim->count--;
96 }
97 }
98
99 /* Execute the buffer and save copied vertices.
100 */
101 if (exec->vtx.vert_count)
102 vbo_exec_vtx_flush(exec);
103 else {
104 exec->vtx.prim_count = 0;
105 exec->vtx.copied.nr = 0;
106 }
107
108 /* Emit a glBegin to start the new list.
109 */
110 assert(exec->vtx.prim_count == 0);
111
112 if (_mesa_inside_begin_end(exec->ctx)) {
113 exec->vtx.prim[0].mode = exec->ctx->Driver.CurrentExecPrimitive;
114 exec->vtx.prim[0].begin = 0;
115 exec->vtx.prim[0].end = 0;
116 exec->vtx.prim[0].start = 0;
117 exec->vtx.prim[0].count = 0;
118 exec->vtx.prim_count++;
119
120 if (exec->vtx.copied.nr == last_count)
121 exec->vtx.prim[0].begin = last_begin;
122 }
123 }
124 }
125
126
127 /**
128 * Deal with buffer wrapping where provoked by the vertex buffer
129 * filling up, as opposed to upgrade_vertex().
130 */
131 static void
132 vbo_exec_vtx_wrap(struct vbo_exec_context *exec)
133 {
134 unsigned numComponents;
135
136 /* Run pipeline on current vertices, copy wrapped vertices
137 * to exec->vtx.copied.
138 */
139 vbo_exec_wrap_buffers(exec);
140
141 if (!exec->vtx.buffer_ptr) {
142 /* probably ran out of memory earlier when allocating the VBO */
143 return;
144 }
145
146 /* Copy stored stored vertices to start of new list.
147 */
148 assert(exec->vtx.max_vert - exec->vtx.vert_count > exec->vtx.copied.nr);
149
150 numComponents = exec->vtx.copied.nr * exec->vtx.vertex_size;
151 memcpy(exec->vtx.buffer_ptr,
152 exec->vtx.copied.buffer,
153 numComponents * sizeof(fi_type));
154 exec->vtx.buffer_ptr += numComponents;
155 exec->vtx.vert_count += exec->vtx.copied.nr;
156
157 exec->vtx.copied.nr = 0;
158 }
159
160
161 /**
162 * Copy the active vertex's values to the ctx->Current fields.
163 */
164 static void
165 vbo_exec_copy_to_current(struct vbo_exec_context *exec)
166 {
167 struct gl_context *ctx = exec->ctx;
168 struct vbo_context *vbo = vbo_context(ctx);
169 GLbitfield64 enabled = exec->vtx.enabled & (~BITFIELD64_BIT(VBO_ATTRIB_POS));
170
171 while (enabled) {
172 const int i = u_bit_scan64(&enabled);
173
174 /* Note: the exec->vtx.current[i] pointers point into the
175 * ctx->Current.Attrib and ctx->Light.Material.Attrib arrays.
176 */
177 GLfloat *current = (GLfloat *)vbo->current[i].Ptr;
178 fi_type tmp[8]; /* space for doubles */
179 int dmul = 1;
180
181 if (exec->vtx.attr[i].type == GL_DOUBLE ||
182 exec->vtx.attr[i].type == GL_UNSIGNED_INT64_ARB)
183 dmul = 2;
184
185 assert(exec->vtx.attr[i].size);
186
187 if (exec->vtx.attr[i].type == GL_DOUBLE ||
188 exec->vtx.attr[i].type == GL_UNSIGNED_INT64_ARB) {
189 memset(tmp, 0, sizeof(tmp));
190 memcpy(tmp, exec->vtx.attrptr[i], exec->vtx.attr[i].size * sizeof(GLfloat));
191 } else {
192 COPY_CLEAN_4V_TYPE_AS_UNION(tmp,
193 exec->vtx.attr[i].size,
194 exec->vtx.attrptr[i],
195 exec->vtx.attr[i].type);
196 }
197
198 if (exec->vtx.attr[i].type != vbo->current[i].Format.Type ||
199 memcmp(current, tmp, 4 * sizeof(GLfloat) * dmul) != 0) {
200 memcpy(current, tmp, 4 * sizeof(GLfloat) * dmul);
201
202 /* Given that we explicitly state size here, there is no need
203 * for the COPY_CLEAN above, could just copy 16 bytes and be
204 * done. The only problem is when Mesa accesses ctx->Current
205 * directly.
206 */
207 /* Size here is in components - not bytes */
208 vbo_set_vertex_format(&vbo->current[i].Format,
209 exec->vtx.attr[i].size / dmul,
210 exec->vtx.attr[i].type);
211
212 /* This triggers rather too much recalculation of Mesa state
213 * that doesn't get used (eg light positions).
214 */
215 if (i >= VBO_ATTRIB_MAT_FRONT_AMBIENT &&
216 i <= VBO_ATTRIB_MAT_BACK_INDEXES)
217 ctx->NewState |= _NEW_LIGHT;
218
219 ctx->NewState |= _NEW_CURRENT_ATTRIB;
220 }
221 }
222
223 /* Colormaterial -- this kindof sucks.
224 */
225 if (ctx->Light.ColorMaterialEnabled &&
226 exec->vtx.attr[VBO_ATTRIB_COLOR0].size) {
227 _mesa_update_color_material(ctx,
228 ctx->Current.Attrib[VBO_ATTRIB_COLOR0]);
229 }
230 }
231
232
233 /**
234 * Flush existing data, set new attrib size, replay copied vertices.
235 * This is called when we transition from a small vertex attribute size
236 * to a larger one. Ex: glTexCoord2f -> glTexCoord4f.
237 * We need to go back over the previous 2-component texcoords and insert
238 * zero and one values.
239 * \param attr VBO_ATTRIB_x vertex attribute value
240 */
241 static void
242 vbo_exec_wrap_upgrade_vertex(struct vbo_exec_context *exec,
243 GLuint attr, GLuint newSize, GLenum newType)
244 {
245 struct gl_context *ctx = exec->ctx;
246 struct vbo_context *vbo = vbo_context(ctx);
247 const GLint lastcount = exec->vtx.vert_count;
248 fi_type *old_attrptr[VBO_ATTRIB_MAX];
249 const GLuint old_vtx_size_no_pos = exec->vtx.vertex_size_no_pos;
250 const GLuint old_vtx_size = exec->vtx.vertex_size; /* floats per vertex */
251 const GLuint oldSize = exec->vtx.attr[attr].size;
252 GLuint i;
253
254 assert(attr < VBO_ATTRIB_MAX);
255
256 /* Run pipeline on current vertices, copy wrapped vertices
257 * to exec->vtx.copied.
258 */
259 vbo_exec_wrap_buffers(exec);
260
261 if (unlikely(exec->vtx.copied.nr)) {
262 /* We're in the middle of a primitive, keep the old vertex
263 * format around to be able to translate the copied vertices to
264 * the new format.
265 */
266 memcpy(old_attrptr, exec->vtx.attrptr, sizeof(old_attrptr));
267 }
268
269 /* Heuristic: Attempt to isolate attributes received outside
270 * begin/end so that they don't bloat the vertices.
271 */
272 if (!_mesa_inside_begin_end(ctx) &&
273 !oldSize && lastcount > 8 && exec->vtx.vertex_size) {
274 vbo_exec_copy_to_current(exec);
275 vbo_reset_all_attr(exec);
276 }
277
278 /* Fix up sizes:
279 */
280 exec->vtx.attr[attr].size = newSize;
281 exec->vtx.attr[attr].active_size = newSize;
282 exec->vtx.attr[attr].type = newType;
283 exec->vtx.vertex_size += newSize - oldSize;
284 exec->vtx.vertex_size_no_pos = exec->vtx.vertex_size - exec->vtx.attr[0].size;
285 exec->vtx.max_vert = vbo_compute_max_verts(exec);
286 exec->vtx.vert_count = 0;
287 exec->vtx.buffer_ptr = exec->vtx.buffer_map;
288 exec->vtx.enabled |= BITFIELD64_BIT(attr);
289
290 if (attr != 0) {
291 if (unlikely(oldSize)) {
292 unsigned offset = exec->vtx.attrptr[attr] - exec->vtx.vertex;
293
294 /* If there are attribs after the resized attrib... */
295 if (offset + oldSize < old_vtx_size_no_pos) {
296 int size_diff = newSize - oldSize;
297 fi_type *old_first = exec->vtx.attrptr[attr] + oldSize;
298 fi_type *new_first = exec->vtx.attrptr[attr] + newSize;
299 fi_type *old_last = exec->vtx.vertex + old_vtx_size_no_pos - 1;
300 fi_type *new_last = exec->vtx.vertex + exec->vtx.vertex_size_no_pos - 1;
301
302 if (size_diff < 0) {
303 /* Decreasing the size: Copy from first to last to move
304 * elements to the left.
305 */
306 fi_type *old_end = old_last + 1;
307 fi_type *old = old_first;
308 fi_type *new = new_first;
309
310 do {
311 *new++ = *old++;
312 } while (old != old_end);
313 } else {
314 /* Increasing the size: Copy from last to first to move
315 * elements to the right.
316 */
317 fi_type *old_end = old_first - 1;
318 fi_type *old = old_last;
319 fi_type *new = new_last;
320
321 do {
322 *new-- = *old--;
323 } while (old != old_end);
324 }
325
326 /* Update pointers to attribs, because we moved them. */
327 GLbitfield64 enabled = exec->vtx.enabled &
328 ~BITFIELD64_BIT(VBO_ATTRIB_POS) &
329 ~BITFIELD64_BIT(attr);
330 while (enabled) {
331 unsigned i = u_bit_scan64(&enabled);
332
333 if (exec->vtx.attrptr[i] > exec->vtx.attrptr[attr])
334 exec->vtx.attrptr[i] += size_diff;
335 }
336 }
337 } else {
338 /* Just have to append the new attribute at the end */
339 exec->vtx.attrptr[attr] = exec->vtx.vertex +
340 exec->vtx.vertex_size_no_pos - newSize;
341 }
342 }
343
344 /* The position is always last. */
345 exec->vtx.attrptr[0] = exec->vtx.vertex + exec->vtx.vertex_size_no_pos;
346
347 /* Replay stored vertices to translate them
348 * to new format here.
349 *
350 * -- No need to replay - just copy piecewise
351 */
352 if (unlikely(exec->vtx.copied.nr)) {
353 fi_type *data = exec->vtx.copied.buffer;
354 fi_type *dest = exec->vtx.buffer_ptr;
355
356 assert(exec->vtx.buffer_ptr == exec->vtx.buffer_map);
357
358 for (i = 0 ; i < exec->vtx.copied.nr ; i++) {
359 GLbitfield64 enabled = exec->vtx.enabled;
360 while (enabled) {
361 const int j = u_bit_scan64(&enabled);
362 GLuint sz = exec->vtx.attr[j].size;
363 GLint old_offset = old_attrptr[j] - exec->vtx.vertex;
364 GLint new_offset = exec->vtx.attrptr[j] - exec->vtx.vertex;
365
366 assert(sz);
367
368 if (j == attr) {
369 if (oldSize) {
370 fi_type tmp[4];
371 COPY_CLEAN_4V_TYPE_AS_UNION(tmp, oldSize,
372 data + old_offset,
373 exec->vtx.attr[j].type);
374 COPY_SZ_4V(dest + new_offset, newSize, tmp);
375 } else {
376 fi_type *current = (fi_type *)vbo->current[j].Ptr;
377 COPY_SZ_4V(dest + new_offset, sz, current);
378 }
379 }
380 else {
381 COPY_SZ_4V(dest + new_offset, sz, data + old_offset);
382 }
383 }
384
385 data += old_vtx_size;
386 dest += exec->vtx.vertex_size;
387 }
388
389 exec->vtx.buffer_ptr = dest;
390 exec->vtx.vert_count += exec->vtx.copied.nr;
391 exec->vtx.copied.nr = 0;
392 }
393 }
394
395
396 /**
397 * This is when a vertex attribute transitions to a different size.
398 * For example, we saw a bunch of glTexCoord2f() calls and now we got a
399 * glTexCoord4f() call. We promote the array from size=2 to size=4.
400 * \param newSize size of new vertex (number of 32-bit words).
401 * \param attr VBO_ATTRIB_x vertex attribute value
402 */
403 static void
404 vbo_exec_fixup_vertex(struct gl_context *ctx, GLuint attr,
405 GLuint newSize, GLenum newType)
406 {
407 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
408
409 assert(attr < VBO_ATTRIB_MAX);
410
411 if (newSize > exec->vtx.attr[attr].size ||
412 newType != exec->vtx.attr[attr].type) {
413 /* New size is larger. Need to flush existing vertices and get
414 * an enlarged vertex format.
415 */
416 vbo_exec_wrap_upgrade_vertex(exec, attr, newSize, newType);
417 }
418 else if (newSize < exec->vtx.attr[attr].active_size) {
419 GLuint i;
420 const fi_type *id =
421 vbo_get_default_vals_as_union(exec->vtx.attr[attr].type);
422
423 /* New size is smaller - just need to fill in some
424 * zeros. Don't need to flush or wrap.
425 */
426 for (i = newSize; i <= exec->vtx.attr[attr].size; i++)
427 exec->vtx.attrptr[attr][i-1] = id[i-1];
428
429 exec->vtx.attr[attr].active_size = newSize;
430 }
431 }
432
433
434 /**
435 * If index=0, does glVertexAttrib*() alias glVertex() to emit a vertex?
436 * It depends on a few things, including whether we're inside or outside
437 * of glBegin/glEnd.
438 */
439 static inline bool
440 is_vertex_position(const struct gl_context *ctx, GLuint index)
441 {
442 return (index == 0 &&
443 _mesa_attr_zero_aliases_vertex(ctx) &&
444 _mesa_inside_begin_end(ctx));
445 }
446
447 /* Write a 64-bit value into a 32-bit pointer by preserving endianness. */
448 #if UTIL_ARCH_LITTLE_ENDIAN
449 #define SET_64BIT(dst32, u64) do { \
450 *(dst32)++ = (u64); \
451 *(dst32)++ = (uint64_t)(u64) >> 32; \
452 } while (0)
453 #else
454 #define SET_64BIT(dst32, u64) do { \
455 *(dst32)++ = (uint64_t)(u64) >> 32; \
456 *(dst32)++ = (u64); \
457 } while (0)
458 #endif
459
460
461 /**
462 * This macro is used to implement all the glVertex, glColor, glTexCoord,
463 * glVertexAttrib, etc functions.
464 * \param A VBO_ATTRIB_x attribute index
465 * \param N attribute size (1..4)
466 * \param T type (GL_FLOAT, GL_DOUBLE, GL_INT, GL_UNSIGNED_INT)
467 * \param C cast type (uint32_t or uint64_t)
468 * \param V0, V1, v2, V3 attribute value
469 */
470 #define ATTR_UNION(A, N, T, C, V0, V1, V2, V3) \
471 do { \
472 struct vbo_exec_context *exec = &vbo_context(ctx)->exec; \
473 int sz = (sizeof(C) / sizeof(GLfloat)); \
474 \
475 assert(sz == 1 || sz == 2); \
476 \
477 /* store a copy of the attribute in exec except for glVertex */ \
478 if ((A) != 0) { \
479 /* Check if attribute size or type is changing. */ \
480 if (unlikely(exec->vtx.attr[A].active_size != N * sz || \
481 exec->vtx.attr[A].type != T)) { \
482 vbo_exec_fixup_vertex(ctx, A, N * sz, T); \
483 } \
484 \
485 C *dest = (C *)exec->vtx.attrptr[A]; \
486 if (N>0) dest[0] = V0; \
487 if (N>1) dest[1] = V1; \
488 if (N>2) dest[2] = V2; \
489 if (N>3) dest[3] = V3; \
490 assert(exec->vtx.attr[A].type == T); \
491 \
492 /* we now have accumulated a per-vertex attribute */ \
493 ctx->Driver.NeedFlush |= FLUSH_UPDATE_CURRENT; \
494 } else { \
495 /* This is a glVertex call */ \
496 int size = exec->vtx.attr[0].size; \
497 \
498 /* Check if attribute size or type is changing. */ \
499 if (unlikely(size < N * sz || \
500 exec->vtx.attr[0].type != T)) { \
501 vbo_exec_wrap_upgrade_vertex(exec, 0, N * sz, T); \
502 } \
503 \
504 uint32_t *dst = (uint32_t *)exec->vtx.buffer_ptr; \
505 uint32_t *src = (uint32_t *)exec->vtx.vertex; \
506 unsigned vertex_size_no_pos = exec->vtx.vertex_size_no_pos; \
507 \
508 /* Copy over attributes from exec. */ \
509 for (unsigned i = 0; i < vertex_size_no_pos; i++) \
510 *dst++ = *src++; \
511 \
512 /* Store the position, which is always last and can have 32 or */ \
513 /* 64 bits per channel. */ \
514 if (sizeof(C) == 4) { \
515 if (N > 0) *dst++ = V0; \
516 if (N > 1) *dst++ = V1; \
517 if (N > 2) *dst++ = V2; \
518 if (N > 3) *dst++ = V3; \
519 \
520 if (unlikely(N < size)) { \
521 if (N < 2 && size >= 2) *dst++ = V1; \
522 if (N < 3 && size >= 3) *dst++ = V2; \
523 if (N < 4 && size >= 4) *dst++ = V3; \
524 } \
525 } else { \
526 /* 64 bits: dst can be unaligned, so copy each 4-byte word */ \
527 /* separately */ \
528 if (N > 0) SET_64BIT(dst, V0); \
529 if (N > 1) SET_64BIT(dst, V1); \
530 if (N > 2) SET_64BIT(dst, V2); \
531 if (N > 3) SET_64BIT(dst, V3); \
532 \
533 if (unlikely(N * 2 < size)) { \
534 if (N < 2 && size >= 4) SET_64BIT(dst, V1); \
535 if (N < 3 && size >= 6) SET_64BIT(dst, V2); \
536 if (N < 4 && size >= 8) SET_64BIT(dst, V3); \
537 } \
538 } \
539 \
540 /* dst now points at the beginning of the next vertex */ \
541 exec->vtx.buffer_ptr = (fi_type*)dst; \
542 \
543 /* Don't set FLUSH_UPDATE_CURRENT because */ \
544 /* Current.Attrib[VBO_ATTRIB_POS] is never used. */ \
545 \
546 if (unlikely(++exec->vtx.vert_count >= exec->vtx.max_vert)) \
547 vbo_exec_vtx_wrap(exec); \
548 } \
549 } while (0)
550
551
552 #undef ERROR
553 #define ERROR(err) _mesa_error(ctx, err, __func__)
554 #define TAG(x) vbo_exec_##x
555
556 #include "vbo_attrib_tmp.h"
557
558
559
560 /**
561 * Execute a glMaterial call. Note that if GL_COLOR_MATERIAL is enabled,
562 * this may be a (partial) no-op.
563 */
564 static void GLAPIENTRY
565 vbo_exec_Materialfv(GLenum face, GLenum pname, const GLfloat *params)
566 {
567 GLbitfield updateMats;
568 GET_CURRENT_CONTEXT(ctx);
569
570 /* This function should be a no-op when it tries to update material
571 * attributes which are currently tracking glColor via glColorMaterial.
572 * The updateMats var will be a mask of the MAT_BIT_FRONT/BACK_x bits
573 * indicating which material attributes can actually be updated below.
574 */
575 if (ctx->Light.ColorMaterialEnabled) {
576 updateMats = ~ctx->Light._ColorMaterialBitmask;
577 }
578 else {
579 /* GL_COLOR_MATERIAL is disabled so don't skip any material updates */
580 updateMats = ALL_MATERIAL_BITS;
581 }
582
583 if (ctx->API == API_OPENGL_COMPAT && face == GL_FRONT) {
584 updateMats &= FRONT_MATERIAL_BITS;
585 }
586 else if (ctx->API == API_OPENGL_COMPAT && face == GL_BACK) {
587 updateMats &= BACK_MATERIAL_BITS;
588 }
589 else if (face != GL_FRONT_AND_BACK) {
590 _mesa_error(ctx, GL_INVALID_ENUM, "glMaterial(invalid face)");
591 return;
592 }
593
594 switch (pname) {
595 case GL_EMISSION:
596 if (updateMats & MAT_BIT_FRONT_EMISSION)
597 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_EMISSION, 4, params);
598 if (updateMats & MAT_BIT_BACK_EMISSION)
599 MAT_ATTR(VBO_ATTRIB_MAT_BACK_EMISSION, 4, params);
600 break;
601 case GL_AMBIENT:
602 if (updateMats & MAT_BIT_FRONT_AMBIENT)
603 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT, 4, params);
604 if (updateMats & MAT_BIT_BACK_AMBIENT)
605 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT, 4, params);
606 break;
607 case GL_DIFFUSE:
608 if (updateMats & MAT_BIT_FRONT_DIFFUSE)
609 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE, 4, params);
610 if (updateMats & MAT_BIT_BACK_DIFFUSE)
611 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE, 4, params);
612 break;
613 case GL_SPECULAR:
614 if (updateMats & MAT_BIT_FRONT_SPECULAR)
615 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SPECULAR, 4, params);
616 if (updateMats & MAT_BIT_BACK_SPECULAR)
617 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SPECULAR, 4, params);
618 break;
619 case GL_SHININESS:
620 if (*params < 0 || *params > ctx->Const.MaxShininess) {
621 _mesa_error(ctx, GL_INVALID_VALUE,
622 "glMaterial(invalid shininess: %f out range [0, %f])",
623 *params, ctx->Const.MaxShininess);
624 return;
625 }
626 if (updateMats & MAT_BIT_FRONT_SHININESS)
627 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_SHININESS, 1, params);
628 if (updateMats & MAT_BIT_BACK_SHININESS)
629 MAT_ATTR(VBO_ATTRIB_MAT_BACK_SHININESS, 1, params);
630 break;
631 case GL_COLOR_INDEXES:
632 if (ctx->API != API_OPENGL_COMPAT) {
633 _mesa_error(ctx, GL_INVALID_ENUM, "glMaterialfv(pname)");
634 return;
635 }
636 if (updateMats & MAT_BIT_FRONT_INDEXES)
637 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_INDEXES, 3, params);
638 if (updateMats & MAT_BIT_BACK_INDEXES)
639 MAT_ATTR(VBO_ATTRIB_MAT_BACK_INDEXES, 3, params);
640 break;
641 case GL_AMBIENT_AND_DIFFUSE:
642 if (updateMats & MAT_BIT_FRONT_AMBIENT)
643 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_AMBIENT, 4, params);
644 if (updateMats & MAT_BIT_FRONT_DIFFUSE)
645 MAT_ATTR(VBO_ATTRIB_MAT_FRONT_DIFFUSE, 4, params);
646 if (updateMats & MAT_BIT_BACK_AMBIENT)
647 MAT_ATTR(VBO_ATTRIB_MAT_BACK_AMBIENT, 4, params);
648 if (updateMats & MAT_BIT_BACK_DIFFUSE)
649 MAT_ATTR(VBO_ATTRIB_MAT_BACK_DIFFUSE, 4, params);
650 break;
651 default:
652 _mesa_error(ctx, GL_INVALID_ENUM, "glMaterialfv(pname)");
653 return;
654 }
655 }
656
657
658 /**
659 * Flush (draw) vertices.
660 *
661 * \param flags bitmask of FLUSH_STORED_VERTICES, FLUSH_UPDATE_CURRENT
662 */
663 static void
664 vbo_exec_FlushVertices_internal(struct vbo_exec_context *exec, unsigned flags)
665 {
666 struct gl_context *ctx = exec->ctx;
667
668 if (flags & FLUSH_STORED_VERTICES) {
669 if (exec->vtx.vert_count) {
670 vbo_exec_vtx_flush(exec);
671 }
672
673 if (exec->vtx.vertex_size) {
674 vbo_exec_copy_to_current(exec);
675 vbo_reset_all_attr(exec);
676 }
677
678 /* All done. */
679 ctx->Driver.NeedFlush = 0;
680 } else {
681 assert(flags == FLUSH_UPDATE_CURRENT);
682
683 /* Note that the vertex size is unchanged.
684 * (vbo_reset_all_attr isn't called)
685 */
686 vbo_exec_copy_to_current(exec);
687
688 /* Only FLUSH_UPDATE_CURRENT is done. */
689 ctx->Driver.NeedFlush = ~FLUSH_UPDATE_CURRENT;
690 }
691 }
692
693
694 static void GLAPIENTRY
695 vbo_exec_EvalCoord1f(GLfloat u)
696 {
697 GET_CURRENT_CONTEXT(ctx);
698 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
699
700 {
701 GLint i;
702 if (exec->eval.recalculate_maps)
703 vbo_exec_eval_update(exec);
704
705 for (i = 0; i <= VBO_ATTRIB_TEX7; i++) {
706 if (exec->eval.map1[i].map)
707 if (exec->vtx.attr[i].active_size != exec->eval.map1[i].sz)
708 vbo_exec_fixup_vertex(ctx, i, exec->eval.map1[i].sz, GL_FLOAT);
709 }
710 }
711
712 memcpy(exec->vtx.copied.buffer, exec->vtx.vertex,
713 exec->vtx.vertex_size * sizeof(GLfloat));
714
715 vbo_exec_do_EvalCoord1f(exec, u);
716
717 memcpy(exec->vtx.vertex, exec->vtx.copied.buffer,
718 exec->vtx.vertex_size * sizeof(GLfloat));
719 }
720
721
722 static void GLAPIENTRY
723 vbo_exec_EvalCoord2f(GLfloat u, GLfloat v)
724 {
725 GET_CURRENT_CONTEXT(ctx);
726 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
727
728 {
729 GLint i;
730 if (exec->eval.recalculate_maps)
731 vbo_exec_eval_update(exec);
732
733 for (i = 0; i <= VBO_ATTRIB_TEX7; i++) {
734 if (exec->eval.map2[i].map)
735 if (exec->vtx.attr[i].active_size != exec->eval.map2[i].sz)
736 vbo_exec_fixup_vertex(ctx, i, exec->eval.map2[i].sz, GL_FLOAT);
737 }
738
739 if (ctx->Eval.AutoNormal)
740 if (exec->vtx.attr[VBO_ATTRIB_NORMAL].active_size != 3)
741 vbo_exec_fixup_vertex(ctx, VBO_ATTRIB_NORMAL, 3, GL_FLOAT);
742 }
743
744 memcpy(exec->vtx.copied.buffer, exec->vtx.vertex,
745 exec->vtx.vertex_size * sizeof(GLfloat));
746
747 vbo_exec_do_EvalCoord2f(exec, u, v);
748
749 memcpy(exec->vtx.vertex, exec->vtx.copied.buffer,
750 exec->vtx.vertex_size * sizeof(GLfloat));
751 }
752
753
754 static void GLAPIENTRY
755 vbo_exec_EvalCoord1fv(const GLfloat *u)
756 {
757 vbo_exec_EvalCoord1f(u[0]);
758 }
759
760
761 static void GLAPIENTRY
762 vbo_exec_EvalCoord2fv(const GLfloat *u)
763 {
764 vbo_exec_EvalCoord2f(u[0], u[1]);
765 }
766
767
768 static void GLAPIENTRY
769 vbo_exec_EvalPoint1(GLint i)
770 {
771 GET_CURRENT_CONTEXT(ctx);
772 GLfloat du = ((ctx->Eval.MapGrid1u2 - ctx->Eval.MapGrid1u1) /
773 (GLfloat) ctx->Eval.MapGrid1un);
774 GLfloat u = i * du + ctx->Eval.MapGrid1u1;
775
776 vbo_exec_EvalCoord1f(u);
777 }
778
779
780 static void GLAPIENTRY
781 vbo_exec_EvalPoint2(GLint i, GLint j)
782 {
783 GET_CURRENT_CONTEXT(ctx);
784 GLfloat du = ((ctx->Eval.MapGrid2u2 - ctx->Eval.MapGrid2u1) /
785 (GLfloat) ctx->Eval.MapGrid2un);
786 GLfloat dv = ((ctx->Eval.MapGrid2v2 - ctx->Eval.MapGrid2v1) /
787 (GLfloat) ctx->Eval.MapGrid2vn);
788 GLfloat u = i * du + ctx->Eval.MapGrid2u1;
789 GLfloat v = j * dv + ctx->Eval.MapGrid2v1;
790
791 vbo_exec_EvalCoord2f(u, v);
792 }
793
794
795 /**
796 * Called via glBegin.
797 */
798 static void GLAPIENTRY
799 vbo_exec_Begin(GLenum mode)
800 {
801 GET_CURRENT_CONTEXT(ctx);
802 struct vbo_context *vbo = vbo_context(ctx);
803 struct vbo_exec_context *exec = &vbo->exec;
804 int i;
805
806 if (_mesa_inside_begin_end(ctx)) {
807 _mesa_error(ctx, GL_INVALID_OPERATION, "glBegin");
808 return;
809 }
810
811 if (!_mesa_valid_prim_mode(ctx, mode, "glBegin")) {
812 return;
813 }
814
815 if (!_mesa_valid_to_render(ctx, "glBegin")) {
816 return;
817 }
818
819 /* Heuristic: attempt to isolate attributes occurring outside
820 * begin/end pairs.
821 *
822 * Use FLUSH_STORED_VERTICES, because it updates current attribs and
823 * sets vertex_size to 0. (FLUSH_UPDATE_CURRENT doesn't change vertex_size)
824 */
825 if (exec->vtx.vertex_size && !exec->vtx.attr[VBO_ATTRIB_POS].size)
826 vbo_exec_FlushVertices_internal(exec, FLUSH_STORED_VERTICES);
827
828 i = exec->vtx.prim_count++;
829 exec->vtx.prim[i].mode = mode;
830 exec->vtx.prim[i].begin = 1;
831 exec->vtx.prim[i].end = 0;
832 exec->vtx.prim[i].start = exec->vtx.vert_count;
833 exec->vtx.prim[i].count = 0;
834
835 ctx->Driver.CurrentExecPrimitive = mode;
836
837 ctx->Exec = ctx->BeginEnd;
838
839 /* We may have been called from a display list, in which case we should
840 * leave dlist.c's dispatch table in place.
841 */
842 if (ctx->CurrentClientDispatch == ctx->MarshalExec) {
843 ctx->CurrentServerDispatch = ctx->Exec;
844 } else if (ctx->CurrentClientDispatch == ctx->OutsideBeginEnd) {
845 ctx->CurrentClientDispatch = ctx->Exec;
846 _glapi_set_dispatch(ctx->CurrentClientDispatch);
847 } else {
848 assert(ctx->CurrentClientDispatch == ctx->Save);
849 }
850 }
851
852
853 /**
854 * Try to merge / concatenate the two most recent VBO primitives.
855 */
856 static void
857 try_vbo_merge(struct vbo_exec_context *exec)
858 {
859 struct _mesa_prim *cur = &exec->vtx.prim[exec->vtx.prim_count - 1];
860
861 assert(exec->vtx.prim_count >= 1);
862
863 vbo_try_prim_conversion(cur);
864
865 if (exec->vtx.prim_count >= 2) {
866 struct _mesa_prim *prev = &exec->vtx.prim[exec->vtx.prim_count - 2];
867 assert(prev == cur - 1);
868
869 if (vbo_merge_draws(exec->ctx, false, prev, cur))
870 exec->vtx.prim_count--; /* drop the last primitive */
871 }
872 }
873
874
875 /**
876 * Called via glEnd.
877 */
878 static void GLAPIENTRY
879 vbo_exec_End(void)
880 {
881 GET_CURRENT_CONTEXT(ctx);
882 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
883
884 if (!_mesa_inside_begin_end(ctx)) {
885 _mesa_error(ctx, GL_INVALID_OPERATION, "glEnd");
886 return;
887 }
888
889 ctx->Exec = ctx->OutsideBeginEnd;
890
891 if (ctx->CurrentClientDispatch == ctx->MarshalExec) {
892 ctx->CurrentServerDispatch = ctx->Exec;
893 } else if (ctx->CurrentClientDispatch == ctx->BeginEnd) {
894 ctx->CurrentClientDispatch = ctx->Exec;
895 _glapi_set_dispatch(ctx->CurrentClientDispatch);
896 }
897
898 if (exec->vtx.prim_count > 0) {
899 /* close off current primitive */
900 struct _mesa_prim *last_prim = &exec->vtx.prim[exec->vtx.prim_count - 1];
901 unsigned count = exec->vtx.vert_count - last_prim->start;
902
903 last_prim->end = 1;
904 last_prim->count = count;
905
906 if (count)
907 ctx->Driver.NeedFlush |= FLUSH_STORED_VERTICES;
908
909 /* Special handling for GL_LINE_LOOP */
910 if (last_prim->mode == GL_LINE_LOOP && last_prim->begin == 0) {
911 /* We're finishing drawing a line loop. Append 0th vertex onto
912 * end of vertex buffer so we can draw it as a line strip.
913 */
914 const fi_type *src = exec->vtx.buffer_map +
915 last_prim->start * exec->vtx.vertex_size;
916 fi_type *dst = exec->vtx.buffer_map +
917 exec->vtx.vert_count * exec->vtx.vertex_size;
918
919 /* copy 0th vertex to end of buffer */
920 memcpy(dst, src, exec->vtx.vertex_size * sizeof(fi_type));
921
922 last_prim->start++; /* skip vertex0 */
923 /* note that last_prim->count stays unchanged */
924 last_prim->mode = GL_LINE_STRIP;
925
926 /* Increment the vertex count so the next primitive doesn't
927 * overwrite the last vertex which we just added.
928 */
929 exec->vtx.vert_count++;
930 exec->vtx.buffer_ptr += exec->vtx.vertex_size;
931 }
932
933 try_vbo_merge(exec);
934 }
935
936 ctx->Driver.CurrentExecPrimitive = PRIM_OUTSIDE_BEGIN_END;
937
938 if (exec->vtx.prim_count == VBO_MAX_PRIM)
939 vbo_exec_vtx_flush(exec);
940
941 if (MESA_DEBUG_FLAGS & DEBUG_ALWAYS_FLUSH) {
942 _mesa_flush(ctx);
943 }
944 }
945
946
947 /**
948 * Called via glPrimitiveRestartNV()
949 */
950 static void GLAPIENTRY
951 vbo_exec_PrimitiveRestartNV(void)
952 {
953 GLenum curPrim;
954 GET_CURRENT_CONTEXT(ctx);
955
956 curPrim = ctx->Driver.CurrentExecPrimitive;
957
958 if (curPrim == PRIM_OUTSIDE_BEGIN_END) {
959 _mesa_error(ctx, GL_INVALID_OPERATION, "glPrimitiveRestartNV");
960 }
961 else {
962 vbo_exec_End();
963 vbo_exec_Begin(curPrim);
964 }
965 }
966
967
968 static void
969 vbo_exec_vtxfmt_init(struct vbo_exec_context *exec)
970 {
971 struct gl_context *ctx = exec->ctx;
972 GLvertexformat *vfmt = &exec->vtxfmt;
973
974 #define NAME_AE(x) _ae_##x
975 #define NAME_CALLLIST(x) _mesa_##x
976 #define NAME(x) vbo_exec_##x
977 #define NAME_ES(x) _es_##x
978
979 #include "vbo_init_tmp.h"
980 }
981
982
983 static void
984 vbo_reset_all_attr(struct vbo_exec_context *exec)
985 {
986 while (exec->vtx.enabled) {
987 const int i = u_bit_scan64(&exec->vtx.enabled);
988
989 /* Reset the vertex attribute by setting its size to zero. */
990 exec->vtx.attr[i].size = 0;
991 exec->vtx.attr[i].type = GL_FLOAT;
992 exec->vtx.attr[i].active_size = 0;
993 exec->vtx.attrptr[i] = NULL;
994 }
995
996 exec->vtx.vertex_size = 0;
997 }
998
999
1000 void
1001 vbo_exec_vtx_init(struct vbo_exec_context *exec, bool use_buffer_objects)
1002 {
1003 struct gl_context *ctx = exec->ctx;
1004
1005 if (use_buffer_objects) {
1006 /* Use buffer objects for immediate mode. */
1007 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
1008
1009 exec->vtx.bufferobj = ctx->Driver.NewBufferObject(ctx, IMM_BUFFER_NAME);
1010
1011 /* Map the buffer. */
1012 vbo_exec_vtx_map(exec);
1013 assert(exec->vtx.buffer_ptr);
1014 } else {
1015 /* Use allocated memory for immediate mode. */
1016 exec->vtx.bufferobj = NULL;
1017 exec->vtx.buffer_map =
1018 _mesa_align_malloc(ctx->Const.glBeginEndBufferSize, 64);
1019 exec->vtx.buffer_ptr = exec->vtx.buffer_map;
1020 }
1021
1022 vbo_exec_vtxfmt_init(exec);
1023 _mesa_noop_vtxfmt_init(ctx, &exec->vtxfmt_noop);
1024
1025 exec->vtx.enabled = u_bit_consecutive64(0, VBO_ATTRIB_MAX); /* reset all */
1026 vbo_reset_all_attr(exec);
1027 }
1028
1029
1030 void
1031 vbo_exec_vtx_destroy(struct vbo_exec_context *exec)
1032 {
1033 /* using a real VBO for vertex data */
1034 struct gl_context *ctx = exec->ctx;
1035
1036 /* True VBOs should already be unmapped
1037 */
1038 if (exec->vtx.buffer_map) {
1039 assert(!exec->vtx.bufferobj ||
1040 exec->vtx.bufferobj->Name == IMM_BUFFER_NAME);
1041 if (!exec->vtx.bufferobj) {
1042 _mesa_align_free(exec->vtx.buffer_map);
1043 exec->vtx.buffer_map = NULL;
1044 exec->vtx.buffer_ptr = NULL;
1045 }
1046 }
1047
1048 /* Free the vertex buffer. Unmap first if needed.
1049 */
1050 if (exec->vtx.bufferobj &&
1051 _mesa_bufferobj_mapped(exec->vtx.bufferobj, MAP_INTERNAL)) {
1052 ctx->Driver.UnmapBuffer(ctx, exec->vtx.bufferobj, MAP_INTERNAL);
1053 }
1054 _mesa_reference_buffer_object(ctx, &exec->vtx.bufferobj, NULL);
1055 }
1056
1057
1058 /**
1059 * If inside glBegin()/glEnd(), it should assert(0). Otherwise, if
1060 * FLUSH_STORED_VERTICES bit in \p flags is set flushes any buffered
1061 * vertices, if FLUSH_UPDATE_CURRENT bit is set updates
1062 * __struct gl_contextRec::Current and gl_light_attrib::Material
1063 *
1064 * Note that the default T&L engine never clears the
1065 * FLUSH_UPDATE_CURRENT bit, even after performing the update.
1066 *
1067 * \param flags bitmask of FLUSH_STORED_VERTICES, FLUSH_UPDATE_CURRENT
1068 */
1069 void
1070 vbo_exec_FlushVertices(struct gl_context *ctx, GLuint flags)
1071 {
1072 struct vbo_exec_context *exec = &vbo_context(ctx)->exec;
1073
1074 #ifndef NDEBUG
1075 /* debug check: make sure we don't get called recursively */
1076 exec->flush_call_depth++;
1077 assert(exec->flush_call_depth == 1);
1078 #endif
1079
1080 if (_mesa_inside_begin_end(ctx)) {
1081 /* We've had glBegin but not glEnd! */
1082 #ifndef NDEBUG
1083 exec->flush_call_depth--;
1084 assert(exec->flush_call_depth == 0);
1085 #endif
1086 return;
1087 }
1088
1089 /* Flush (draw). */
1090 vbo_exec_FlushVertices_internal(exec, flags);
1091
1092 #ifndef NDEBUG
1093 exec->flush_call_depth--;
1094 assert(exec->flush_call_depth == 0);
1095 #endif
1096 }
1097
1098
1099 void GLAPIENTRY
1100 _es_Color4f(GLfloat r, GLfloat g, GLfloat b, GLfloat a)
1101 {
1102 vbo_exec_Color4f(r, g, b, a);
1103 }
1104
1105
1106 void GLAPIENTRY
1107 _es_Normal3f(GLfloat x, GLfloat y, GLfloat z)
1108 {
1109 vbo_exec_Normal3f(x, y, z);
1110 }
1111
1112
1113 void GLAPIENTRY
1114 _es_MultiTexCoord4f(GLenum target, GLfloat s, GLfloat t, GLfloat r, GLfloat q)
1115 {
1116 vbo_exec_MultiTexCoord4f(target, s, t, r, q);
1117 }
1118
1119
1120 void GLAPIENTRY
1121 _es_Materialfv(GLenum face, GLenum pname, const GLfloat *params)
1122 {
1123 vbo_exec_Materialfv(face, pname, params);
1124 }
1125
1126
1127 void GLAPIENTRY
1128 _es_Materialf(GLenum face, GLenum pname, GLfloat param)
1129 {
1130 GLfloat p[4];
1131 p[0] = param;
1132 p[1] = p[2] = p[3] = 0.0F;
1133 vbo_exec_Materialfv(face, pname, p);
1134 }
1135
1136
1137 /**
1138 * A special version of glVertexAttrib4f that does not treat index 0 as
1139 * VBO_ATTRIB_POS.
1140 */
1141 static void
1142 VertexAttrib4f_nopos(GLuint index, GLfloat x, GLfloat y, GLfloat z, GLfloat w)
1143 {
1144 GET_CURRENT_CONTEXT(ctx);
1145 if (index < MAX_VERTEX_GENERIC_ATTRIBS)
1146 ATTRF(VBO_ATTRIB_GENERIC0 + index, 4, x, y, z, w);
1147 else
1148 ERROR(GL_INVALID_VALUE);
1149 }
1150
1151 void GLAPIENTRY
1152 _es_VertexAttrib4f(GLuint index, GLfloat x, GLfloat y, GLfloat z, GLfloat w)
1153 {
1154 VertexAttrib4f_nopos(index, x, y, z, w);
1155 }
1156
1157
1158 void GLAPIENTRY
1159 _es_VertexAttrib1f(GLuint indx, GLfloat x)
1160 {
1161 VertexAttrib4f_nopos(indx, x, 0.0f, 0.0f, 1.0f);
1162 }
1163
1164
1165 void GLAPIENTRY
1166 _es_VertexAttrib1fv(GLuint indx, const GLfloat* values)
1167 {
1168 VertexAttrib4f_nopos(indx, values[0], 0.0f, 0.0f, 1.0f);
1169 }
1170
1171
1172 void GLAPIENTRY
1173 _es_VertexAttrib2f(GLuint indx, GLfloat x, GLfloat y)
1174 {
1175 VertexAttrib4f_nopos(indx, x, y, 0.0f, 1.0f);
1176 }
1177
1178
1179 void GLAPIENTRY
1180 _es_VertexAttrib2fv(GLuint indx, const GLfloat* values)
1181 {
1182 VertexAttrib4f_nopos(indx, values[0], values[1], 0.0f, 1.0f);
1183 }
1184
1185
1186 void GLAPIENTRY
1187 _es_VertexAttrib3f(GLuint indx, GLfloat x, GLfloat y, GLfloat z)
1188 {
1189 VertexAttrib4f_nopos(indx, x, y, z, 1.0f);
1190 }
1191
1192
1193 void GLAPIENTRY
1194 _es_VertexAttrib3fv(GLuint indx, const GLfloat* values)
1195 {
1196 VertexAttrib4f_nopos(indx, values[0], values[1], values[2], 1.0f);
1197 }
1198
1199
1200 void GLAPIENTRY
1201 _es_VertexAttrib4fv(GLuint indx, const GLfloat* values)
1202 {
1203 VertexAttrib4f_nopos(indx, values[0], values[1], values[2], values[3]);
1204 }