1 /**************************************************************************
3 * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Keith Whitwell <keith@tungstengraphics.com>
33 #include "pipe/p_util.h"
34 #include "pipe/p_state.h"
35 #include "translate.h"
40 typedef void (*fetch_func
)(const void *ptr
, float *attrib
);
41 typedef void (*emit_func
)(const float *attrib
, void *ptr
);
45 struct translate_generic
{
46 struct translate translate
;
51 unsigned input_offset
;
54 unsigned output_offset
;
57 unsigned input_stride
;
59 } attrib
[PIPE_MAX_ATTRIBS
];
65 static struct translate_generic
*translate_generic( struct translate
*translate
)
67 return (struct translate_generic
*)translate
;
71 * Fetch a float[4] vertex attribute from memory, doing format/type
72 * conversion as needed.
74 * This is probably needed/dupliocated elsewhere, eg format
75 * conversion, texture sampling etc.
77 #define ATTRIB( NAME, SZ, TYPE, FROM, TO ) \
79 fetch_##NAME(const void *ptr, float *attrib) \
81 const float defaults[4] = { 0.0f,0.0f,0.0f,1.0f }; \
84 for (i = 0; i < SZ; i++) { \
85 attrib[i] = FROM(i); \
88 for (; i < 4; i++) { \
89 attrib[i] = defaults[i]; \
94 emit_##NAME(const float *attrib, void *ptr) \
97 TYPE *out = (TYPE *)ptr; \
99 for (i = 0; i < SZ; i++) { \
100 out[i] = TO(attrib[i]); \
105 #define FROM_64_FLOAT(i) ((float) ((double *) ptr)[i])
106 #define FROM_32_FLOAT(i) (((float *) ptr)[i])
108 #define FROM_8_USCALED(i) ((float) ((unsigned char *) ptr)[i])
109 #define FROM_16_USCALED(i) ((float) ((unsigned short *) ptr)[i])
110 #define FROM_32_USCALED(i) ((float) ((unsigned int *) ptr)[i])
112 #define FROM_8_SSCALED(i) ((float) ((char *) ptr)[i])
113 #define FROM_16_SSCALED(i) ((float) ((short *) ptr)[i])
114 #define FROM_32_SSCALED(i) ((float) ((int *) ptr)[i])
116 #define FROM_8_UNORM(i) ((float) ((unsigned char *) ptr)[i] / 255.0f)
117 #define FROM_16_UNORM(i) ((float) ((unsigned short *) ptr)[i] / 65535.0f)
118 #define FROM_32_UNORM(i) ((float) ((unsigned int *) ptr)[i] / 4294967295.0f)
120 #define FROM_8_SNORM(i) ((float) ((char *) ptr)[i] / 127.0f)
121 #define FROM_16_SNORM(i) ((float) ((short *) ptr)[i] / 32767.0f)
122 #define FROM_32_SNORM(i) ((float) ((int *) ptr)[i] / 2147483647.0f)
124 #define TO_64_FLOAT(x) ((double) x)
125 #define TO_32_FLOAT(x) (x)
127 #define TO_8_USCALED(x) ((unsigned char) x)
128 #define TO_16_USCALED(x) ((unsigned short) x)
129 #define TO_32_USCALED(x) ((unsigned int) x)
131 #define TO_8_SSCALED(x) ((char) x)
132 #define TO_16_SSCALED(x) ((short) x)
133 #define TO_32_SSCALED(x) ((int) x)
135 #define TO_8_UNORM(x) ((unsigned char) (x * 255.0f))
136 #define TO_16_UNORM(x) ((unsigned short) (x * 65535.0f))
137 #define TO_32_UNORM(x) ((unsigned int) (x * 4294967295.0f))
139 #define TO_8_SNORM(x) ((char) (x * 127.0f))
140 #define TO_16_SNORM(x) ((short) (x * 32767.0f))
141 #define TO_32_SNORM(x) ((int) (x * 2147483647.0f))
145 ATTRIB( R64G64B64A64_FLOAT
, 4, double, FROM_64_FLOAT
, TO_64_FLOAT
)
146 ATTRIB( R64G64B64_FLOAT
, 3, double, FROM_64_FLOAT
, TO_64_FLOAT
)
147 ATTRIB( R64G64_FLOAT
, 2, double, FROM_64_FLOAT
, TO_64_FLOAT
)
148 ATTRIB( R64_FLOAT
, 1, double, FROM_64_FLOAT
, TO_64_FLOAT
)
150 ATTRIB( R32G32B32A32_FLOAT
, 4, float, FROM_32_FLOAT
, TO_32_FLOAT
)
151 ATTRIB( R32G32B32_FLOAT
, 3, float, FROM_32_FLOAT
, TO_32_FLOAT
)
152 ATTRIB( R32G32_FLOAT
, 2, float, FROM_32_FLOAT
, TO_32_FLOAT
)
153 ATTRIB( R32_FLOAT
, 1, float, FROM_32_FLOAT
, TO_32_FLOAT
)
155 ATTRIB( R32G32B32A32_USCALED
, 4, unsigned, FROM_32_USCALED
, TO_32_USCALED
)
156 ATTRIB( R32G32B32_USCALED
, 3, unsigned, FROM_32_USCALED
, TO_32_USCALED
)
157 ATTRIB( R32G32_USCALED
, 2, unsigned, FROM_32_USCALED
, TO_32_USCALED
)
158 ATTRIB( R32_USCALED
, 1, unsigned, FROM_32_USCALED
, TO_32_USCALED
)
160 ATTRIB( R32G32B32A32_SSCALED
, 4, int, FROM_32_SSCALED
, TO_32_SSCALED
)
161 ATTRIB( R32G32B32_SSCALED
, 3, int, FROM_32_SSCALED
, TO_32_SSCALED
)
162 ATTRIB( R32G32_SSCALED
, 2, int, FROM_32_SSCALED
, TO_32_SSCALED
)
163 ATTRIB( R32_SSCALED
, 1, int, FROM_32_SSCALED
, TO_32_SSCALED
)
165 ATTRIB( R32G32B32A32_UNORM
, 4, unsigned, FROM_32_UNORM
, TO_32_UNORM
)
166 ATTRIB( R32G32B32_UNORM
, 3, unsigned, FROM_32_UNORM
, TO_32_UNORM
)
167 ATTRIB( R32G32_UNORM
, 2, unsigned, FROM_32_UNORM
, TO_32_UNORM
)
168 ATTRIB( R32_UNORM
, 1, unsigned, FROM_32_UNORM
, TO_32_UNORM
)
170 ATTRIB( R32G32B32A32_SNORM
, 4, int, FROM_32_SNORM
, TO_32_SNORM
)
171 ATTRIB( R32G32B32_SNORM
, 3, int, FROM_32_SNORM
, TO_32_SNORM
)
172 ATTRIB( R32G32_SNORM
, 2, int, FROM_32_SNORM
, TO_32_SNORM
)
173 ATTRIB( R32_SNORM
, 1, int, FROM_32_SNORM
, TO_32_SNORM
)
175 ATTRIB( R16G16B16A16_USCALED
, 4, ushort
, FROM_16_USCALED
, TO_16_USCALED
)
176 ATTRIB( R16G16B16_USCALED
, 3, ushort
, FROM_16_USCALED
, TO_16_USCALED
)
177 ATTRIB( R16G16_USCALED
, 2, ushort
, FROM_16_USCALED
, TO_16_USCALED
)
178 ATTRIB( R16_USCALED
, 1, ushort
, FROM_16_USCALED
, TO_16_USCALED
)
180 ATTRIB( R16G16B16A16_SSCALED
, 4, short, FROM_16_SSCALED
, TO_16_SSCALED
)
181 ATTRIB( R16G16B16_SSCALED
, 3, short, FROM_16_SSCALED
, TO_16_SSCALED
)
182 ATTRIB( R16G16_SSCALED
, 2, short, FROM_16_SSCALED
, TO_16_SSCALED
)
183 ATTRIB( R16_SSCALED
, 1, short, FROM_16_SSCALED
, TO_16_SSCALED
)
185 ATTRIB( R16G16B16A16_UNORM
, 4, ushort
, FROM_16_UNORM
, TO_16_UNORM
)
186 ATTRIB( R16G16B16_UNORM
, 3, ushort
, FROM_16_UNORM
, TO_16_UNORM
)
187 ATTRIB( R16G16_UNORM
, 2, ushort
, FROM_16_UNORM
, TO_16_UNORM
)
188 ATTRIB( R16_UNORM
, 1, ushort
, FROM_16_UNORM
, TO_16_UNORM
)
190 ATTRIB( R16G16B16A16_SNORM
, 4, short, FROM_16_SNORM
, TO_16_SNORM
)
191 ATTRIB( R16G16B16_SNORM
, 3, short, FROM_16_SNORM
, TO_16_SNORM
)
192 ATTRIB( R16G16_SNORM
, 2, short, FROM_16_SNORM
, TO_16_SNORM
)
193 ATTRIB( R16_SNORM
, 1, short, FROM_16_SNORM
, TO_16_SNORM
)
195 ATTRIB( R8G8B8A8_USCALED
, 4, ubyte
, FROM_8_USCALED
, TO_8_USCALED
)
196 ATTRIB( R8G8B8_USCALED
, 3, ubyte
, FROM_8_USCALED
, TO_8_USCALED
)
197 ATTRIB( R8G8_USCALED
, 2, ubyte
, FROM_8_USCALED
, TO_8_USCALED
)
198 ATTRIB( R8_USCALED
, 1, ubyte
, FROM_8_USCALED
, TO_8_USCALED
)
200 ATTRIB( R8G8B8A8_SSCALED
, 4, char, FROM_8_SSCALED
, TO_8_SSCALED
)
201 ATTRIB( R8G8B8_SSCALED
, 3, char, FROM_8_SSCALED
, TO_8_SSCALED
)
202 ATTRIB( R8G8_SSCALED
, 2, char, FROM_8_SSCALED
, TO_8_SSCALED
)
203 ATTRIB( R8_SSCALED
, 1, char, FROM_8_SSCALED
, TO_8_SSCALED
)
205 ATTRIB( R8G8B8A8_UNORM
, 4, ubyte
, FROM_8_UNORM
, TO_8_UNORM
)
206 ATTRIB( R8G8B8_UNORM
, 3, ubyte
, FROM_8_UNORM
, TO_8_UNORM
)
207 ATTRIB( R8G8_UNORM
, 2, ubyte
, FROM_8_UNORM
, TO_8_UNORM
)
208 ATTRIB( R8_UNORM
, 1, ubyte
, FROM_8_UNORM
, TO_8_UNORM
)
210 ATTRIB( R8G8B8A8_SNORM
, 4, char, FROM_8_SNORM
, TO_8_SNORM
)
211 ATTRIB( R8G8B8_SNORM
, 3, char, FROM_8_SNORM
, TO_8_SNORM
)
212 ATTRIB( R8G8_SNORM
, 2, char, FROM_8_SNORM
, TO_8_SNORM
)
213 ATTRIB( R8_SNORM
, 1, char, FROM_8_SNORM
, TO_8_SNORM
)
215 ATTRIB( A8R8G8B8_UNORM
, 4, ubyte
, FROM_8_UNORM
, TO_8_UNORM
)
216 //ATTRIB( R8G8B8A8_UNORM, 4, ubyte, FROM_8_UNORM, TO_8_UNORM )
221 fetch_B8G8R8A8_UNORM(const void *ptr
, float *attrib
)
223 attrib
[2] = FROM_8_UNORM(0);
224 attrib
[1] = FROM_8_UNORM(1);
225 attrib
[0] = FROM_8_UNORM(2);
226 attrib
[3] = FROM_8_UNORM(3);
230 emit_B8G8R8A8_UNORM( const float *attrib
, void *ptr
)
232 ubyte
*out
= (ubyte
*)ptr
;
233 out
[2] = TO_8_UNORM(attrib
[0]);
234 out
[1] = TO_8_UNORM(attrib
[1]);
235 out
[0] = TO_8_UNORM(attrib
[2]);
236 out
[3] = TO_8_UNORM(attrib
[3]);
240 fetch_NULL( const void *ptr
, float *attrib
)
249 emit_NULL( const float *attrib
, void *ptr
)
251 /* do nothing is the only sensible option */
254 static fetch_func
get_fetch_func( enum pipe_format format
)
257 case PIPE_FORMAT_R64_FLOAT
:
258 return fetch_R64_FLOAT
;
259 case PIPE_FORMAT_R64G64_FLOAT
:
260 return fetch_R64G64_FLOAT
;
261 case PIPE_FORMAT_R64G64B64_FLOAT
:
262 return fetch_R64G64B64_FLOAT
;
263 case PIPE_FORMAT_R64G64B64A64_FLOAT
:
264 return fetch_R64G64B64A64_FLOAT
;
266 case PIPE_FORMAT_R32_FLOAT
:
267 return fetch_R32_FLOAT
;
268 case PIPE_FORMAT_R32G32_FLOAT
:
269 return fetch_R32G32_FLOAT
;
270 case PIPE_FORMAT_R32G32B32_FLOAT
:
271 return fetch_R32G32B32_FLOAT
;
272 case PIPE_FORMAT_R32G32B32A32_FLOAT
:
273 return fetch_R32G32B32A32_FLOAT
;
275 case PIPE_FORMAT_R32_UNORM
:
276 return fetch_R32_UNORM
;
277 case PIPE_FORMAT_R32G32_UNORM
:
278 return fetch_R32G32_UNORM
;
279 case PIPE_FORMAT_R32G32B32_UNORM
:
280 return fetch_R32G32B32_UNORM
;
281 case PIPE_FORMAT_R32G32B32A32_UNORM
:
282 return fetch_R32G32B32A32_UNORM
;
284 case PIPE_FORMAT_R32_USCALED
:
285 return fetch_R32_USCALED
;
286 case PIPE_FORMAT_R32G32_USCALED
:
287 return fetch_R32G32_USCALED
;
288 case PIPE_FORMAT_R32G32B32_USCALED
:
289 return fetch_R32G32B32_USCALED
;
290 case PIPE_FORMAT_R32G32B32A32_USCALED
:
291 return fetch_R32G32B32A32_USCALED
;
293 case PIPE_FORMAT_R32_SNORM
:
294 return fetch_R32_SNORM
;
295 case PIPE_FORMAT_R32G32_SNORM
:
296 return fetch_R32G32_SNORM
;
297 case PIPE_FORMAT_R32G32B32_SNORM
:
298 return fetch_R32G32B32_SNORM
;
299 case PIPE_FORMAT_R32G32B32A32_SNORM
:
300 return fetch_R32G32B32A32_SNORM
;
302 case PIPE_FORMAT_R32_SSCALED
:
303 return fetch_R32_SSCALED
;
304 case PIPE_FORMAT_R32G32_SSCALED
:
305 return fetch_R32G32_SSCALED
;
306 case PIPE_FORMAT_R32G32B32_SSCALED
:
307 return fetch_R32G32B32_SSCALED
;
308 case PIPE_FORMAT_R32G32B32A32_SSCALED
:
309 return fetch_R32G32B32A32_SSCALED
;
311 case PIPE_FORMAT_R16_UNORM
:
312 return fetch_R16_UNORM
;
313 case PIPE_FORMAT_R16G16_UNORM
:
314 return fetch_R16G16_UNORM
;
315 case PIPE_FORMAT_R16G16B16_UNORM
:
316 return fetch_R16G16B16_UNORM
;
317 case PIPE_FORMAT_R16G16B16A16_UNORM
:
318 return fetch_R16G16B16A16_UNORM
;
320 case PIPE_FORMAT_R16_USCALED
:
321 return fetch_R16_USCALED
;
322 case PIPE_FORMAT_R16G16_USCALED
:
323 return fetch_R16G16_USCALED
;
324 case PIPE_FORMAT_R16G16B16_USCALED
:
325 return fetch_R16G16B16_USCALED
;
326 case PIPE_FORMAT_R16G16B16A16_USCALED
:
327 return fetch_R16G16B16A16_USCALED
;
329 case PIPE_FORMAT_R16_SNORM
:
330 return fetch_R16_SNORM
;
331 case PIPE_FORMAT_R16G16_SNORM
:
332 return fetch_R16G16_SNORM
;
333 case PIPE_FORMAT_R16G16B16_SNORM
:
334 return fetch_R16G16B16_SNORM
;
335 case PIPE_FORMAT_R16G16B16A16_SNORM
:
336 return fetch_R16G16B16A16_SNORM
;
338 case PIPE_FORMAT_R16_SSCALED
:
339 return fetch_R16_SSCALED
;
340 case PIPE_FORMAT_R16G16_SSCALED
:
341 return fetch_R16G16_SSCALED
;
342 case PIPE_FORMAT_R16G16B16_SSCALED
:
343 return fetch_R16G16B16_SSCALED
;
344 case PIPE_FORMAT_R16G16B16A16_SSCALED
:
345 return fetch_R16G16B16A16_SSCALED
;
347 case PIPE_FORMAT_R8_UNORM
:
348 return fetch_R8_UNORM
;
349 case PIPE_FORMAT_R8G8_UNORM
:
350 return fetch_R8G8_UNORM
;
351 case PIPE_FORMAT_R8G8B8_UNORM
:
352 return fetch_R8G8B8_UNORM
;
353 case PIPE_FORMAT_R8G8B8A8_UNORM
:
354 return fetch_R8G8B8A8_UNORM
;
356 case PIPE_FORMAT_R8_USCALED
:
357 return fetch_R8_USCALED
;
358 case PIPE_FORMAT_R8G8_USCALED
:
359 return fetch_R8G8_USCALED
;
360 case PIPE_FORMAT_R8G8B8_USCALED
:
361 return fetch_R8G8B8_USCALED
;
362 case PIPE_FORMAT_R8G8B8A8_USCALED
:
363 return fetch_R8G8B8A8_USCALED
;
365 case PIPE_FORMAT_R8_SNORM
:
366 return fetch_R8_SNORM
;
367 case PIPE_FORMAT_R8G8_SNORM
:
368 return fetch_R8G8_SNORM
;
369 case PIPE_FORMAT_R8G8B8_SNORM
:
370 return fetch_R8G8B8_SNORM
;
371 case PIPE_FORMAT_R8G8B8A8_SNORM
:
372 return fetch_R8G8B8A8_SNORM
;
374 case PIPE_FORMAT_R8_SSCALED
:
375 return fetch_R8_SSCALED
;
376 case PIPE_FORMAT_R8G8_SSCALED
:
377 return fetch_R8G8_SSCALED
;
378 case PIPE_FORMAT_R8G8B8_SSCALED
:
379 return fetch_R8G8B8_SSCALED
;
380 case PIPE_FORMAT_R8G8B8A8_SSCALED
:
381 return fetch_R8G8B8A8_SSCALED
;
383 case PIPE_FORMAT_A8R8G8B8_UNORM
:
384 return fetch_A8R8G8B8_UNORM
;
386 case PIPE_FORMAT_B8G8R8A8_UNORM
:
387 return fetch_B8G8R8A8_UNORM
;
398 static emit_func
get_emit_func( enum pipe_format format
)
401 case PIPE_FORMAT_R64_FLOAT
:
402 return emit_R64_FLOAT
;
403 case PIPE_FORMAT_R64G64_FLOAT
:
404 return emit_R64G64_FLOAT
;
405 case PIPE_FORMAT_R64G64B64_FLOAT
:
406 return emit_R64G64B64_FLOAT
;
407 case PIPE_FORMAT_R64G64B64A64_FLOAT
:
408 return emit_R64G64B64A64_FLOAT
;
410 case PIPE_FORMAT_R32_FLOAT
:
411 return emit_R32_FLOAT
;
412 case PIPE_FORMAT_R32G32_FLOAT
:
413 return emit_R32G32_FLOAT
;
414 case PIPE_FORMAT_R32G32B32_FLOAT
:
415 return emit_R32G32B32_FLOAT
;
416 case PIPE_FORMAT_R32G32B32A32_FLOAT
:
417 return emit_R32G32B32A32_FLOAT
;
419 case PIPE_FORMAT_R32_UNORM
:
420 return emit_R32_UNORM
;
421 case PIPE_FORMAT_R32G32_UNORM
:
422 return emit_R32G32_UNORM
;
423 case PIPE_FORMAT_R32G32B32_UNORM
:
424 return emit_R32G32B32_UNORM
;
425 case PIPE_FORMAT_R32G32B32A32_UNORM
:
426 return emit_R32G32B32A32_UNORM
;
428 case PIPE_FORMAT_R32_USCALED
:
429 return emit_R32_USCALED
;
430 case PIPE_FORMAT_R32G32_USCALED
:
431 return emit_R32G32_USCALED
;
432 case PIPE_FORMAT_R32G32B32_USCALED
:
433 return emit_R32G32B32_USCALED
;
434 case PIPE_FORMAT_R32G32B32A32_USCALED
:
435 return emit_R32G32B32A32_USCALED
;
437 case PIPE_FORMAT_R32_SNORM
:
438 return emit_R32_SNORM
;
439 case PIPE_FORMAT_R32G32_SNORM
:
440 return emit_R32G32_SNORM
;
441 case PIPE_FORMAT_R32G32B32_SNORM
:
442 return emit_R32G32B32_SNORM
;
443 case PIPE_FORMAT_R32G32B32A32_SNORM
:
444 return emit_R32G32B32A32_SNORM
;
446 case PIPE_FORMAT_R32_SSCALED
:
447 return emit_R32_SSCALED
;
448 case PIPE_FORMAT_R32G32_SSCALED
:
449 return emit_R32G32_SSCALED
;
450 case PIPE_FORMAT_R32G32B32_SSCALED
:
451 return emit_R32G32B32_SSCALED
;
452 case PIPE_FORMAT_R32G32B32A32_SSCALED
:
453 return emit_R32G32B32A32_SSCALED
;
455 case PIPE_FORMAT_R16_UNORM
:
456 return emit_R16_UNORM
;
457 case PIPE_FORMAT_R16G16_UNORM
:
458 return emit_R16G16_UNORM
;
459 case PIPE_FORMAT_R16G16B16_UNORM
:
460 return emit_R16G16B16_UNORM
;
461 case PIPE_FORMAT_R16G16B16A16_UNORM
:
462 return emit_R16G16B16A16_UNORM
;
464 case PIPE_FORMAT_R16_USCALED
:
465 return emit_R16_USCALED
;
466 case PIPE_FORMAT_R16G16_USCALED
:
467 return emit_R16G16_USCALED
;
468 case PIPE_FORMAT_R16G16B16_USCALED
:
469 return emit_R16G16B16_USCALED
;
470 case PIPE_FORMAT_R16G16B16A16_USCALED
:
471 return emit_R16G16B16A16_USCALED
;
473 case PIPE_FORMAT_R16_SNORM
:
474 return emit_R16_SNORM
;
475 case PIPE_FORMAT_R16G16_SNORM
:
476 return emit_R16G16_SNORM
;
477 case PIPE_FORMAT_R16G16B16_SNORM
:
478 return emit_R16G16B16_SNORM
;
479 case PIPE_FORMAT_R16G16B16A16_SNORM
:
480 return emit_R16G16B16A16_SNORM
;
482 case PIPE_FORMAT_R16_SSCALED
:
483 return emit_R16_SSCALED
;
484 case PIPE_FORMAT_R16G16_SSCALED
:
485 return emit_R16G16_SSCALED
;
486 case PIPE_FORMAT_R16G16B16_SSCALED
:
487 return emit_R16G16B16_SSCALED
;
488 case PIPE_FORMAT_R16G16B16A16_SSCALED
:
489 return emit_R16G16B16A16_SSCALED
;
491 case PIPE_FORMAT_R8_UNORM
:
492 return emit_R8_UNORM
;
493 case PIPE_FORMAT_R8G8_UNORM
:
494 return emit_R8G8_UNORM
;
495 case PIPE_FORMAT_R8G8B8_UNORM
:
496 return emit_R8G8B8_UNORM
;
497 case PIPE_FORMAT_R8G8B8A8_UNORM
:
498 return emit_R8G8B8A8_UNORM
;
500 case PIPE_FORMAT_R8_USCALED
:
501 return emit_R8_USCALED
;
502 case PIPE_FORMAT_R8G8_USCALED
:
503 return emit_R8G8_USCALED
;
504 case PIPE_FORMAT_R8G8B8_USCALED
:
505 return emit_R8G8B8_USCALED
;
506 case PIPE_FORMAT_R8G8B8A8_USCALED
:
507 return emit_R8G8B8A8_USCALED
;
509 case PIPE_FORMAT_R8_SNORM
:
510 return emit_R8_SNORM
;
511 case PIPE_FORMAT_R8G8_SNORM
:
512 return emit_R8G8_SNORM
;
513 case PIPE_FORMAT_R8G8B8_SNORM
:
514 return emit_R8G8B8_SNORM
;
515 case PIPE_FORMAT_R8G8B8A8_SNORM
:
516 return emit_R8G8B8A8_SNORM
;
518 case PIPE_FORMAT_R8_SSCALED
:
519 return emit_R8_SSCALED
;
520 case PIPE_FORMAT_R8G8_SSCALED
:
521 return emit_R8G8_SSCALED
;
522 case PIPE_FORMAT_R8G8B8_SSCALED
:
523 return emit_R8G8B8_SSCALED
;
524 case PIPE_FORMAT_R8G8B8A8_SSCALED
:
525 return emit_R8G8B8A8_SSCALED
;
527 case PIPE_FORMAT_A8R8G8B8_UNORM
:
528 return emit_A8R8G8B8_UNORM
;
530 case PIPE_FORMAT_B8G8R8A8_UNORM
:
531 return emit_B8G8R8A8_UNORM
;
542 * Fetch vertex attributes for 'count' vertices.
544 static void generic_run_elts( struct translate
*translate
,
545 const unsigned *elts
,
547 void *output_buffer
)
549 struct translate_generic
*tg
= translate_generic(translate
);
550 char *vert
= output_buffer
;
551 unsigned nr_attrs
= tg
->nr_attrib
;
555 /* loop over vertex attributes (vertex shader inputs)
557 for (i
= 0; i
< count
; i
++) {
558 unsigned elt
= *elts
++;
560 for (attr
= 0; attr
< nr_attrs
; attr
++) {
563 const char *src
= (tg
->attrib
[attr
].input_ptr
+
564 tg
->attrib
[attr
].input_stride
* elt
);
567 tg
->attrib
[attr
].output_offset
);
569 tg
->attrib
[attr
].fetch( src
, data
);
571 if (0) debug_printf("vert %d/%d attr %d: %f %f %f %f\n",
572 i
, elt
, attr
, data
[0], data
[1], data
[2], data
[3]);
574 tg
->attrib
[attr
].emit( data
, dst
);
577 vert
+= tg
->translate
.key
.output_stride
;
583 static void generic_run( struct translate
*translate
,
586 void *output_buffer
)
588 struct translate_generic
*tg
= translate_generic(translate
);
589 char *vert
= output_buffer
;
590 unsigned nr_attrs
= tg
->nr_attrib
;
594 /* loop over vertex attributes (vertex shader inputs)
596 for (i
= 0; i
< count
; i
++) {
597 unsigned elt
= start
+ i
;
599 for (attr
= 0; attr
< nr_attrs
; attr
++) {
602 const char *src
= (tg
->attrib
[attr
].input_ptr
+
603 tg
->attrib
[attr
].input_stride
* elt
);
606 tg
->attrib
[attr
].output_offset
);
608 tg
->attrib
[attr
].fetch( src
, data
);
610 if (0) debug_printf("vert %d attr %d: %f %f %f %f\n",
611 i
, attr
, data
[0], data
[1], data
[2], data
[3]);
613 tg
->attrib
[attr
].emit( data
, dst
);
616 vert
+= tg
->translate
.key
.output_stride
;
622 static void generic_set_buffer( struct translate
*translate
,
627 struct translate_generic
*tg
= translate_generic(translate
);
630 for (i
= 0; i
< tg
->nr_attrib
; i
++) {
631 if (tg
->attrib
[i
].buffer
== buf
) {
632 tg
->attrib
[i
].input_ptr
= ((char *)ptr
+
633 tg
->attrib
[i
].input_offset
);
634 tg
->attrib
[i
].input_stride
= stride
;
640 static void generic_release( struct translate
*translate
)
647 struct translate
*translate_generic_create( const struct translate_key
*key
)
649 struct translate_generic
*tg
= CALLOC_STRUCT(translate_generic
);
655 tg
->translate
.key
= *key
;
656 tg
->translate
.release
= generic_release
;
657 tg
->translate
.set_buffer
= generic_set_buffer
;
658 tg
->translate
.run_elts
= generic_run_elts
;
659 tg
->translate
.run
= generic_run
;
661 for (i
= 0; i
< key
->nr_elements
; i
++) {
663 tg
->attrib
[i
].fetch
= get_fetch_func(key
->element
[i
].input_format
);
664 tg
->attrib
[i
].buffer
= key
->element
[i
].input_buffer
;
665 tg
->attrib
[i
].input_offset
= key
->element
[i
].input_offset
;
667 tg
->attrib
[i
].emit
= get_emit_func(key
->element
[i
].output_format
);
668 tg
->attrib
[i
].output_offset
= key
->element
[i
].output_offset
;
672 tg
->nr_attrib
= key
->nr_elements
;
675 return &tg
->translate
;