/**************************************************************************
*
- * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
+ * Copyright 2007 VMware, Inc.
* All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
- * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
+ * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
#include "pipe/p_defines.h"
#include "pipe/p_context.h"
+#include "util/u_draw.h"
#include "util/u_prim.h"
#include "lp_context.h"
if (!llvmpipe_check_render_cond(lp))
return;
+ if (info->indirect) {
+ util_draw_indirect(pipe, info);
+ return;
+ }
+
if (lp->dirty)
llvmpipe_update_derived( lp );
* Map vertex buffers
*/
for (i = 0; i < lp->num_vertex_buffers; i++) {
- const void *buf = lp->vertex_buffer[i].user_buffer;
+ const void *buf = lp->vertex_buffer[i].is_user_buffer ?
+ lp->vertex_buffer[i].buffer.user : NULL;
size_t size = ~0;
if (!buf) {
- if (!lp->vertex_buffer[i].buffer) {
+ if (!lp->vertex_buffer[i].buffer.resource) {
continue;
}
- buf = llvmpipe_resource_data(lp->vertex_buffer[i].buffer);
- size = lp->vertex_buffer[i].buffer->width0;
+ buf = llvmpipe_resource_data(lp->vertex_buffer[i].buffer.resource);
+ size = lp->vertex_buffer[i].buffer.resource->width0;
}
draw_set_mapped_vertex_buffer(draw, i, buf, size);
}
/* Map index buffer, if present */
- if (info->indexed) {
- mapped_indices = lp->index_buffer.user_buffer;
- if (!mapped_indices)
- mapped_indices = llvmpipe_resource_data(lp->index_buffer.buffer);
-
- draw_set_indexes(draw,
- (ubyte *) mapped_indices + lp->index_buffer.offset,
- lp->index_buffer.index_size);
- }
-
- for (i = 0; i < lp->num_so_targets; i++) {
- void *buf = 0;
- if (lp->so_targets[i]) {
- buf = llvmpipe_resource(lp->so_targets[i]->target.buffer)->data;
- lp->so_targets[i]->mapping = buf;
+ if (info->index_size) {
+ unsigned available_space = ~0;
+ mapped_indices = info->has_user_indices ? info->index.user : NULL;
+ if (!mapped_indices) {
+ mapped_indices = llvmpipe_resource_data(info->index.resource);
+ available_space = info->index.resource->width0;
}
+ draw_set_indexes(draw,
+ (ubyte *) mapped_indices,
+ info->index_size, available_space);
}
- draw_set_mapped_so_targets(draw, lp->num_so_targets,
- lp->so_targets);
llvmpipe_prepare_vertex_sampling(lp,
lp->num_sampler_views[PIPE_SHADER_VERTEX],
llvmpipe_prepare_geometry_sampling(lp,
lp->num_sampler_views[PIPE_SHADER_GEOMETRY],
lp->sampler_views[PIPE_SHADER_GEOMETRY]);
- if (lp->gs && !lp->gs->shader.tokens) {
+
+ llvmpipe_prepare_vertex_images(lp,
+ lp->num_images[PIPE_SHADER_VERTEX],
+ lp->images[PIPE_SHADER_VERTEX]);
+ llvmpipe_prepare_geometry_images(lp,
+ lp->num_images[PIPE_SHADER_GEOMETRY],
+ lp->images[PIPE_SHADER_GEOMETRY]);
+ if (lp->gs && lp->gs->no_tokens) {
/* we have an empty geometry shader with stream output, so
attach the stream output info to the current vertex shader */
if (lp->vs) {
- draw_vs_attach_so(lp->vs->draw_data, &lp->gs->shader.stream_output);
+ draw_vs_attach_so(lp->vs, &lp->gs->stream_output);
}
}
draw_collect_pipeline_statistics(draw,
draw_set_mapped_vertex_buffer(draw, i, NULL, 0);
}
if (mapped_indices) {
- draw_set_indexes(draw, NULL, 0);
+ draw_set_indexes(draw, NULL, 0, 0);
}
- draw_set_mapped_so_targets(draw, 0, NULL);
- if (lp->gs && !lp->gs->shader.tokens) {
+ if (lp->gs && lp->gs->no_tokens) {
/* we have attached stream output to the vs for rendering,
now lets reset it */
if (lp->vs) {
- draw_vs_reset_so(lp->vs->draw_data);
+ draw_vs_reset_so(lp->vs);
}
}
-
- llvmpipe_cleanup_vertex_sampling(lp);
- llvmpipe_cleanup_geometry_sampling(lp);
/*
* TODO: Flush only when a user vertex/index buffer is present