struct translate_key transkey;
unsigned i;
- assert(num_elements);
-
so = MALLOC(sizeof(*so) +
- (num_elements - 1) * sizeof(struct nv50_vertex_element));
+ num_elements * sizeof(struct nv50_vertex_element));
if (!so)
return NULL;
so->num_elements = num_elements;
{
const void *data;
struct nouveau_channel *chan = nv50->screen->base.channel;
- struct nv50_resource *res = nv50_resource(vb->buffer);
+ struct nv04_resource *res = nv04_resource(vb->buffer);
float v[4];
const unsigned nc = util_format_get_nr_components(ve->src_format);
- data = nv50_resource_map_offset(nv50, res, vb->buffer_offset +
- ve->src_offset, NOUVEAU_BO_RD);
+ data = nouveau_resource_map_offset(&nv50->base, res, vb->buffer_offset +
+ ve->src_offset, NOUVEAU_BO_RD);
util_format_read_4f(ve->src_format, v, 0, data, 0, 0, 0, 1, 1);
nv50_prevalidate_vbufs(struct nv50_context *nv50)
{
struct pipe_vertex_buffer *vb;
- struct nv50_resource *buf;
+ struct nv04_resource *buf;
int i;
uint32_t base, size;
vb = &nv50->vtxbuf[i];
if (!vb->stride)
continue;
- buf = nv50_resource(vb->buffer);
+ buf = nv04_resource(vb->buffer);
/* NOTE: user buffers with temporary storage count as mapped by GPU */
- if (!nv50_resource_mapped_by_gpu(vb->buffer)) {
+ if (!nouveau_resource_mapped_by_gpu(vb->buffer)) {
if (nv50->vbo_push_hint) {
nv50->vbo_fifo = ~0;
continue;
} else {
- if (buf->status & NV50_BUFFER_STATUS_USER_MEMORY) {
+ if (buf->status & NOUVEAU_BUFFER_STATUS_USER_MEMORY) {
nv50->vbo_user |= 1 << i;
assert(vb->stride > vb->buffer_offset);
nv50_vbuf_range(nv50, i, &base, &size);
- nv50_user_buffer_upload(buf, base, size);
+ nouveau_user_buffer_upload(buf, base, size);
} else {
- nv50_buffer_migrate(nv50, buf, NOUVEAU_BO_GART);
+ nouveau_buffer_migrate(&nv50->base, buf, NOUVEAU_BO_GART);
}
- nv50->vbo_dirty = TRUE;
+ nv50->base.vbo_dirty = TRUE;
}
}
nv50_bufctx_add_resident(nv50, NV50_BUFCTX_VERTEX, buf, NOUVEAU_BO_RD);
- nv50_buffer_adjust_score(nv50, buf, 1);
+ nouveau_buffer_adjust_score(&nv50->base, buf, 1);
}
}
struct pipe_vertex_element *ve = &nv50->vertex->element[i].pipe;
const int b = ve->vertex_buffer_index;
struct pipe_vertex_buffer *vb = &nv50->vtxbuf[b];
- struct nv50_resource *buf = nv50_resource(vb->buffer);
+ struct nv04_resource *buf = nv04_resource(vb->buffer);
if (!(nv50->vbo_user & (1 << b)))
continue;
if (!(written & (1 << b))) {
written |= 1 << b;
- nv50_user_buffer_upload(buf, base, size);
+ nouveau_user_buffer_upload(buf, base, size);
}
offset = vb->buffer_offset + ve->src_offset;
OUT_RESRCh(chan, buf, offset, NOUVEAU_BO_RD);
OUT_RESRCl(chan, buf, offset, NOUVEAU_BO_RD);
}
- nv50->vbo_dirty = TRUE;
+ nv50->base.vbo_dirty = TRUE;
}
static INLINE void
int i = ffs(vbo_user) - 1;
vbo_user &= ~(1 << i);
- nv50_buffer_release_gpu_storage(nv50_resource(nv50->vtxbuf[i].buffer));
+ nouveau_buffer_release_gpu_storage(nv04_resource(nv50->vtxbuf[i].buffer));
}
}
}
for (i = 0; i < vertex->num_elements; ++i) {
- struct nv50_resource *res;
+ struct nv04_resource *res;
unsigned size, offset;
ve = &vertex->element[i];
OUT_RING (chan, 0);
}
- res = nv50_resource(vb->buffer);
+ res = nv04_resource(vb->buffer);
if (nv50->vbo_fifo || unlikely(vb->stride == 0)) {
if (!nv50->vbo_fifo)
{
struct nv50_context *nv50 = chan->user_private;
- nv50_screen_fence_update(nv50->screen, TRUE);
+ nouveau_fence_update(&nv50->screen->base, TRUE);
nv50_bufctx_emit_relocs(nv50);
}
nv50->state.index_bias = index_bias;
}
- if (nv50_resource_mapped_by_gpu(nv50->idxbuf.buffer) && 0) {
- struct nv50_resource *res = nv50_resource(nv50->idxbuf.buffer);
- unsigned offset = res->offset + nv50->idxbuf.offset;
+ if (nouveau_resource_mapped_by_gpu(nv50->idxbuf.buffer)) {
+ struct nv04_resource *res = nv04_resource(nv50->idxbuf.buffer);
+
+ start += nv50->idxbuf.offset >> (index_size >> 1);
- nv50_buffer_adjust_score(nv50, res, 1);
+ nouveau_buffer_adjust_score(&nv50->base, res, 1);
while (instance_count--) {
BEGIN_RING(chan, RING_3D(VERTEX_BEGIN_GL), 1);
WAIT_RING (chan, 2);
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U32) | 0x30000, 0);
OUT_RING (chan, count);
- nouveau_pushbuf_submit(chan, res->bo,
- (start << 2) + offset,
- (count << 2));
+ nouveau_pushbuf_submit(chan, res->bo, res->offset + start * 4,
+ count * 4);
}
break;
case 2:
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U16_SETUP), 1);
OUT_RING (chan, (start << 31) | count);
+ WAIT_RING (chan, 2);
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U16) | 0x30000, 0);
OUT_RING (chan, pb_words);
- nouveau_pushbuf_submit(chan, res->bo,
- (pb_start << 1) + offset, pb_words << 2);
+ nouveau_pushbuf_submit(chan, res->bo, res->offset + pb_start * 2,
+ pb_words * 4);
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U16_SETUP), 1);
OUT_RING (chan, 0);
break;
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U8_SETUP), 1);
OUT_RING (chan, (start << 30) | count);
+ WAIT_RING (chan, 2);
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U8) | 0x30000, 0);
OUT_RING (chan, pb_words);
- nouveau_pushbuf_submit(chan, res->bo,
- pb_start + offset, pb_words << 2);
+ nouveau_pushbuf_submit(chan, res->bo, res->offset + pb_start,
+ pb_words * 4);
BEGIN_RING(chan, RING_3D(VB_ELEMENT_U8_SETUP), 1);
OUT_RING (chan, 0);
break;
assert(0);
return;
}
+ BEGIN_RING(chan, RING_3D(VERTEX_END_GL), 1);
+ OUT_RING (chan, 0);
nv50_resource_fence(res, NOUVEAU_BO_RD);
mode |= NV50_3D_VERTEX_BEGIN_GL_INSTANCE_NEXT;
}
} else {
- data = nv50_resource_map_offset(nv50, nv50_resource(nv50->idxbuf.buffer),
- nv50->idxbuf.offset, NOUVEAU_BO_RD);
+ data = nouveau_resource_map_offset(&nv50->base,
+ nv04_resource(nv50->idxbuf.buffer),
+ nv50->idxbuf.offset, NOUVEAU_BO_RD);
if (!data)
return;
OUT_RING (chan, info->start_instance);
}
- if (nv50->vbo_dirty) {
+ if (nv50->base.vbo_dirty) {
BEGIN_RING(chan, RING_3D(VERTEX_ARRAY_FLUSH), 1);
OUT_RING (chan, 0);
- nv50->vbo_dirty = FALSE;
+ nv50->base.vbo_dirty = FALSE;
}
if (!info->indexed) {