+#include "nouveau_screen.h"
+#include "nouveau_context.h"
+#include "nouveau_video.h"
+
+#include "nouveau/nouveau_buffer.h"
+#include "util/u_video.h"
+#include "util/u_format.h"
+#include "util/u_sampler.h"
+
+static int
+nouveau_vpe_init(struct nouveau_decoder *dec) {
+ int ret;
+ if (dec->cmds)
+ return 0;
+ ret = nouveau_bo_map(dec->cmd_bo, NOUVEAU_BO_RDWR, dec->client);
+ if (ret) {
+ debug_printf("Mapping cmd bo: %s\n", strerror(-ret));
+ return ret;
+ }
+ ret = nouveau_bo_map(dec->data_bo, NOUVEAU_BO_RDWR, dec->client);
+ if (ret) {
+ debug_printf("Mapping data bo: %s\n", strerror(-ret));
+ return ret;
+ }
+ dec->cmds = dec->cmd_bo->map;
+ dec->data = dec->data_bo->map;
+ return ret;
+}
+
+static void
+nouveau_vpe_synch(struct nouveau_decoder *dec) {
+ struct nouveau_pushbuf *push = dec->push;
+#if 0
+ if (dec->fence_map) {
+ BEGIN_NV04(push, NV84_MPEG(QUERY_COUNTER), 1);
+ PUSH_DATA (push, ++dec->fence_seq);
+ PUSH_KICK (push);
+ while (dec->fence_map[0] != dec->fence_seq)
+ usleep(1000);
+ } else
+#endif
+ PUSH_KICK(push);
+}
+
+static void
+nouveau_vpe_fini(struct nouveau_decoder *dec) {
+ struct nouveau_pushbuf *push = dec->push;
+ if (!dec->cmds)
+ return;
+
+ nouveau_pushbuf_space(push, 8, 2, 0);
+ nouveau_bufctx_reset(dec->bufctx, NV31_VIDEO_BIND_CMD);
+
+#define BCTX_ARGS dec->bufctx, NV31_VIDEO_BIND_CMD, NOUVEAU_BO_RD
+
+ BEGIN_NV04(push, NV31_MPEG(CMD_OFFSET), 2);
+ PUSH_MTHDl(push, NV31_MPEG(CMD_OFFSET), dec->cmd_bo, 0, BCTX_ARGS);
+ PUSH_DATA (push, dec->ofs * 4);
+
+ BEGIN_NV04(push, NV31_MPEG(DATA_OFFSET), 2);
+ PUSH_MTHDl(push, NV31_MPEG(DATA_OFFSET), dec->data_bo, 0, BCTX_ARGS);
+ PUSH_DATA (push, dec->data_pos * 4);
+
+#undef BCTX_ARGS
+
+ if (unlikely(nouveau_pushbuf_validate(dec->push)))
+ return;
+
+ BEGIN_NV04(push, NV31_MPEG(EXEC), 1);
+ PUSH_DATA (push, 1);
+
+ nouveau_vpe_synch(dec);
+ dec->ofs = dec->data_pos = dec->num_surfaces = 0;
+ dec->cmds = dec->data = NULL;
+ dec->current = dec->future = dec->past = 8;
+}
+
+static INLINE void
+nouveau_vpe_mb_dct_blocks(struct nouveau_decoder *dec, const struct pipe_mpeg12_macroblock *mb)
+{
+ int cbb;
+ unsigned cbp = mb->coded_block_pattern;
+ short *db = mb->blocks;
+ for (cbb = 0x20; cbb > 0; cbb >>= 1) {
+ if (cbb & cbp) {
+ static const int lookup[64] = {
+ 0, 1, 8,16, 9, 2, 3,10,
+ 17,24,32,25,18,11, 4, 5,
+ 12,19,26,33,40,48,41,34,
+ 27,20,13, 6, 7,14,21,28,
+ 35,42,49,56,57,50,43,36,
+ 29,22,15,23,30,37,44,51,
+ 58,59,52,45,38,31,39,46,
+ 53,60,61,54,47,55,62,63
+ };
+ int i, j = 0, found = 0;
+ for (i = 0; i < 64; ++i) {
+ if (!db[lookup[i]]) { j += 2; continue; }
+ dec->data[dec->data_pos++] = (db[lookup[i]] << 16) | j;
+ j = 0;
+ found = 1;
+ }
+ if (found)
+ dec->data[dec->data_pos - 1] |= 1;
+ else
+ dec->data[dec->data_pos++] = 1;
+ db += 64;
+ } else if (mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA) {
+ dec->data[dec->data_pos++] = 1;
+ }
+ }
+}
+
+static INLINE void
+nouveau_vpe_mb_data_blocks(struct nouveau_decoder *dec, const struct pipe_mpeg12_macroblock *mb)
+{
+ int cbb;
+ unsigned cbp = mb->coded_block_pattern;
+ short *db = mb->blocks;
+ for (cbb = 0x20; cbb > 0; cbb >>= 1) {
+ if (cbb & cbp) {
+ memcpy(&dec->data[dec->data_pos], db, 128);
+ dec->data_pos += 32;
+ db += 64;
+ } else if (mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA) {
+ memset(&dec->data[dec->data_pos], 0, 128);
+ dec->data_pos += 32;
+ }
+ }
+}
+
+static INLINE void
+nouveau_vpe_mb_dct_header(struct nouveau_decoder *dec,
+ const struct pipe_mpeg12_macroblock *mb,
+ bool luma)
+{
+ unsigned base_dct, cbp;
+ bool intra = mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA;
+ unsigned x = mb->x * 16;
+ unsigned y = luma ? mb->y * 16 : mb->y * 8;
+
+ /* Setup the base dct header */
+ base_dct = dec->current << NV17_MPEG_CMD_CHROMA_MB_HEADER_SURFACE__SHIFT;
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_RUN_SINGLE;
+
+ if (!(mb->x & 1))
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_X_COORD_EVEN;
+ if (intra)
+ cbp = 0x3f;
+ else
+ cbp = mb->coded_block_pattern;
+
+ if (dec->picture_structure == PIPE_MPEG12_PICTURE_STRUCTURE_FRAME) {
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_TYPE_FRAME;
+ if (luma && mb->macroblock_modes.bits.dct_type == PIPE_MPEG12_DCT_TYPE_FIELD)
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_FRAME_DCT_TYPE_FIELD;
+ } else {
+ if (dec->picture_structure == PIPE_MPEG12_PICTURE_STRUCTURE_FIELD_BOTTOM)
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_FIELD_BOTTOM;
+ if (!intra)
+ y *= 2;
+ }
+
+ if (luma) {
+ base_dct |= NV17_MPEG_CMD_LUMA_MB_HEADER_OP_LUMA_MB_HEADER;
+ base_dct |= (cbp >> 2) << NV17_MPEG_CMD_LUMA_MB_HEADER_CBP__SHIFT;
+ } else {
+ base_dct |= NV17_MPEG_CMD_CHROMA_MB_HEADER_OP_CHROMA_MB_HEADER;
+ base_dct |= (cbp & 3) << NV17_MPEG_CMD_CHROMA_MB_HEADER_CBP__SHIFT;
+ }
+ nouveau_vpe_write(dec, base_dct);
+ nouveau_vpe_write(dec, NV17_MPEG_CMD_MB_COORDS_OP_MB_COORDS |
+ x | (y << NV17_MPEG_CMD_MB_COORDS_Y__SHIFT));
+}
+
+static INLINE unsigned int
+nouveau_vpe_mb_mv_flags(bool luma, int mv_h, int mv_v, bool forward, bool first, bool vert)
+{
+ unsigned mc_header = 0;
+ if (luma)
+ mc_header |= NV17_MPEG_CMD_LUMA_MV_HEADER_OP_LUMA_MV_HEADER;
+ else
+ mc_header |= NV17_MPEG_CMD_CHROMA_MV_HEADER_OP_CHROMA_MV_HEADER;
+ if (mv_h & 1)
+ mc_header |= NV17_MPEG_CMD_CHROMA_MV_HEADER_X_HALF;
+ if (mv_v & 1)
+ mc_header |= NV17_MPEG_CMD_CHROMA_MV_HEADER_Y_HALF;
+ if (!forward)
+ mc_header |= NV17_MPEG_CMD_CHROMA_MV_HEADER_DIRECTION_BACKWARD;
+ if (!first)
+ mc_header |= NV17_MPEG_CMD_CHROMA_MV_HEADER_IDX;
+ if (vert)
+ mc_header |= NV17_MPEG_CMD_LUMA_MV_HEADER_FIELD_BOTTOM;
+ return mc_header;
+}
+
+static unsigned pos(int pos, int mov, int max) {
+ int ret = pos + mov;
+ if (pos < 0)
+ return 0;
+ if (pos >= max)
+ return max-1;
+ return ret;
+}
+
+/* because we want -1 / 2 = -1 */
+static int div_down(int val, int mult) {
+ val &= ~(mult - 1);
+ return val / mult;
+}
+
+static int div_up(int val, int mult) {
+ val += mult - 1;
+ return val / mult;
+}
+
+static INLINE void
+nouveau_vpe_mb_mv(struct nouveau_decoder *dec, unsigned mc_header,
+ bool luma, bool frame, bool forward, bool vert,
+ int x, int y, const short motions[2],
+ unsigned surface, bool first)
+{
+ unsigned mc_vector;
+ int mv_horizontal = motions[0];
+ int mv_vertical = motions[1];
+ int mv2 = mc_header & NV17_MPEG_CMD_CHROMA_MV_HEADER_COUNT_2;
+ unsigned width = dec->base.width;
+ unsigned height = dec->base.height;
+ if (mv2)
+ mv_vertical = div_down(mv_vertical, 2);
+ assert(frame); // Untested for non-frames
+ if (!frame)
+ height *= 2;
+
+ mc_header |= surface << NV17_MPEG_CMD_CHROMA_MV_HEADER_SURFACE__SHIFT;
+ if (!luma) {
+ mv_vertical = div_up(mv_vertical, 2);
+ mv_horizontal = div_up(mv_horizontal, 2);
+ height /= 2;
+ }
+ mc_header |= nouveau_vpe_mb_mv_flags(luma, mv_horizontal, mv_vertical, forward, first, vert);
+ nouveau_vpe_write(dec, mc_header);
+
+ mc_vector = NV17_MPEG_CMD_MV_COORDS_OP_MV_COORDS;
+ if (luma)
+ mc_vector |= pos(x, div_down(mv_horizontal, 2), width);
+ else
+ mc_vector |= pos(x, mv_horizontal & ~1, width);
+ if (!mv2)
+ mc_vector |= pos(y, div_down(mv_vertical, 2), height) << NV17_MPEG_CMD_MV_COORDS_Y__SHIFT;
+ else
+ mc_vector |= pos(y, mv_vertical & ~1, height) << NV17_MPEG_CMD_MV_COORDS_Y__SHIFT;
+ nouveau_vpe_write(dec, mc_vector);
+}
+
+static void
+nouveau_vpe_mb_mv_header(struct nouveau_decoder *dec,
+ const struct pipe_mpeg12_macroblock *mb,
+ bool luma)
+{
+ bool frame = dec->picture_structure == PIPE_MPEG12_PICTURE_STRUCTURE_FRAME;
+ unsigned base;
+ bool forward, backward;
+ int y, y2, x = mb->x * 16;
+ if (luma)
+ y = mb->y * (frame ? 16 : 32);
+ else
+ y = mb->y * (frame ? 8 : 16);
+ if (frame)
+ y2 = y;
+ else
+ y2 = y + (luma ? 16 : 8);
+
+ forward = mb->macroblock_type & PIPE_MPEG12_MB_TYPE_MOTION_FORWARD;
+ backward = mb->macroblock_type & PIPE_MPEG12_MB_TYPE_MOTION_BACKWARD;
+ assert(!forward || dec->past < 8);
+ assert(!backward || dec->future < 8);
+ if (frame) {
+ switch (mb->macroblock_modes.bits.frame_motion_type) {
+ case PIPE_MPEG12_MO_TYPE_FRAME: goto mv1;
+ case PIPE_MPEG12_MO_TYPE_FIELD: goto mv2;
+ case PIPE_MPEG12_MO_TYPE_DUAL_PRIME: {
+ base = NV17_MPEG_CMD_CHROMA_MV_HEADER_COUNT_2;
+ if (forward) {
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE, FALSE,
+ x, y, mb->PMV[0][0], dec->past, TRUE);
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE, TRUE,
+ x, y2, mb->PMV[0][0], dec->past, FALSE);
+ }
+ if (backward && forward) {
+ nouveau_vpe_mb_mv(dec, base, luma, frame, !forward, TRUE,
+ x, y, mb->PMV[1][0], dec->future, TRUE);
+ nouveau_vpe_mb_mv(dec, base, luma, frame, !forward, FALSE,
+ x, y2, mb->PMV[1][1], dec->future, FALSE);
+ } else assert(!backward);
+ break;
+ }
+ default: assert(0);
+ }
+ } else {
+ switch (mb->macroblock_modes.bits.field_motion_type) {
+ case PIPE_MPEG12_MO_TYPE_FIELD: goto mv1;
+ case PIPE_MPEG12_MO_TYPE_16x8: goto mv2;
+ case PIPE_MPEG12_MO_TYPE_DUAL_PRIME: {
+ base = NV17_MPEG_CMD_CHROMA_MV_HEADER_MV_SPLIT_HALF_MB;
+ if (frame)
+ base |= NV17_MPEG_CMD_CHROMA_MV_HEADER_TYPE_FRAME;
+ if (forward)
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE,
+ dec->picture_structure != PIPE_MPEG12_PICTURE_STRUCTURE_FIELD_TOP,
+ x, y, mb->PMV[0][0], dec->past, TRUE);
+ if (backward && forward)
+ nouveau_vpe_mb_mv(dec, base, luma, frame, FALSE,
+ dec->picture_structure == PIPE_MPEG12_PICTURE_STRUCTURE_FIELD_TOP,
+ x, y, mb->PMV[0][1], dec->future, TRUE);
+ else assert(!backward);
+ break;
+ }
+ default: assert(0);
+ }
+ }
+ return;
+
+mv1:
+ base = NV17_MPEG_CMD_CHROMA_MV_HEADER_MV_SPLIT_HALF_MB;
+ if (frame)
+ base |= NV17_MPEG_CMD_CHROMA_MV_HEADER_TYPE_FRAME;
+ /* frame 16x16 */
+ if (forward)
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE, FALSE,
+ x, y, mb->PMV[0][0], dec->past, TRUE);
+ if (backward)
+ nouveau_vpe_mb_mv(dec, base, luma, frame, !forward, FALSE,
+ x, y, mb->PMV[0][1], dec->future, TRUE);
+ return;
+
+mv2:
+ base = NV17_MPEG_CMD_CHROMA_MV_HEADER_COUNT_2;
+ if (!frame)
+ base |= NV17_MPEG_CMD_CHROMA_MV_HEADER_MV_SPLIT_HALF_MB;
+ if (forward) {
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE,
+ mb->motion_vertical_field_select & PIPE_MPEG12_FS_FIRST_FORWARD,
+ x, y, mb->PMV[0][0], dec->past, TRUE);
+ nouveau_vpe_mb_mv(dec, base, luma, frame, TRUE,
+ mb->motion_vertical_field_select & PIPE_MPEG12_FS_SECOND_FORWARD,
+ x, y2, mb->PMV[1][0], dec->past, FALSE);
+ }
+ if (backward) {
+ nouveau_vpe_mb_mv(dec, base, luma, frame, !forward,
+ mb->motion_vertical_field_select & PIPE_MPEG12_FS_FIRST_BACKWARD,
+ x, y, mb->PMV[0][1], dec->future, TRUE);
+ nouveau_vpe_mb_mv(dec, base, luma, frame, !forward,
+ mb->motion_vertical_field_select & PIPE_MPEG12_FS_SECOND_BACKWARD,
+ x, y2, mb->PMV[1][1], dec->future, FALSE);
+ }
+}
+
+static unsigned
+nouveau_decoder_surface_index(struct nouveau_decoder *dec,
+ struct pipe_video_buffer *buffer)
+{
+ struct nouveau_video_buffer *buf = (struct nouveau_video_buffer *)buffer;
+ struct nouveau_pushbuf *push = dec->push;
+ struct nouveau_bo *bo_y = nv04_resource(buf->resources[0])->bo;
+ struct nouveau_bo *bo_c = nv04_resource(buf->resources[1])->bo;
+
+ unsigned i;
+
+ if (!buf)
+ return 8;
+ for (i = 0; i < dec->num_surfaces; ++i) {
+ if (dec->surfaces[i] == buf)
+ return i;
+ }
+ assert(i < 8);
+ dec->surfaces[i] = buf;
+ dec->num_surfaces++;
+
+ nouveau_bufctx_reset(dec->bufctx, NV31_VIDEO_BIND_IMG(i));
+
+#define BCTX_ARGS dec->bufctx, NV31_VIDEO_BIND_IMG(i), NOUVEAU_BO_RDWR
+ BEGIN_NV04(push, NV31_MPEG(IMAGE_Y_OFFSET(i)), 2);
+ PUSH_MTHDl(push, NV31_MPEG(IMAGE_Y_OFFSET(i)), bo_y, 0, BCTX_ARGS);
+ PUSH_MTHDl(push, NV31_MPEG(IMAGE_C_OFFSET(i)), bo_c, 0, BCTX_ARGS);
+#undef BCTX_ARGS
+
+ return i;
+}
+
+static void
+nouveau_decoder_begin_frame(struct pipe_video_decoder *decoder,
+ struct pipe_video_buffer *target,
+ struct pipe_picture_desc *picture)
+{
+}
+
+static void
+nouveau_decoder_decode_macroblock(struct pipe_video_decoder *decoder,
+ struct pipe_video_buffer *target,
+ struct pipe_picture_desc *picture,
+ const struct pipe_macroblock *pipe_mb,
+ unsigned num_macroblocks)
+{
+ struct nouveau_decoder *dec = (struct nouveau_decoder *)decoder;
+ struct pipe_mpeg12_picture_desc *desc = (struct pipe_mpeg12_picture_desc*)picture;
+ const struct pipe_mpeg12_macroblock *mb;
+ unsigned i;
+ assert(target->width == decoder->width);
+ assert(target->height == decoder->height);
+
+ dec->current = nouveau_decoder_surface_index(dec, target);
+ assert(dec->current < 8);
+ dec->picture_structure = desc->picture_structure;
+ if (desc->ref[1])
+ dec->future = nouveau_decoder_surface_index(dec, desc->ref[1]);
+ if (desc->ref[0])
+ dec->past = nouveau_decoder_surface_index(dec, desc->ref[0]);
+
+ if (nouveau_vpe_init(dec)) return;
+ mb = (const struct pipe_mpeg12_macroblock *)pipe_mb;
+ for (i = 0; i < num_macroblocks; ++i, mb++) {
+ if (mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA) {
+ nouveau_vpe_mb_dct_header(dec, mb, TRUE);
+ nouveau_vpe_mb_dct_header(dec, mb, FALSE);
+ } else {
+ nouveau_vpe_mb_mv_header(dec, mb, TRUE);
+ nouveau_vpe_mb_dct_header(dec, mb, TRUE);
+
+ nouveau_vpe_mb_mv_header(dec, mb, FALSE);
+ nouveau_vpe_mb_dct_header(dec, mb, FALSE);
+ }
+ if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT)
+ nouveau_vpe_mb_dct_blocks(dec, mb);
+ else
+ nouveau_vpe_mb_data_blocks(dec, mb);
+ }
+}
+
+static void
+nouveau_decoder_end_frame(struct pipe_video_decoder *decoder,
+ struct pipe_video_buffer *target,
+ struct pipe_picture_desc *picture)
+{
+}
+
+static void
+nouveau_decoder_flush(struct pipe_video_decoder *decoder)
+{
+ struct nouveau_decoder *dec = (struct nouveau_decoder *)decoder;
+ if (dec->ofs)
+ nouveau_vpe_fini(dec);
+}
+
+static void
+nouveau_decoder_destroy(struct pipe_video_decoder *decoder)
+{
+ struct nouveau_decoder *dec = (struct nouveau_decoder*)decoder;
+
+ if (dec->data_bo)
+ nouveau_bo_ref(NULL, &dec->data_bo);
+ if (dec->cmd_bo)
+ nouveau_bo_ref(NULL, &dec->cmd_bo);
+ if (dec->fence_bo)
+ nouveau_bo_ref(NULL, &dec->fence_bo);
+
+ nouveau_object_del(&dec->mpeg);
+
+ if (dec->bufctx)
+ nouveau_bufctx_del(&dec->bufctx);
+ if (dec->push)
+ nouveau_pushbuf_del(&dec->push);
+ if (dec->client)
+ nouveau_client_del(&dec->client);
+ if (dec->chan)
+ nouveau_object_del(&dec->chan);
+
+ FREE(dec);
+}
+
+static struct pipe_video_decoder *
+nouveau_create_decoder(struct pipe_context *context,
+ struct nouveau_screen *screen,
+ enum pipe_video_profile profile,
+ enum pipe_video_entrypoint entrypoint,
+ enum pipe_video_chroma_format chroma_format,
+ unsigned width, unsigned height,
+ unsigned max_references, bool expect_chunked_decode)
+{
+ struct nv04_fifo nv04_data = { .vram = 0xbeef0201, .gart = 0xbeef0202 };
+ struct nouveau_object *mpeg = NULL;
+ struct nouveau_decoder *dec;
+ struct nouveau_pushbuf *push;
+ int ret;
+ bool is8274 = screen->device->chipset > 0x80;
+
+ debug_printf("Acceleration level: %s\n", entrypoint <= PIPE_VIDEO_ENTRYPOINT_BITSTREAM ? "bit":
+ entrypoint == PIPE_VIDEO_ENTRYPOINT_IDCT ? "IDCT" : "MC");
+
+ if (getenv("XVMC_VL"))
+ goto vl;
+ if (u_reduce_video_profile(profile) != PIPE_VIDEO_CODEC_MPEG12)
+ goto vl;
+ if (screen->device->chipset >= 0x98 && screen->device->chipset != 0xa0)
+ goto vl;
+
+ dec = CALLOC_STRUCT(nouveau_decoder);
+ if (!dec)
+ return NULL;
+
+ ret = nouveau_object_new(&screen->device->object, 0,
+ NOUVEAU_FIFO_CHANNEL_CLASS,
+ &nv04_data, sizeof(nv04_data), &dec->chan);
+ if (ret)
+ goto fail;
+ ret = nouveau_client_new(screen->device, &dec->client);
+ if (ret)
+ goto fail;
+ ret = nouveau_pushbuf_new(dec->client, dec->chan, 2, 4096, 1, &dec->push);
+ if (ret)
+ goto fail;
+ ret = nouveau_bufctx_new(dec->client, NV31_VIDEO_BIND_COUNT, &dec->bufctx);
+ if (ret)
+ goto fail;
+ push = dec->push;
+
+ width = align(width, 64);
+ height = align(height, 64);
+
+ if (is8274)
+ ret = nouveau_object_new(dec->chan, 0xbeef8274, NV84_MPEG_CLASS, NULL, 0,
+ &mpeg);
+ else
+ ret = nouveau_object_new(dec->chan, 0xbeef3174, NV31_MPEG_CLASS, NULL, 0,
+ &mpeg);
+ if (ret < 0) {
+ debug_printf("Creation failed: %s (%i)\n", strerror(-ret), ret);
+ return NULL;
+ }
+
+ dec->mpeg = mpeg;
+ dec->base.context = context;
+ dec->base.profile = profile;
+ dec->base.entrypoint = entrypoint;
+ dec->base.chroma_format = chroma_format;
+ dec->base.width = width;
+ dec->base.height = height;
+ dec->base.max_references = max_references;
+ dec->base.destroy = nouveau_decoder_destroy;
+ dec->base.begin_frame = nouveau_decoder_begin_frame;
+ dec->base.decode_macroblock = nouveau_decoder_decode_macroblock;
+ dec->base.end_frame = nouveau_decoder_end_frame;
+ dec->base.flush = nouveau_decoder_flush;
+ dec->screen = screen;
+
+ ret = nouveau_bo_new(dec->screen->device, NOUVEAU_BO_GART | NOUVEAU_BO_MAP,
+ 0, 1024 * 1024, NULL, &dec->cmd_bo);
+ if (ret)
+ goto fail;
+
+ ret = nouveau_bo_new(dec->screen->device, NOUVEAU_BO_GART | NOUVEAU_BO_MAP,
+ 0, width * height * 6, NULL, &dec->data_bo);
+ if (ret)
+ goto fail;
+
+ /* we don't need the fence, the kernel sync's for us */
+#if 0
+ ret = nouveau_bo_new(dec->screen->device, NOUVEAU_BO_GART | NOUVEAU_BO_MAP,
+ 0, 4096, NULL, &dec->fence_bo);
+ if (ret)
+ goto fail;
+ nouveau_bo_map(dec->fence_bo, NOUVEAU_BO_RDWR, NULL);
+ dec->fence_map = dec->fence_bo->map;
+ dec->fence_map[0] = 0;
+#endif
+
+ nouveau_pushbuf_bufctx(dec->push, dec->bufctx);
+ nouveau_pushbuf_space(push, 32, 4, 0);
+
+ BEGIN_NV04(push, SUBC_MPEG(NV01_SUBCHAN_OBJECT), 1);
+ PUSH_DATA (push, dec->mpeg->handle);
+
+ BEGIN_NV04(push, NV31_MPEG(DMA_CMD), 1);
+ PUSH_DATA (push, nv04_data.gart);
+
+ BEGIN_NV04(push, NV31_MPEG(DMA_DATA), 1);
+ PUSH_DATA (push, nv04_data.gart);
+
+ BEGIN_NV04(push, NV31_MPEG(DMA_IMAGE), 1);
+ PUSH_DATA (push, nv04_data.vram);
+
+ BEGIN_NV04(push, NV31_MPEG(PITCH), 2);
+ PUSH_DATA (push, width | NV31_MPEG_PITCH_UNK);
+ PUSH_DATA (push, (height << NV31_MPEG_SIZE_H__SHIFT) | width);
+
+ BEGIN_NV04(push, NV31_MPEG(FORMAT), 2);
+ PUSH_DATA (push, 0);
+ switch (entrypoint) {
+ case PIPE_VIDEO_ENTRYPOINT_BITSTREAM: PUSH_DATA (push, 0x100); break;
+ case PIPE_VIDEO_ENTRYPOINT_IDCT: PUSH_DATA (push, 1); break;
+ case PIPE_VIDEO_ENTRYPOINT_MC: PUSH_DATA (push, 0); break;
+ default: assert(0);
+ }
+
+ if (is8274) {
+ BEGIN_NV04(push, NV84_MPEG(DMA_QUERY), 1);
+ PUSH_DATA (push, nv04_data.vram);
+#if 0
+ BEGIN_NV04(push, NV84_MPEG(QUERY_OFFSET), 2);
+ PUSH_DATA (push, dec->fence_bo->offset);
+ PUSH_DATA (push, dec->fence_seq);
+#endif
+ }
+
+ ret = nouveau_vpe_init(dec);
+ if (ret)
+ goto fail;
+ nouveau_vpe_fini(dec);
+ return &dec->base;
+
+fail:
+ nouveau_decoder_destroy(&dec->base);
+ return NULL;
+
+vl:
+ debug_printf("Using g3dvl renderer\n");
+ return vl_create_decoder(context, profile, entrypoint,
+ chroma_format, width, height,
+ max_references, expect_chunked_decode);
+}
+
+static struct pipe_sampler_view **
+nouveau_video_buffer_sampler_view_planes(struct pipe_video_buffer *buffer)
+{
+ struct nouveau_video_buffer *buf = (struct nouveau_video_buffer *)buffer;
+ struct pipe_sampler_view sv_templ;
+ struct pipe_context *pipe;
+ unsigned i;
+
+ assert(buf);
+
+ pipe = buf->base.context;
+
+ for (i = 0; i < buf->num_planes; ++i ) {
+ if (!buf->sampler_view_planes[i]) {
+ memset(&sv_templ, 0, sizeof(sv_templ));
+ u_sampler_view_default_template(&sv_templ, buf->resources[i], buf->resources[i]->format);
+
+ if (util_format_get_nr_components(buf->resources[i]->format) == 1)
+ sv_templ.swizzle_r = sv_templ.swizzle_g = sv_templ.swizzle_b = sv_templ.swizzle_a = PIPE_SWIZZLE_RED;
+
+ buf->sampler_view_planes[i] = pipe->create_sampler_view(pipe, buf->resources[i], &sv_templ);
+ if (!buf->sampler_view_planes[i])
+ goto error;
+ }
+ }
+
+ return buf->sampler_view_planes;
+
+error:
+ for (i = 0; i < buf->num_planes; ++i )
+ pipe_sampler_view_reference(&buf->sampler_view_planes[i], NULL);
+
+ return NULL;
+}
+
+static struct pipe_sampler_view **
+nouveau_video_buffer_sampler_view_components(struct pipe_video_buffer *buffer)
+{
+ struct nouveau_video_buffer *buf = (struct nouveau_video_buffer *)buffer;
+ struct pipe_sampler_view sv_templ;
+ struct pipe_context *pipe;
+ unsigned i, j, component;
+
+ assert(buf);
+
+ pipe = buf->base.context;
+
+ for (component = 0, i = 0; i < buf->num_planes; ++i ) {
+ unsigned nr_components = util_format_get_nr_components(buf->resources[i]->format);
+
+ for (j = 0; j < nr_components; ++j, ++component) {
+ assert(component < VL_NUM_COMPONENTS);
+
+ if (!buf->sampler_view_components[component]) {
+ memset(&sv_templ, 0, sizeof(sv_templ));
+ u_sampler_view_default_template(&sv_templ, buf->resources[i], buf->resources[i]->format);
+ sv_templ.swizzle_r = sv_templ.swizzle_g = sv_templ.swizzle_b = PIPE_SWIZZLE_RED + j;
+ sv_templ.swizzle_a = PIPE_SWIZZLE_ONE;
+ buf->sampler_view_components[component] = pipe->create_sampler_view(pipe, buf->resources[i], &sv_templ);
+ if (!buf->sampler_view_components[component])
+ goto error;
+ }
+ }
+ }
+
+ return buf->sampler_view_components;
+
+error:
+ for (i = 0; i < 3; ++i )
+ pipe_sampler_view_reference(&buf->sampler_view_components[i], NULL);
+
+ return NULL;
+}
+
+static struct pipe_surface **
+nouveau_video_buffer_surfaces(struct pipe_video_buffer *buffer)
+{
+ struct nouveau_video_buffer *buf = (struct nouveau_video_buffer *)buffer;
+ struct pipe_surface surf_templ;
+ struct pipe_context *pipe;
+ unsigned i;
+
+ assert(buf);
+
+ pipe = buf->base.context;
+
+ for (i = 0; i < buf->num_planes; ++i ) {
+ if (!buf->surfaces[i]) {
+ memset(&surf_templ, 0, sizeof(surf_templ));
+ surf_templ.format = buf->resources[i]->format;
+ surf_templ.usage = PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET;
+ buf->surfaces[i] = pipe->create_surface(pipe, buf->resources[i], &surf_templ);
+ if (!buf->surfaces[i])
+ goto error;
+ }
+ }
+
+ return buf->surfaces;
+
+error:
+ for (i = 0; i < buf->num_planes; ++i )
+ pipe_surface_reference(&buf->surfaces[i], NULL);
+
+ return NULL;
+}
+
+static void
+nouveau_video_buffer_destroy(struct pipe_video_buffer *buffer)
+{
+ struct nouveau_video_buffer *buf = (struct nouveau_video_buffer *)buffer;
+ unsigned i;
+
+ assert(buf);
+
+ for (i = 0; i < buf->num_planes; ++i) {
+ pipe_surface_reference(&buf->surfaces[i], NULL);
+ pipe_sampler_view_reference(&buf->sampler_view_planes[i], NULL);
+ pipe_sampler_view_reference(&buf->sampler_view_components[i], NULL);
+ pipe_resource_reference(&buf->resources[i], NULL);
+ }
+ for (;i < 3;++i)
+ pipe_sampler_view_reference(&buf->sampler_view_components[i], NULL);
+
+ FREE(buffer);
+}
+
+static struct pipe_video_buffer *
+nouveau_video_buffer_create(struct pipe_context *pipe,
+ struct nouveau_screen *screen,
+ const struct pipe_video_buffer *templat)
+{
+ struct nouveau_video_buffer *buffer;
+ struct pipe_resource templ;
+ unsigned width, height;
+
+ /* Only do a linear surface when a hardware decoder is used
+ * hardware decoder is only supported on some chipsets
+ * and it only supports the NV12 format
+ */
+ if (templat->buffer_format != PIPE_FORMAT_NV12 || getenv("XVMC_VL") ||
+ (screen->device->chipset >= 0x98 && screen->device->chipset != 0xa0))
+ return vl_video_buffer_create(pipe, templat);
+
+ assert(templat->chroma_format == PIPE_VIDEO_CHROMA_FORMAT_420);
+ width = align(templat->width, 64);
+ height = align(templat->height, 64);
+
+ buffer = CALLOC_STRUCT(nouveau_video_buffer);
+ if (!buffer)
+ return NULL;
+
+ buffer->base.context = pipe;
+ buffer->base.destroy = nouveau_video_buffer_destroy;
+ buffer->base.get_sampler_view_planes = nouveau_video_buffer_sampler_view_planes;
+ buffer->base.get_sampler_view_components = nouveau_video_buffer_sampler_view_components;
+ buffer->base.get_surfaces = nouveau_video_buffer_surfaces;
+ buffer->base.chroma_format = templat->chroma_format;
+ buffer->base.width = width;
+ buffer->base.height = height;
+ buffer->num_planes = 2;
+
+ memset(&templ, 0, sizeof(templ));
+ templ.target = PIPE_TEXTURE_2D;
+ templ.format = PIPE_FORMAT_R8_UNORM;
+ templ.width0 = width;
+ templ.height0 = height;
+ templ.depth0 = 1;
+ templ.array_size = 1;
+ templ.bind = PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET;
+ templ.usage = PIPE_USAGE_STATIC;
+ templ.flags = NOUVEAU_RESOURCE_FLAG_LINEAR;
+
+ buffer->resources[0] = pipe->screen->resource_create(pipe->screen, &templ);
+ if (!buffer->resources[0])
+ goto error;
+ templ.width0 /= 2;
+ templ.height0 /= 2;
+ templ.format = PIPE_FORMAT_R8G8_UNORM;
+ buffer->resources[1] = pipe->screen->resource_create(pipe->screen, &templ);
+ if (!buffer->resources[1])
+ goto error;
+ return &buffer->base;
+
+error:
+ nouveau_video_buffer_destroy(&buffer->base);
+ return NULL;
+}