st/nine: Access pipe_context via NineDevice9_GetPipe
[mesa.git] / src / gallium / state_trackers / vdpau / decode.c
index ec3995b98db3e7cfa025adad1dca5a0388d5b0fa..f85bce823bb032dd5ad768f1b3fac54e2b448e6a 100644 (file)
  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
  * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
- * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
+ * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
  * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
  * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
  * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
  *
  **************************************************************************/
 
+#include "util/u_memory.h"
+#include "util/u_math.h"
+#include "util/u_debug.h"
+#include "util/u_video.h"
+
+#include "vl/vl_vlc.h"
+
 #include "vdpau_private.h"
-#include <util/u_memory.h>
-#include <util/u_math.h>
-#include <pipe/p_video_context.h>
-#include <util/u_debug.h>
 
+/**
+ * Create a VdpDecoder.
+ */
 VdpStatus
-vlVdpDecoderCreate (   VdpDevice device, 
-                                               VdpDecoderProfile profile, 
-                                               uint32_t width, uint32_t height, 
-                                               uint32_t max_references, 
-                                               VdpDecoder *decoder 
-)
+vlVdpDecoderCreate(VdpDevice device,
+                   VdpDecoderProfile profile,
+                   uint32_t width, uint32_t height,
+                   uint32_t max_references,
+                   VdpDecoder *decoder)
 {
-       struct vl_screen *vscreen;
-       enum pipe_video_profile p_profile;
-       VdpStatus ret;
-       vlVdpDecoder *vldecoder;
-       
-       debug_printf("[VDPAU] Creating decoder\n");
-       
-       if (!decoder)
-               return VDP_STATUS_INVALID_POINTER;
-       
-       if (!(width && height))
-               return VDP_STATUS_INVALID_VALUE;
-               
-   vlVdpDevice *dev = vlGetDataHTAB(device);
-   if (!dev)  {
-      ret = VDP_STATUS_INVALID_HANDLE;
-      goto inv_device;
+   struct pipe_video_codec templat = {};
+   struct pipe_context *pipe;
+   struct pipe_screen *screen;
+   vlVdpDevice *dev;
+   vlVdpDecoder *vldecoder;
+   VdpStatus ret;
+   bool supported;
+   uint32_t maxwidth, maxheight;
+
+   if (!decoder)
+      return VDP_STATUS_INVALID_POINTER;
+   *decoder = 0;
+
+   if (!(width && height))
+      return VDP_STATUS_INVALID_VALUE;
+
+   templat.profile = ProfileToPipe(profile);
+   if (templat.profile == PIPE_VIDEO_PROFILE_UNKNOWN)
+      return VDP_STATUS_INVALID_DECODER_PROFILE;
+
+   dev = vlGetDataHTAB(device);
+   if (!dev)
+      return VDP_STATUS_INVALID_HANDLE;
+
+   pipe = dev->context;
+   screen = dev->vscreen->pscreen;
+
+   pipe_mutex_lock(dev->mutex);
+
+   supported = screen->get_video_param
+   (
+      screen,
+      templat.profile,
+      PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+      PIPE_VIDEO_CAP_SUPPORTED
+   );
+   if (!supported) {
+      pipe_mutex_unlock(dev->mutex);
+      return VDP_STATUS_INVALID_DECODER_PROFILE;
+   }
+
+   maxwidth = screen->get_video_param
+   (
+      screen,
+      templat.profile,
+      PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+      PIPE_VIDEO_CAP_MAX_WIDTH
+   );
+   maxheight = screen->get_video_param
+   (
+      screen,
+      templat.profile,
+      PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+      PIPE_VIDEO_CAP_MAX_HEIGHT
+   );
+   if (width > maxwidth || height > maxheight) {
+      pipe_mutex_unlock(dev->mutex);
+      return VDP_STATUS_INVALID_SIZE;
    }
-   
+
    vldecoder = CALLOC(1,sizeof(vlVdpDecoder));
-   if (!vldecoder)   {
-          ret = VDP_STATUS_RESOURCES;
-          goto no_decoder;
+   if (!vldecoder) {
+      pipe_mutex_unlock(dev->mutex);
+      return VDP_STATUS_RESOURCES;
    }
-   
-   p_profile = ProfileToPipe(profile);
-   if (p_profile == PIPE_VIDEO_PROFILE_UNKNOWN)        {
-          ret = VDP_STATUS_INVALID_DECODER_PROFILE;
-          goto inv_profile;
+
+   DeviceReference(&vldecoder->device, dev);
+
+   templat.entrypoint = PIPE_VIDEO_ENTRYPOINT_BITSTREAM;
+   templat.chroma_format = PIPE_VIDEO_CHROMA_FORMAT_420;
+   templat.width = width;
+   templat.height = height;
+   templat.max_references = max_references;
+
+   if (u_reduce_video_profile(templat.profile) ==
+       PIPE_VIDEO_FORMAT_MPEG4_AVC)
+      templat.level = u_get_h264_level(templat.width, templat.height,
+                            &templat.max_references);
+
+   vldecoder->decoder = pipe->create_video_codec(pipe, &templat);
+
+   if (!vldecoder->decoder) {
+      ret = VDP_STATUS_ERROR;
+      goto error_decoder;
    }
 
-       // TODO: Define max_references. Used mainly for H264
-       
-       vldecoder->profile = p_profile;
-       vldecoder->device = dev;
-               
-       *decoder = vlAddDataHTAB(vldecoder);
-       if (*decoder == 0) {
+   *decoder = vlAddDataHTAB(vldecoder);
+   if (*decoder == 0) {
       ret = VDP_STATUS_ERROR;
-      goto no_handle;
-       }
-       debug_printf("[VDPAU] Decoder created succesfully\n");
-       
-       return VDP_STATUS_OK;
-       
-       no_handle:
-       FREE(vldecoder);
-       inv_profile:
-       no_screen:
-       no_decoder:
-       inv_device:
-    return ret;
+      goto error_handle;
+   }
+
+   pipe_mutex_init(vldecoder->mutex);
+   pipe_mutex_unlock(dev->mutex);
+
+   return VDP_STATUS_OK;
+
+error_handle:
+   vldecoder->decoder->destroy(vldecoder->decoder);
+
+error_decoder:
+   pipe_mutex_unlock(dev->mutex);
+   DeviceReference(&vldecoder->device, NULL);
+   FREE(vldecoder);
+   return ret;
 }
 
+/**
+ * Destroy a VdpDecoder.
+ */
 VdpStatus
-vlVdpDecoderDestroy  (VdpDecoder decoder
-)
+vlVdpDecoderDestroy(VdpDecoder decoder)
 {
-       vlVdpDecoder *vldecoder;
-       
-       vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder);
-       if (!vldecoder)  {
+   vlVdpDecoder *vldecoder;
+
+   vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder);
+   if (!vldecoder)
       return VDP_STATUS_INVALID_HANDLE;
-       }
-       
-       if (vldecoder->vctx->vscreen)
-               vl_screen_destroy(vldecoder->vctx->vscreen);
-       
-       if (vldecoder->vctx)
-               vl_video_destroy(vldecoder->vctx);
-               
-       FREE(vldecoder);
-       
-       return VDP_STATUS_OK;
+
+   pipe_mutex_lock(vldecoder->mutex);
+   vldecoder->decoder->destroy(vldecoder->decoder);
+   pipe_mutex_unlock(vldecoder->mutex);
+   pipe_mutex_destroy(vldecoder->mutex);
+
+   vlRemoveDataHTAB(decoder);
+   DeviceReference(&vldecoder->device, NULL);
+   FREE(vldecoder);
+
+   return VDP_STATUS_OK;
 }
 
+/**
+ * Retrieve the parameters used to create a VdpDecoder.
+ */
 VdpStatus
-vlVdpCreateSurfaceTarget   (vlVdpDecoder *vldecoder,
-                                                       vlVdpSurface *vlsurf
-)
+vlVdpDecoderGetParameters(VdpDecoder decoder,
+                          VdpDecoderProfile *profile,
+                          uint32_t *width,
+                          uint32_t *height)
+{
+   vlVdpDecoder *vldecoder;
+
+   vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder);
+   if (!vldecoder)
+      return VDP_STATUS_INVALID_HANDLE;
+
+   *profile = PipeToProfile(vldecoder->decoder->profile);
+   *width = vldecoder->decoder->width;
+   *height = vldecoder->decoder->height;
+
+   return VDP_STATUS_OK;
+}
+
+static VdpStatus
+vlVdpGetReferenceFrame(VdpVideoSurface handle, struct pipe_video_buffer **ref_frame)
 {
-       struct pipe_resource tmplt;
-       struct pipe_resource *surf_tex;
-       struct pipe_video_context *vpipe;
-               
-       if(!(vldecoder && vlsurf))
-               return VDP_STATUS_INVALID_POINTER;
-               
-       vpipe = vldecoder->vctx;
-               
-       memset(&tmplt, 0, sizeof(struct pipe_resource));
-       tmplt.target = PIPE_TEXTURE_2D;
-       tmplt.format = vlsurf->format;
-       tmplt.last_level = 0;
-       if (vpipe->is_format_supported(vpipe, tmplt.format,
-                                  PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET,
-                                  PIPE_TEXTURE_GEOM_NON_POWER_OF_TWO)) {
-      tmplt.width0 = vlsurf->width;
-      tmplt.height0 = vlsurf->height;
-    }
-    else {
-      assert(vpipe->is_format_supported(vpipe, tmplt.format,
-                                       PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET,
-                                       PIPE_TEXTURE_GEOM_NON_SQUARE));
-      tmplt.width0 = util_next_power_of_two(vlsurf->width);
-      tmplt.height0 = util_next_power_of_two(vlsurf->height);
-    }
-       tmplt.depth0 = 1;
-       tmplt.usage = PIPE_USAGE_DEFAULT;
-       tmplt.bind = PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET;
-       tmplt.flags = 0;
-       
-       surf_tex = vpipe->screen->resource_create(vpipe->screen, &tmplt);
-       
-       vlsurf->psurface = vpipe->screen->get_tex_surface(vpipe->screen, surf_tex, 0, 0, 0,
-                                         PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET);
-                                                                                
-       pipe_resource_reference(&surf_tex, NULL);
-       
-       if (!vlsurf->psurface)
-               return VDP_STATUS_RESOURCES;
-       
-       
-       return VDP_STATUS_OK;
+   vlVdpSurface *surface;
+
+   /* if surfaces equals VDP_STATUS_INVALID_HANDLE, they are not used */
+   if (handle ==  VDP_INVALID_HANDLE) {
+      *ref_frame = NULL;
+      return VDP_STATUS_OK;
+   }
+
+   surface = vlGetDataHTAB(handle);
+   if (!surface)
+      return VDP_STATUS_INVALID_HANDLE;
+
+   *ref_frame = surface->video_buffer;
+   if (!*ref_frame)
+         return VDP_STATUS_INVALID_HANDLE;
+
+   return VDP_STATUS_OK;
 }
 
-static void
-vlVdpMacroBlocksToPipe(struct pipe_screen *screen,
-                  VdpBitstreamBuffer const *bitstream_buffers,
-                  unsigned int num_macroblocks,
-                  struct pipe_mpeg12_macroblock *pipe_macroblocks)
+/**
+ * Decode a mpeg 1/2 video.
+ */
+static VdpStatus
+vlVdpDecoderRenderMpeg12(struct pipe_mpeg12_picture_desc *picture,
+                         VdpPictureInfoMPEG1Or2 *picture_info)
 {
-       debug_printf("NAF!\n");
+   VdpStatus r;
+
+   VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding MPEG12\n");
+
+   r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   picture->picture_coding_type = picture_info->picture_coding_type;
+   picture->picture_structure = picture_info->picture_structure;
+   picture->frame_pred_frame_dct = picture_info->frame_pred_frame_dct;
+   picture->q_scale_type = picture_info->q_scale_type;
+   picture->alternate_scan = picture_info->alternate_scan;
+   picture->intra_vlc_format = picture_info->intra_vlc_format;
+   picture->concealment_motion_vectors = picture_info->concealment_motion_vectors;
+   picture->intra_dc_precision = picture_info->intra_dc_precision;
+   picture->f_code[0][0] = picture_info->f_code[0][0] - 1;
+   picture->f_code[0][1] = picture_info->f_code[0][1] - 1;
+   picture->f_code[1][0] = picture_info->f_code[1][0] - 1;
+   picture->f_code[1][1] = picture_info->f_code[1][1] - 1;
+   picture->num_slices = picture_info->slice_count;
+   picture->top_field_first = picture_info->top_field_first;
+   picture->full_pel_forward_vector = picture_info->full_pel_forward_vector;
+   picture->full_pel_backward_vector = picture_info->full_pel_backward_vector;
+   picture->intra_matrix = picture_info->intra_quantizer_matrix;
+   picture->non_intra_matrix = picture_info->non_intra_quantizer_matrix;
+
+   return VDP_STATUS_OK;
 }
 
-VdpStatus
-vlVdpDecoderRenderMpeg2    (vlVdpDecoder *vldecoder,
-                                                       vlVdpSurface *vlsurf,
-                                                       VdpPictureInfoMPEG1Or2 *picture_info,
-                                                       uint32_t bitstream_buffer_count,
-                                                       VdpBitstreamBuffer const *bitstream_buffers
-                                                       )
+/**
+ * Decode a mpeg 4 video.
+ */
+static VdpStatus
+vlVdpDecoderRenderMpeg4(struct pipe_mpeg4_picture_desc *picture,
+                        VdpPictureInfoMPEG4Part2 *picture_info)
+{
+   VdpStatus r;
+   unsigned i;
+
+   VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding MPEG4\n");
+
+   r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   for (i = 0; i < 2; ++i) {
+      picture->trd[i] = picture_info->trd[i];
+      picture->trb[i] = picture_info->trb[i];
+   }
+   picture->vop_time_increment_resolution = picture_info->vop_time_increment_resolution;
+   picture->vop_coding_type = picture_info->vop_coding_type;
+   picture->vop_fcode_forward = picture_info->vop_fcode_forward;
+   picture->vop_fcode_backward = picture_info->vop_fcode_backward;
+   picture->resync_marker_disable = picture_info->resync_marker_disable;
+   picture->interlaced = picture_info->interlaced;
+   picture->quant_type = picture_info->quant_type;
+   picture->quarter_sample = picture_info->quarter_sample;
+   picture->short_video_header = picture_info->short_video_header;
+   picture->rounding_control = picture_info->rounding_control;
+   picture->alternate_vertical_scan_flag = picture_info->alternate_vertical_scan_flag;
+   picture->top_field_first = picture_info->top_field_first;
+   picture->intra_matrix = picture_info->intra_quantizer_matrix;
+   picture->non_intra_matrix = picture_info->non_intra_quantizer_matrix;
+
+   return VDP_STATUS_OK;
+}
+
+static VdpStatus
+vlVdpDecoderRenderVC1(struct pipe_vc1_picture_desc *picture,
+                      VdpPictureInfoVC1 *picture_info)
+{
+   VdpStatus r;
+
+   VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding VC-1\n");
+
+   r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]);
+   if (r != VDP_STATUS_OK)
+      return r;
+
+   picture->slice_count = picture_info->slice_count;
+   picture->picture_type = picture_info->picture_type;
+   picture->frame_coding_mode = picture_info->frame_coding_mode;
+   picture->postprocflag = picture_info->postprocflag;
+   picture->pulldown = picture_info->pulldown;
+   picture->interlace = picture_info->interlace;
+   picture->tfcntrflag = picture_info->tfcntrflag;
+   picture->finterpflag = picture_info->finterpflag;
+   picture->psf = picture_info->psf;
+   picture->dquant = picture_info->dquant;
+   picture->panscan_flag = picture_info->panscan_flag;
+   picture->refdist_flag = picture_info->refdist_flag;
+   picture->quantizer = picture_info->quantizer;
+   picture->extended_mv = picture_info->extended_mv;
+   picture->extended_dmv = picture_info->extended_dmv;
+   picture->overlap = picture_info->overlap;
+   picture->vstransform = picture_info->vstransform;
+   picture->loopfilter = picture_info->loopfilter;
+   picture->fastuvmc = picture_info->fastuvmc;
+   picture->range_mapy_flag = picture_info->range_mapy_flag;
+   picture->range_mapy = picture_info->range_mapy;
+   picture->range_mapuv_flag = picture_info->range_mapuv_flag;
+   picture->range_mapuv = picture_info->range_mapuv;
+   picture->multires = picture_info->multires;
+   picture->syncmarker = picture_info->syncmarker;
+   picture->rangered = picture_info->rangered;
+   picture->maxbframes = picture_info->maxbframes;
+   picture->deblockEnable = picture_info->deblockEnable;
+   picture->pquant = picture_info->pquant;
+
+   return VDP_STATUS_OK;
+}
+
+static VdpStatus
+vlVdpDecoderRenderH264(struct pipe_h264_picture_desc *picture,
+                       VdpPictureInfoH264 *picture_info)
+{
+   unsigned i;
+
+   VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding H264\n");
+
+   picture->pps->sps->mb_adaptive_frame_field_flag = picture_info->mb_adaptive_frame_field_flag;
+   picture->pps->sps->frame_mbs_only_flag = picture_info->frame_mbs_only_flag;
+   picture->pps->sps->log2_max_frame_num_minus4 = picture_info->log2_max_frame_num_minus4;
+   picture->pps->sps->pic_order_cnt_type = picture_info->pic_order_cnt_type;
+   picture->pps->sps->log2_max_pic_order_cnt_lsb_minus4 = picture_info->log2_max_pic_order_cnt_lsb_minus4;
+   picture->pps->sps->delta_pic_order_always_zero_flag = picture_info->delta_pic_order_always_zero_flag;
+   picture->pps->sps->direct_8x8_inference_flag = picture_info->direct_8x8_inference_flag;
+
+   picture->pps->transform_8x8_mode_flag = picture_info->transform_8x8_mode_flag;
+   picture->pps->chroma_qp_index_offset = picture_info->chroma_qp_index_offset;
+   picture->pps->second_chroma_qp_index_offset = picture_info->second_chroma_qp_index_offset;
+   picture->pps->pic_init_qp_minus26 = picture_info->pic_init_qp_minus26;
+   picture->pps->entropy_coding_mode_flag = picture_info->entropy_coding_mode_flag;
+   picture->pps->deblocking_filter_control_present_flag = picture_info->deblocking_filter_control_present_flag;
+   picture->pps->redundant_pic_cnt_present_flag = picture_info->redundant_pic_cnt_present_flag;
+   picture->pps->constrained_intra_pred_flag = picture_info->constrained_intra_pred_flag;
+   picture->pps->weighted_pred_flag = picture_info->weighted_pred_flag;
+   picture->pps->weighted_bipred_idc = picture_info->weighted_bipred_idc;
+   picture->pps->bottom_field_pic_order_in_frame_present_flag = picture_info->pic_order_present_flag;
+   memcpy(picture->pps->ScalingList4x4, picture_info->scaling_lists_4x4, 6*16);
+   memcpy(picture->pps->ScalingList8x8, picture_info->scaling_lists_8x8, 2*64);
+
+   picture->slice_count = picture_info->slice_count;
+   picture->field_order_cnt[0] = picture_info->field_order_cnt[0];
+   picture->field_order_cnt[1] = picture_info->field_order_cnt[1];
+   picture->is_reference = picture_info->is_reference;
+   picture->frame_num = picture_info->frame_num;
+   picture->field_pic_flag = picture_info->field_pic_flag;
+   picture->bottom_field_flag = picture_info->bottom_field_flag;
+   picture->num_ref_frames = picture_info->num_ref_frames;
+
+   picture->num_ref_idx_l0_active_minus1 = picture_info->num_ref_idx_l0_active_minus1;
+   picture->num_ref_idx_l1_active_minus1 = picture_info->num_ref_idx_l1_active_minus1;
+
+   for (i = 0; i < 16; ++i) {
+      VdpStatus ret = vlVdpGetReferenceFrame
+      (
+         picture_info->referenceFrames[i].surface,
+         &picture->ref[i]
+      );
+      if (ret != VDP_STATUS_OK)
+         return ret;
+
+      picture->is_long_term[i] = picture_info->referenceFrames[i].is_long_term;
+      picture->top_is_reference[i] = picture_info->referenceFrames[i].top_is_reference;
+      picture->bottom_is_reference[i] = picture_info->referenceFrames[i].bottom_is_reference;
+      picture->field_order_cnt_list[i][0] = picture_info->referenceFrames[i].field_order_cnt[0];
+      picture->field_order_cnt_list[i][1] = picture_info->referenceFrames[i].field_order_cnt[1];
+      picture->frame_num_list[i] = picture_info->referenceFrames[i].frame_idx;
+   }
+
+   return VDP_STATUS_OK;
+}
+
+static VdpStatus
+vlVdpDecoderRenderH265(struct pipe_h265_picture_desc *picture,
+                       VdpPictureInfoHEVC *picture_info)
+{
+   unsigned i;
+
+   picture->pps->sps->chroma_format_idc = picture_info->chroma_format_idc;
+   picture->pps->sps->separate_colour_plane_flag = picture_info->separate_colour_plane_flag;
+   picture->pps->sps->pic_width_in_luma_samples = picture_info->pic_width_in_luma_samples;
+   picture->pps->sps->pic_height_in_luma_samples = picture_info->pic_height_in_luma_samples;
+   picture->pps->sps->bit_depth_luma_minus8 = picture_info->bit_depth_luma_minus8;
+   picture->pps->sps->bit_depth_chroma_minus8 = picture_info->bit_depth_chroma_minus8;
+   picture->pps->sps->log2_max_pic_order_cnt_lsb_minus4 = picture_info->log2_max_pic_order_cnt_lsb_minus4;
+   picture->pps->sps->sps_max_dec_pic_buffering_minus1 = picture_info->sps_max_dec_pic_buffering_minus1;
+   picture->pps->sps->log2_min_luma_coding_block_size_minus3 = picture_info->log2_min_luma_coding_block_size_minus3;
+   picture->pps->sps->log2_diff_max_min_luma_coding_block_size = picture_info->log2_diff_max_min_luma_coding_block_size;
+   picture->pps->sps->log2_min_transform_block_size_minus2 = picture_info->log2_min_transform_block_size_minus2;
+   picture->pps->sps->log2_diff_max_min_transform_block_size = picture_info->log2_diff_max_min_transform_block_size;
+   picture->pps->sps->max_transform_hierarchy_depth_inter = picture_info->max_transform_hierarchy_depth_inter;
+   picture->pps->sps->max_transform_hierarchy_depth_intra = picture_info->max_transform_hierarchy_depth_intra;
+   picture->pps->sps->scaling_list_enabled_flag = picture_info->scaling_list_enabled_flag;
+   memcpy(picture->pps->sps->ScalingList4x4, picture_info->ScalingList4x4, 6*16);
+   memcpy(picture->pps->sps->ScalingList8x8, picture_info->ScalingList8x8, 6*64);
+   memcpy(picture->pps->sps->ScalingList16x16, picture_info->ScalingList16x16, 6*64);
+   memcpy(picture->pps->sps->ScalingList32x32, picture_info->ScalingList32x32, 2*64);
+   memcpy(picture->pps->sps->ScalingListDCCoeff16x16, picture_info->ScalingListDCCoeff16x16, 6);
+   memcpy(picture->pps->sps->ScalingListDCCoeff32x32, picture_info->ScalingListDCCoeff32x32, 2);
+   picture->pps->sps->amp_enabled_flag = picture_info->amp_enabled_flag;
+   picture->pps->sps->sample_adaptive_offset_enabled_flag = picture_info->sample_adaptive_offset_enabled_flag;
+   picture->pps->sps->pcm_enabled_flag = picture_info->pcm_enabled_flag;
+   picture->pps->sps->pcm_sample_bit_depth_luma_minus1 = picture_info->pcm_sample_bit_depth_luma_minus1;
+   picture->pps->sps->pcm_sample_bit_depth_chroma_minus1 = picture_info->pcm_sample_bit_depth_chroma_minus1;
+   picture->pps->sps->log2_min_pcm_luma_coding_block_size_minus3 = picture_info->log2_min_pcm_luma_coding_block_size_minus3;
+   picture->pps->sps->log2_diff_max_min_pcm_luma_coding_block_size = picture_info->log2_diff_max_min_pcm_luma_coding_block_size;
+   picture->pps->sps->pcm_loop_filter_disabled_flag = picture_info->pcm_loop_filter_disabled_flag;
+   picture->pps->sps->num_short_term_ref_pic_sets = picture_info->num_short_term_ref_pic_sets;
+   picture->pps->sps->long_term_ref_pics_present_flag = picture_info->long_term_ref_pics_present_flag;
+   picture->pps->sps->num_long_term_ref_pics_sps = picture_info->num_long_term_ref_pics_sps;
+   picture->pps->sps->sps_temporal_mvp_enabled_flag = picture_info->sps_temporal_mvp_enabled_flag;
+   picture->pps->sps->strong_intra_smoothing_enabled_flag = picture_info->strong_intra_smoothing_enabled_flag;
+
+   picture->pps->dependent_slice_segments_enabled_flag = picture_info->dependent_slice_segments_enabled_flag;
+   picture->pps->output_flag_present_flag = picture_info->output_flag_present_flag;
+   picture->pps->num_extra_slice_header_bits = picture_info->num_extra_slice_header_bits;
+   picture->pps->sign_data_hiding_enabled_flag = picture_info->sign_data_hiding_enabled_flag;
+   picture->pps->cabac_init_present_flag = picture_info->cabac_init_present_flag;
+   picture->pps->num_ref_idx_l0_default_active_minus1 = picture_info->num_ref_idx_l0_default_active_minus1;
+   picture->pps->num_ref_idx_l1_default_active_minus1 = picture_info->num_ref_idx_l1_default_active_minus1;
+   picture->pps->init_qp_minus26 = picture_info->init_qp_minus26;
+   picture->pps->constrained_intra_pred_flag = picture_info->constrained_intra_pred_flag;
+   picture->pps->transform_skip_enabled_flag = picture_info->transform_skip_enabled_flag;
+   picture->pps->cu_qp_delta_enabled_flag = picture_info->cu_qp_delta_enabled_flag;
+   picture->pps->diff_cu_qp_delta_depth = picture_info->diff_cu_qp_delta_depth;
+   picture->pps->pps_cb_qp_offset = picture_info->pps_cb_qp_offset;
+   picture->pps->pps_cr_qp_offset = picture_info->pps_cr_qp_offset;
+   picture->pps->pps_slice_chroma_qp_offsets_present_flag = picture_info->pps_slice_chroma_qp_offsets_present_flag;
+   picture->pps->weighted_pred_flag = picture_info->weighted_pred_flag;
+   picture->pps->weighted_bipred_flag = picture_info->weighted_bipred_flag;
+   picture->pps->transquant_bypass_enabled_flag = picture_info->transquant_bypass_enabled_flag;
+   picture->pps->tiles_enabled_flag = picture_info->tiles_enabled_flag;
+   picture->pps->entropy_coding_sync_enabled_flag = picture_info->entropy_coding_sync_enabled_flag;
+   picture->pps->num_tile_columns_minus1 = picture_info->num_tile_columns_minus1;
+   picture->pps->num_tile_rows_minus1 = picture_info->num_tile_rows_minus1;
+   picture->pps->uniform_spacing_flag = picture_info->uniform_spacing_flag;
+   memcpy(picture->pps->column_width_minus1, picture_info->column_width_minus1, 20 * 2);
+   memcpy(picture->pps->row_height_minus1, picture_info->row_height_minus1, 22 * 2);
+   picture->pps->loop_filter_across_tiles_enabled_flag = picture_info->loop_filter_across_tiles_enabled_flag;
+   picture->pps->pps_loop_filter_across_slices_enabled_flag = picture_info->pps_loop_filter_across_slices_enabled_flag;
+   picture->pps->deblocking_filter_control_present_flag = picture_info->deblocking_filter_control_present_flag;
+   picture->pps->deblocking_filter_override_enabled_flag = picture_info->deblocking_filter_override_enabled_flag;
+   picture->pps->pps_deblocking_filter_disabled_flag = picture_info->pps_deblocking_filter_disabled_flag;
+   picture->pps->pps_beta_offset_div2 = picture_info->pps_beta_offset_div2;
+   picture->pps->pps_tc_offset_div2 = picture_info->pps_tc_offset_div2;
+   picture->pps->lists_modification_present_flag = picture_info->lists_modification_present_flag;
+   picture->pps->log2_parallel_merge_level_minus2 = picture_info->log2_parallel_merge_level_minus2;
+   picture->pps->slice_segment_header_extension_present_flag = picture_info->slice_segment_header_extension_present_flag;
+
+   picture->IDRPicFlag = picture_info->IDRPicFlag;
+   picture->RAPPicFlag = picture_info->RAPPicFlag;
+   picture->CurrRpsIdx = picture_info->CurrRpsIdx;
+   picture->NumPocTotalCurr = picture_info->NumPocTotalCurr;
+   picture->NumDeltaPocsOfRefRpsIdx = picture_info->NumDeltaPocsOfRefRpsIdx;
+   picture->NumShortTermPictureSliceHeaderBits = picture_info->NumShortTermPictureSliceHeaderBits;
+   picture->NumLongTermPictureSliceHeaderBits = picture_info->NumLongTermPictureSliceHeaderBits;
+   picture->CurrPicOrderCntVal = picture_info->CurrPicOrderCntVal;
+
+   for (i = 0; i < 16; ++i) {
+      VdpStatus ret = vlVdpGetReferenceFrame
+      (
+         picture_info->RefPics[i],
+         &picture->ref[i]
+      );
+      if (ret != VDP_STATUS_OK)
+         return ret;
+
+      picture->PicOrderCntVal[i] = picture_info->PicOrderCntVal[i];
+      picture->IsLongTerm[i] = picture_info->IsLongTerm[i];
+   }
+
+   picture->NumPocStCurrBefore = picture_info->NumPocStCurrBefore;
+   picture->NumPocStCurrAfter = picture_info->NumPocStCurrAfter;
+   picture->NumPocLtCurr = picture_info->NumPocLtCurr;
+   memcpy(picture->RefPicSetStCurrBefore, picture_info->RefPicSetStCurrBefore, 8);
+   memcpy(picture->RefPicSetStCurrAfter, picture_info->RefPicSetStCurrAfter, 8);
+   memcpy(picture->RefPicSetLtCurr, picture_info->RefPicSetLtCurr, 8);
+   picture->UseRefPicList = false;
+
+   return VDP_STATUS_OK;
+}
+
+static void
+vlVdpDecoderFixVC1Startcode(uint32_t *num_buffers, const void *buffers[], unsigned sizes[])
 {
-       struct pipe_video_context *vpipe;
-       vlVdpSurface *t_vdp_surf;
-       vlVdpSurface *p_vdp_surf;
-       vlVdpSurface *f_vdp_surf;
-       struct pipe_surface *t_surf;
-       struct pipe_surface *p_surf;
-       struct pipe_surface *f_surf;
-       uint32_t num_macroblocks;
-       VdpStatus ret;
-       
-
-       vpipe = vldecoder->vctx->vpipe;
-       t_vdp_surf = vlsurf;
-       
-       /* if surfaces equals VDP_STATUS_INVALID_HANDLE, they are not used */
-       if (picture_info->backward_reference ==  VDP_INVALID_HANDLE) 
-               p_vdp_surf = NULL;
-       else    {
-               p_vdp_surf = (vlVdpSurface *)vlGetDataHTAB(picture_info->backward_reference);
-               if (!p_vdp_surf)
-                       return VDP_STATUS_INVALID_HANDLE;
-       }
-
-       if (picture_info->forward_reference ==  VDP_INVALID_HANDLE) 
-               f_vdp_surf = NULL;
-       else    {
-               f_vdp_surf = (vlVdpSurface *)vlGetDataHTAB(picture_info->forward_reference);
-               if (!f_vdp_surf)
-                       return VDP_STATUS_INVALID_HANDLE;
-       }
-               
-       
-       if (f_vdp_surf ==  VDP_INVALID_HANDLE) f_vdp_surf = NULL;
-       
-       ret = vlVdpCreateSurfaceTarget(vldecoder,t_vdp_surf);
-               
-       num_macroblocks = bitstream_buffer_count;
-       struct pipe_mpeg12_macroblock pipe_macroblocks[num_macroblocks];
-       
-       vlVdpMacroBlocksToPipe(vpipe->screen, bitstream_buffers,
-                     num_macroblocks, pipe_macroblocks);
-               
-       vpipe->set_decode_target(vpipe,t_surf);
-       vpipe->decode_macroblocks(vpipe, p_surf, f_surf, num_macroblocks, &pipe_macroblocks->base, NULL);
-       return ret;
+   static const uint8_t vc1_startcode[] = { 0x00, 0x00, 0x01, 0x0D };
+   struct vl_vlc vlc;
+   unsigned i;
+
+   /* search the first 64 bytes for a startcode */
+   vl_vlc_init(&vlc, *num_buffers, buffers, sizes);
+   while (vl_vlc_search_byte(&vlc, 64*8, 0x00) && vl_vlc_bits_left(&vlc) >= 32) {
+      uint32_t value = vl_vlc_peekbits(&vlc, 32);
+      if (value == 0x0000010D ||
+          value == 0x0000010C ||
+          value == 0x0000010B)
+         return;
+      vl_vlc_eatbits(&vlc, 8);
+   }
+
+   /* none found, ok add one manually */
+   VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Manually adding VC-1 startcode\n");
+   for (i = *num_buffers; i > 0; --i) {
+      buffers[i] = buffers[i - 1];
+      sizes[i] = sizes[i - 1];
+   }
+   ++(*num_buffers);
+   buffers[0] = vc1_startcode;
+   sizes[0] = 4;
 }
 
+/**
+ * Decode a compressed field/frame and render the result into a VdpVideoSurface.
+ */
 VdpStatus
-vlVdpDecoderRender (VdpDecoder decoder, 
-                                       VdpVideoSurface target, 
-                                       VdpPictureInfo const *picture_info, 
-                                       uint32_t bitstream_buffer_count, 
-                                       VdpBitstreamBuffer const *bitstream_buffers
-)
+vlVdpDecoderRender(VdpDecoder decoder,
+                   VdpVideoSurface target,
+                   VdpPictureInfo const *picture_info,
+                   uint32_t bitstream_buffer_count,
+                   VdpBitstreamBuffer const *bitstream_buffers)
 {
-       vlVdpDecoder *vldecoder;
-       vlVdpSurface *vlsurf;
-       struct vl_screen *vscreen;
-       VdpStatus ret;
-       debug_printf("[VDPAU] Decoding\n");
-               
-       if (!(picture_info && bitstream_buffers))
-               return VDP_STATUS_INVALID_POINTER;
-       
-       
-       vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder);
-       if (!vldecoder)
-               return VDP_STATUS_INVALID_HANDLE;
-
-       vlsurf = (vlVdpSurface *)vlGetDataHTAB(target);
-       if (!vlsurf)
-               return VDP_STATUS_INVALID_HANDLE;
-       
-       if (vlsurf->device != vldecoder->device)
-               return VDP_STATUS_HANDLE_DEVICE_MISMATCH;
-               
-       if (vlsurf->chroma_format != vldecoder->chroma_format)
-               return VDP_STATUS_INVALID_CHROMA_TYPE;
-               
-       vscreen = vl_screen_create(vldecoder->device->display, vldecoder->device->screen);
-       if (!vscreen)
-               return VDP_STATUS_RESOURCES;
-       
-       vldecoder->vctx = vl_video_create(vscreen, vldecoder->profile, vlsurf->format, vlsurf->width, vlsurf->height);
-       if (!vldecoder->vctx)
-               return VDP_STATUS_RESOURCES;
-               
-       vldecoder->vctx->vscreen = vscreen;
-               
-    // TODO: Right now only mpeg2 is supported.
-       switch (vldecoder->vctx->vpipe->profile)   {
-               case PIPE_VIDEO_PROFILE_MPEG2_SIMPLE:
-               case PIPE_VIDEO_PROFILE_MPEG2_MAIN:
-                       ret = vlVdpDecoderRenderMpeg2(vldecoder,vlsurf,(VdpPictureInfoMPEG1Or2 *)picture_info,
-                                                                                       bitstream_buffer_count,bitstream_buffers);
-                       break;
-               default:
-                       return VDP_STATUS_INVALID_DECODER_PROFILE;
-       }
-       assert(0);
-
-       return ret;
-}
\ No newline at end of file
+   const void * buffers[bitstream_buffer_count + 1];
+   unsigned sizes[bitstream_buffer_count + 1];
+   vlVdpDecoder *vldecoder;
+   vlVdpSurface *vlsurf;
+   VdpStatus ret;
+   struct pipe_screen *screen;
+   struct pipe_video_codec *dec;
+   bool buffer_support[2];
+   unsigned i;
+   struct pipe_h264_sps sps_h264 = {};
+   struct pipe_h264_pps pps_h264 = { &sps_h264 };
+   struct pipe_h265_sps sps_h265 = {};
+   struct pipe_h265_pps pps_h265 = { &sps_h265 };
+   union {
+      struct pipe_picture_desc base;
+      struct pipe_mpeg12_picture_desc mpeg12;
+      struct pipe_mpeg4_picture_desc mpeg4;
+      struct pipe_vc1_picture_desc vc1;
+      struct pipe_h264_picture_desc h264;
+      struct pipe_h265_picture_desc h265;
+   } desc;
+
+   if (!(picture_info && bitstream_buffers))
+      return VDP_STATUS_INVALID_POINTER;
+
+   vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder);
+   if (!vldecoder)
+      return VDP_STATUS_INVALID_HANDLE;
+   dec = vldecoder->decoder;
+   screen = dec->context->screen;
+
+   vlsurf = (vlVdpSurface *)vlGetDataHTAB(target);
+   if (!vlsurf)
+      return VDP_STATUS_INVALID_HANDLE;
+
+   if (vlsurf->device != vldecoder->device)
+      return VDP_STATUS_HANDLE_DEVICE_MISMATCH;
+
+   if (vlsurf->video_buffer != NULL && vlsurf->video_buffer->chroma_format != dec->chroma_format)
+      // TODO: Recreate decoder with correct chroma
+      return VDP_STATUS_INVALID_CHROMA_TYPE;
+
+   buffer_support[0] = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+                                               PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE);
+   buffer_support[1] = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+                                               PIPE_VIDEO_CAP_SUPPORTS_INTERLACED);
+
+   if (vlsurf->video_buffer == NULL ||
+       !screen->is_video_format_supported(screen, vlsurf->video_buffer->buffer_format,
+                                          dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM) ||
+       !buffer_support[vlsurf->video_buffer->interlaced]) {
+
+      pipe_mutex_lock(vlsurf->device->mutex);
+
+      /* destroy the old one */
+      if (vlsurf->video_buffer)
+         vlsurf->video_buffer->destroy(vlsurf->video_buffer);
+
+      /* set the buffer format to the prefered one */
+      vlsurf->templat.buffer_format = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+                                                              PIPE_VIDEO_CAP_PREFERED_FORMAT);
+
+      /* also set interlacing to decoders preferences */
+      vlsurf->templat.interlaced = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM,
+                                                           PIPE_VIDEO_CAP_PREFERS_INTERLACED);
+
+      /* and recreate the video buffer */
+      vlsurf->video_buffer = dec->context->create_video_buffer(dec->context, &vlsurf->templat);
+
+      /* still no luck? get me out of here... */
+      if (!vlsurf->video_buffer) {
+         pipe_mutex_unlock(vlsurf->device->mutex);
+         return VDP_STATUS_NO_IMPLEMENTATION;
+      }
+      vlVdpVideoSurfaceClear(vlsurf);
+      pipe_mutex_unlock(vlsurf->device->mutex);
+   }
+
+   for (i = 0; i < bitstream_buffer_count; ++i) {
+      buffers[i] = bitstream_buffers[i].bitstream;
+      sizes[i] = bitstream_buffers[i].bitstream_bytes;
+   }
+
+   memset(&desc, 0, sizeof(desc));
+   desc.base.profile = dec->profile;
+   switch (u_reduce_video_profile(dec->profile)) {
+   case PIPE_VIDEO_FORMAT_MPEG12:
+      ret = vlVdpDecoderRenderMpeg12(&desc.mpeg12, (VdpPictureInfoMPEG1Or2 *)picture_info);
+      break;
+   case PIPE_VIDEO_FORMAT_MPEG4:
+      ret = vlVdpDecoderRenderMpeg4(&desc.mpeg4, (VdpPictureInfoMPEG4Part2 *)picture_info);
+      break;
+   case PIPE_VIDEO_FORMAT_VC1:
+      if (dec->profile == PIPE_VIDEO_PROFILE_VC1_ADVANCED)
+         vlVdpDecoderFixVC1Startcode(&bitstream_buffer_count, buffers, sizes);
+      ret = vlVdpDecoderRenderVC1(&desc.vc1, (VdpPictureInfoVC1 *)picture_info);
+      break;
+   case PIPE_VIDEO_FORMAT_MPEG4_AVC:
+      desc.h264.pps = &pps_h264;
+      ret = vlVdpDecoderRenderH264(&desc.h264, (VdpPictureInfoH264 *)picture_info);
+      break;
+   case PIPE_VIDEO_FORMAT_HEVC:
+      desc.h265.pps = &pps_h265;
+      ret = vlVdpDecoderRenderH265(&desc.h265, (VdpPictureInfoHEVC *)picture_info);
+      break;
+   default:
+      return VDP_STATUS_INVALID_DECODER_PROFILE;
+   }
+
+   if (ret != VDP_STATUS_OK)
+      return ret;
+
+   pipe_mutex_lock(vldecoder->mutex);
+   dec->begin_frame(dec, vlsurf->video_buffer, &desc.base);
+   dec->decode_bitstream(dec, vlsurf->video_buffer, &desc.base, bitstream_buffer_count, buffers, sizes);
+   dec->end_frame(dec, vlsurf->video_buffer, &desc.base);
+   pipe_mutex_unlock(vldecoder->mutex);
+   return ret;
+}