FREE(ctx);
}
-static int
-vl_context_get_param(struct pipe_video_context *context, int param)
-{
- struct vl_context *ctx = (struct vl_context*)context;
-
- assert(context);
-
- if (param == PIPE_CAP_NPOT_TEXTURES)
- return !ctx->pot_buffers;
-
- debug_printf("vl_context: Unknown PIPE_CAP %d\n", param);
- return 0;
-}
-
static boolean
vl_context_is_format_supported(struct pipe_video_context *context,
enum pipe_format format,
{
struct vl_context *ctx = (struct vl_context*)context;
unsigned buffer_width, buffer_height;
+ bool pot_buffers;
assert(context);
assert(width > 0 && height > 0);
+
+ pot_buffers = !ctx->base.screen->get_video_param(ctx->base.screen, profile, PIPE_VIDEO_CAP_NPOT_TEXTURES);
- buffer_width = ctx->pot_buffers ? util_next_power_of_two(width) : align(width, MACROBLOCK_WIDTH);
- buffer_height = ctx->pot_buffers ? util_next_power_of_two(height) : align(height, MACROBLOCK_HEIGHT);
+ buffer_width = pot_buffers ? util_next_power_of_two(width) : align(width, MACROBLOCK_WIDTH);
+ buffer_height = pot_buffers ? util_next_power_of_two(height) : align(height, MACROBLOCK_HEIGHT);
switch (u_reduce_video_profile(profile)) {
case PIPE_VIDEO_CODEC_MPEG12:
const enum pipe_format *resource_formats;
struct pipe_video_buffer *result;
unsigned buffer_width, buffer_height;
+ bool pot_buffers;
assert(context);
assert(width > 0 && height > 0);
+ pot_buffers = !ctx->base.screen->get_video_param
+ (
+ ctx->base.screen,
+ PIPE_VIDEO_PROFILE_UNKNOWN,
+ PIPE_VIDEO_CAP_NPOT_TEXTURES
+ );
+
resource_formats = vl_video_buffer_formats(ctx->pipe, buffer_format);
if (!resource_formats)
return NULL;
- buffer_width = ctx->pot_buffers ? util_next_power_of_two(width) : align(width, MACROBLOCK_WIDTH);
- buffer_height = ctx->pot_buffers ? util_next_power_of_two(height) : align(height, MACROBLOCK_HEIGHT);
+ buffer_width = pot_buffers ? util_next_power_of_two(width) : align(width, MACROBLOCK_WIDTH);
+ buffer_height = pot_buffers ? util_next_power_of_two(height) : align(height, MACROBLOCK_HEIGHT);
result = vl_video_buffer_init(context, ctx->pipe,
buffer_width, buffer_height, 1,
}
struct pipe_video_context *
-vl_create_context(struct pipe_context *pipe, bool pot_buffers)
+vl_create_context(struct pipe_context *pipe)
{
struct vl_context *ctx;
ctx->base.screen = pipe->screen;
ctx->base.destroy = vl_context_destroy;
- ctx->base.get_param = vl_context_get_param;
ctx->base.is_format_supported = vl_context_is_format_supported;
ctx->base.create_surface = vl_context_create_surface;
ctx->base.create_sampler_view = vl_context_create_sampler_view;
ctx->base.create_compositor = vl_context_create_compositor;
ctx->pipe = pipe;
- ctx->pot_buffers = pot_buffers;
return &ctx->base;
}
{
struct pipe_video_context base;
struct pipe_context *pipe;
- bool pot_buffers;
};
/* drivers can call this function in their pipe_video_context constructors and pass it
an accelerated pipe_context along with suitable buffering modes, etc */
struct pipe_video_context *
-vl_create_context(struct pipe_context *pipe, bool pot_buffers);
+vl_create_context(struct pipe_context *pipe);
#endif /* vl_context_h */
}
}
+static int r300_get_video_param(struct pipe_screen *screen,
+ enum pipe_video_profile profile,
+ enum pipe_video_cap param)
+{
+ switch (param) {
+ case PIPE_VIDEO_CAP_NPOT_TEXTURES:
+ return 0;
+ default:
+ return 0;
+ }
+}
+
static boolean r300_is_format_supported(struct pipe_screen* screen,
enum pipe_format format,
enum pipe_texture_target target,
r300screen->screen.get_param = r300_get_param;
r300screen->screen.get_shader_param = r300_get_shader_param;
r300screen->screen.get_paramf = r300_get_paramf;
+ r300screen->screen.get_video_param = r300_get_video_param;
r300screen->screen.is_format_supported = r300_is_format_supported;
r300screen->screen.context_create = r300_create_context;
r300screen->screen.video_context_create = r300_video_create;
if (!pipe)
return NULL;
- return vl_create_context(pipe, false);
+ return vl_create_context(pipe);
}
}
}
+static int r600_get_video_param(struct pipe_screen *screen,
+ enum pipe_video_profile profile,
+ enum pipe_video_cap param)
+{
+ switch (param) {
+ case PIPE_VIDEO_CAP_NPOT_TEXTURES:
+ return 1;
+ default:
+ return 0;
+ }
+}
+
static boolean r600_is_format_supported(struct pipe_screen* screen,
enum pipe_format format,
enum pipe_texture_target target,
rscreen->screen.get_param = r600_get_param;
rscreen->screen.get_shader_param = r600_get_shader_param;
rscreen->screen.get_paramf = r600_get_paramf;
+ rscreen->screen.get_video_param = r600_get_video_param;
rscreen->screen.is_format_supported = r600_is_format_supported;
rscreen->screen.context_create = r600_create_context;
rscreen->screen.video_context_create = r600_video_create;
if (!pipe)
return NULL;
- return vl_create_context(pipe, false);
+ return vl_create_context(pipe);
}
}
}
+static int
+softpipe_get_video_param(struct pipe_screen *screen,
+ enum pipe_video_profile profile,
+ enum pipe_video_cap param)
+{
+ switch (param) {
+ case PIPE_VIDEO_CAP_NPOT_TEXTURES:
+ return 0;
+ default:
+ return 0;
+ }
+}
/**
* Query format support for creating a texture, drawing surface, etc.
return NULL;
/* TODO: Use slice buffering for softpipe when implemented, no advantage to buffering an entire picture with softpipe */
- return vl_create_context(pipe, true);
+ return vl_create_context(pipe);
}
/**
screen->base.get_param = softpipe_get_param;
screen->base.get_shader_param = softpipe_get_shader_param;
screen->base.get_paramf = softpipe_get_paramf;
+ screen->base.get_video_param = softpipe_get_video_param;
screen->base.is_format_supported = softpipe_is_format_supported;
screen->base.context_create = softpipe_create_context;
screen->base.flush_frontbuffer = softpipe_flush_frontbuffer;
PIPE_SHADER_CAP_SUBROUTINES = 16, /* BGNSUB, ENDSUB, CAL, RET */
};
+/* Video caps, can be different for each codec/profile */
+enum pipe_video_cap
+{
+ PIPE_VIDEO_CAP_NPOT_TEXTURES = 0,
+};
enum pipe_video_codec
{
*/
int (*get_shader_param)( struct pipe_screen *, unsigned shader, enum pipe_shader_cap param );
+ /**
+ * Query an integer-valued capability/parameter/limit for a codec/profile
+ * \param param one of PIPE_VIDEO_CAP_x
+ */
+ int (*get_video_param)( struct pipe_screen *, enum pipe_video_profile profile, enum pipe_video_cap param );
+
struct pipe_context * (*context_create)( struct pipe_screen *, void *priv );
struct pipe_video_context * (*video_context_create)( struct pipe_screen *screen, void *priv );
*/
void (*destroy)(struct pipe_video_context *context);
- /**
- * Query an integer-valued capability/parameter/limit
- * \param param one of PIPE_CAP_x
- */
- int (*get_param)(struct pipe_video_context *context, int param);
-
/**
* Check if the given pipe_format is supported as a video buffer
*/
tex_templ.target = PIPE_TEXTURE_2D;
tex_templ.format = XvIDToPipe(xvimage_id);
tex_templ.last_level = 0;
- if (vpipe->get_param(vpipe, PIPE_CAP_NPOT_TEXTURES)) {
+ if (vpipe->screen->get_video_param(vpipe->screen,
+ PIPE_VIDEO_PROFILE_UNKNOWN,
+ PIPE_VIDEO_CAP_NPOT_TEXTURES)) {
tex_templ.width0 = width;
tex_templ.height0 = height;
}