[FFmpeg-cvslog] vaapi_vp9: Convert to use the new VAAPI hwaccel code
Mark Thompson
git at videolan.org
Wed Jan 18 01:38:50 EET 2017
ffmpeg | branch: master | Mark Thompson <sw at jkqxz.net> | Sun Jan 8 15:44:01 2017 +0000| [defbb8bc26c7ff1a452a95dc64395bb30c7d7c4f] | committer: Mark Thompson
vaapi_vp9: Convert to use the new VAAPI hwaccel code
> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=defbb8bc26c7ff1a452a95dc64395bb30c7d7c4f
---
libavcodec/vaapi_vp9.c | 208 ++++++++++++++++++++++++++-----------------------
1 file changed, 110 insertions(+), 98 deletions(-)
diff --git a/libavcodec/vaapi_vp9.c b/libavcodec/vaapi_vp9.c
index 9b3e81a..a656ffb 100644
--- a/libavcodec/vaapi_vp9.c
+++ b/libavcodec/vaapi_vp9.c
@@ -21,135 +21,146 @@
*/
#include "libavutil/pixdesc.h"
-#include "vaapi_internal.h"
+#include "vaapi_decode.h"
#include "vp9.h"
-static void fill_picture_parameters(AVCodecContext *avctx,
- const VP9SharedContext *h,
- VADecPictureParameterBufferVP9 *pp)
+static VASurfaceID vaapi_vp9_surface_id(const VP9Frame *vf)
{
+ if (vf)
+ return ff_vaapi_get_surface_id(vf->tf.f);
+ else
+ return VA_INVALID_SURFACE;
+}
+
+static int vaapi_vp9_start_frame(AVCodecContext *avctx,
+ av_unused const uint8_t *buffer,
+ av_unused uint32_t size)
+{
+ const VP9SharedContext *h = avctx->priv_data;
+ VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private;
+ VADecPictureParameterBufferVP9 pic_param;
const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->sw_pix_fmt);
- int i;
-
- pp->frame_width = avctx->width;
- pp->frame_height = avctx->height;
-
- pp->frame_header_length_in_bytes = h->h.uncompressed_header_size;
- pp->first_partition_size = h->h.compressed_header_size;
-
- pp->profile = h->h.profile;
- pp->bit_depth = h->h.bpp;
-
- pp->filter_level = h->h.filter.level;
- pp->sharpness_level = h->h.filter.sharpness;
- pp->log2_tile_rows = h->h.tiling.log2_tile_rows;
- pp->log2_tile_columns = h->h.tiling.log2_tile_cols;
-
- pp->pic_fields.bits.subsampling_x = pixdesc->log2_chroma_w;
- pp->pic_fields.bits.subsampling_y = pixdesc->log2_chroma_h;
- pp->pic_fields.bits.frame_type = !h->h.keyframe;
- pp->pic_fields.bits.show_frame = !h->h.invisible;
- pp->pic_fields.bits.error_resilient_mode = h->h.errorres;
- pp->pic_fields.bits.intra_only = h->h.intraonly;
- pp->pic_fields.bits.allow_high_precision_mv = h->h.keyframe ? 0 : h->h.highprecisionmvs;
- pp->pic_fields.bits.mcomp_filter_type = h->h.filtermode ^ (h->h.filtermode <= 1);
- pp->pic_fields.bits.frame_parallel_decoding_mode = h->h.parallelmode;
- pp->pic_fields.bits.reset_frame_context = h->h.resetctx;
- pp->pic_fields.bits.refresh_frame_context = h->h.refreshctx;
- pp->pic_fields.bits.frame_context_idx = h->h.framectxid;
-
- pp->pic_fields.bits.segmentation_enabled = h->h.segmentation.enabled;
- pp->pic_fields.bits.segmentation_temporal_update = h->h.segmentation.temporal;
- pp->pic_fields.bits.segmentation_update_map = h->h.segmentation.update_map;
-
- pp->pic_fields.bits.last_ref_frame = h->h.refidx[0];
- pp->pic_fields.bits.last_ref_frame_sign_bias = h->h.signbias[0];
- pp->pic_fields.bits.golden_ref_frame = h->h.refidx[1];
- pp->pic_fields.bits.golden_ref_frame_sign_bias = h->h.signbias[1];
- pp->pic_fields.bits.alt_ref_frame = h->h.refidx[2];
- pp->pic_fields.bits.alt_ref_frame_sign_bias = h->h.signbias[2];
- pp->pic_fields.bits.lossless_flag = h->h.lossless;
+ int err, i;
+
+ pic->output_surface = vaapi_vp9_surface_id(&h->frames[CUR_FRAME]);
+
+ pic_param = (VADecPictureParameterBufferVP9) {
+ .frame_width = avctx->width,
+ .frame_height = avctx->height,
+
+ .pic_fields.bits = {
+ .subsampling_x = pixdesc->log2_chroma_w,
+ .subsampling_y = pixdesc->log2_chroma_h,
+ .frame_type = !h->h.keyframe,
+ .show_frame = !h->h.invisible,
+ .error_resilient_mode = h->h.errorres,
+ .intra_only = h->h.intraonly,
+ .allow_high_precision_mv = h->h.keyframe ? 0 : h->h.highprecisionmvs,
+ .mcomp_filter_type = h->h.filtermode ^ (h->h.filtermode <= 1),
+ .frame_parallel_decoding_mode = h->h.parallelmode,
+ .reset_frame_context = h->h.resetctx,
+ .refresh_frame_context = h->h.refreshctx,
+ .frame_context_idx = h->h.framectxid,
+
+ .segmentation_enabled = h->h.segmentation.enabled,
+ .segmentation_temporal_update = h->h.segmentation.temporal,
+ .segmentation_update_map = h->h.segmentation.update_map,
+
+ .last_ref_frame = h->h.refidx[0],
+ .last_ref_frame_sign_bias = h->h.signbias[0],
+ .golden_ref_frame = h->h.refidx[1],
+ .golden_ref_frame_sign_bias = h->h.signbias[1],
+ .alt_ref_frame = h->h.refidx[2],
+ .alt_ref_frame_sign_bias = h->h.signbias[2],
+ .lossless_flag = h->h.lossless,
+ },
+
+ .filter_level = h->h.filter.level,
+ .sharpness_level = h->h.filter.sharpness,
+ .log2_tile_rows = h->h.tiling.log2_tile_rows,
+ .log2_tile_columns = h->h.tiling.log2_tile_cols,
+
+ .frame_header_length_in_bytes = h->h.uncompressed_header_size,
+ .first_partition_size = h->h.compressed_header_size,
+
+ .profile = h->h.profile,
+ .bit_depth = h->h.bpp,
+ };
for (i = 0; i < 7; i++)
- pp->mb_segment_tree_probs[i] = h->h.segmentation.prob[i];
+ pic_param.mb_segment_tree_probs[i] = h->h.segmentation.prob[i];
if (h->h.segmentation.temporal) {
for (i = 0; i < 3; i++)
- pp->segment_pred_probs[i] = h->h.segmentation.pred_prob[i];
+ pic_param.segment_pred_probs[i] = h->h.segmentation.pred_prob[i];
} else {
- memset(pp->segment_pred_probs, 255, sizeof(pp->segment_pred_probs));
+ memset(pic_param.segment_pred_probs, 255, sizeof(pic_param.segment_pred_probs));
}
for (i = 0; i < 8; i++) {
- if (h->refs[i].f->buf[0]) {
- pp->reference_frames[i] = ff_vaapi_get_surface_id(h->refs[i].f);
- } else {
- pp->reference_frames[i] = VA_INVALID_ID;
- }
+ if (h->refs[i].f->buf[0])
+ pic_param.reference_frames[i] = ff_vaapi_get_surface_id(h->refs[i].f);
+ else
+ pic_param.reference_frames[i] = VA_INVALID_ID;
}
-}
-
-static int vaapi_vp9_start_frame(AVCodecContext *avctx,
- av_unused const uint8_t *buffer,
- av_unused uint32_t size)
-{
- const VP9SharedContext *h = avctx->priv_data;
- FFVAContext * const vactx = ff_vaapi_get_context(avctx);
- VADecPictureParameterBufferVP9 *pic_param;
-
- vactx->slice_param_size = sizeof(VASliceParameterBufferVP9);
- pic_param = ff_vaapi_alloc_pic_param(vactx, sizeof(VADecPictureParameterBufferVP9));
- if (!pic_param)
- return -1;
- fill_picture_parameters(avctx, h, pic_param);
+ err = ff_vaapi_decode_make_param_buffer(avctx, pic,
+ VAPictureParameterBufferType,
+ &pic_param, sizeof(pic_param));
+ if (err < 0) {
+ ff_vaapi_decode_cancel(avctx, pic);
+ return err;
+ }
return 0;
}
static int vaapi_vp9_end_frame(AVCodecContext *avctx)
{
- FFVAContext * const vactx = ff_vaapi_get_context(avctx);
const VP9SharedContext *h = avctx->priv_data;
- int ret;
-
- ret = ff_vaapi_commit_slices(vactx);
- if (ret < 0)
- goto finish;
-
- ret = ff_vaapi_render_picture(vactx, ff_vaapi_get_surface_id(h->frames[CUR_FRAME].tf.f));
- if (ret < 0)
- goto finish;
+ VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private;
-finish:
- ff_vaapi_common_end_frame(avctx);
- return ret;
+ return ff_vaapi_decode_issue(avctx, pic);
}
static int vaapi_vp9_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer,
uint32_t size)
{
- FFVAContext * const vactx = ff_vaapi_get_context(avctx);
const VP9SharedContext *h = avctx->priv_data;
- VASliceParameterBufferVP9 *slice_param;
- int i;
+ VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private;
+ VASliceParameterBufferVP9 slice_param;
+ int err, i;
- slice_param = (VASliceParameterBufferVP9*)ff_vaapi_alloc_slice(vactx, buffer, size);
- if (!slice_param)
- return -1;
+ slice_param = (VASliceParameterBufferVP9) {
+ .slice_data_size = size,
+ .slice_data_offset = 0,
+ .slice_data_flag = VA_SLICE_DATA_FLAG_ALL,
+ };
for (i = 0; i < 8; i++) {
- slice_param->seg_param[i].segment_flags.fields.segment_reference_enabled = h->h.segmentation.feat[i].ref_enabled;
- slice_param->seg_param[i].segment_flags.fields.segment_reference = h->h.segmentation.feat[i].ref_val;
- slice_param->seg_param[i].segment_flags.fields.segment_reference_skipped = h->h.segmentation.feat[i].skip_enabled;
-
- memcpy(slice_param->seg_param[i].filter_level, h->h.segmentation.feat[i].lflvl, sizeof(slice_param->seg_param[i].filter_level));
+ slice_param.seg_param[i] = (VASegmentParameterVP9) {
+ .segment_flags.fields = {
+ .segment_reference_enabled = h->h.segmentation.feat[i].ref_enabled,
+ .segment_reference = h->h.segmentation.feat[i].ref_val,
+ .segment_reference_skipped = h->h.segmentation.feat[i].skip_enabled,
+ },
+
+ .luma_dc_quant_scale = h->h.segmentation.feat[i].qmul[0][0],
+ .luma_ac_quant_scale = h->h.segmentation.feat[i].qmul[0][1],
+ .chroma_dc_quant_scale = h->h.segmentation.feat[i].qmul[1][0],
+ .chroma_ac_quant_scale = h->h.segmentation.feat[i].qmul[1][1],
+ };
+
+ memcpy(slice_param.seg_param[i].filter_level, h->h.segmentation.feat[i].lflvl, sizeof(slice_param.seg_param[i].filter_level));
+ }
- slice_param->seg_param[i].luma_dc_quant_scale = h->h.segmentation.feat[i].qmul[0][0];
- slice_param->seg_param[i].luma_ac_quant_scale = h->h.segmentation.feat[i].qmul[0][1];
- slice_param->seg_param[i].chroma_dc_quant_scale = h->h.segmentation.feat[i].qmul[1][0];
- slice_param->seg_param[i].chroma_ac_quant_scale = h->h.segmentation.feat[i].qmul[1][1];
+ err = ff_vaapi_decode_make_slice_buffer(avctx, pic,
+ &slice_param, sizeof(slice_param),
+ buffer, size);
+ if (err) {
+ ff_vaapi_decode_cancel(avctx, pic);
+ return err;
}
return 0;
@@ -163,7 +174,8 @@ AVHWAccel ff_vp9_vaapi_hwaccel = {
.start_frame = vaapi_vp9_start_frame,
.end_frame = vaapi_vp9_end_frame,
.decode_slice = vaapi_vp9_decode_slice,
- .init = ff_vaapi_context_init,
- .uninit = ff_vaapi_context_fini,
- .priv_data_size = sizeof(FFVAContext),
+ .frame_priv_data_size = sizeof(VAAPIDecodePicture),
+ .init = ff_vaapi_decode_init,
+ .uninit = ff_vaapi_decode_uninit,
+ .priv_data_size = sizeof(VAAPIDecodeContext),
};
More information about the ffmpeg-cvslog
mailing list