diff options
author | Boyuan Zhang <[email protected]> | 2016-07-21 19:40:18 -0400 |
---|---|---|
committer | Christian König <[email protected]> | 2016-07-25 13:39:52 +0200 |
commit | 85d807f2e04eb4f096fa619bc61c65cdcef446c8 (patch) | |
tree | d04dd1bf67658764c37e8daf9126ba052af4546f /src/gallium | |
parent | 10c1cc47a6adba9ea83d4df5c154c07f95ac3a40 (diff) |
st/va: add functions for VAAPI encode
Add necessary functions/changes for VAAPI encoding to buffer and picture. These changes will allow driver to handle all Vaapi encode related operations. This patch doesn't change the Vaapi decode behaviour.
Signed-off-by: Boyuan Zhang <[email protected]>
Diffstat (limited to 'src/gallium')
-rw-r--r-- | src/gallium/state_trackers/va/buffer.c | 6 | ||||
-rw-r--r-- | src/gallium/state_trackers/va/picture.c | 172 | ||||
-rw-r--r-- | src/gallium/state_trackers/va/va_private.h | 2 |
3 files changed, 178 insertions, 2 deletions
diff --git a/src/gallium/state_trackers/va/buffer.c b/src/gallium/state_trackers/va/buffer.c index 7d3167b2ce2..dfcebbe4bd5 100644 --- a/src/gallium/state_trackers/va/buffer.c +++ b/src/gallium/state_trackers/va/buffer.c @@ -133,6 +133,12 @@ vlVaMapBuffer(VADriverContextP ctx, VABufferID buf_id, void **pbuff) if (!buf->derived_surface.transfer || !*pbuff) return VA_STATUS_ERROR_INVALID_BUFFER; + if (buf->type == VAEncCodedBufferType) { + ((VACodedBufferSegment*)buf->data)->buf = *pbuff; + ((VACodedBufferSegment*)buf->data)->size = buf->coded_size; + ((VACodedBufferSegment*)buf->data)->next = NULL; + *pbuff = buf->data; + } } else { pipe_mutex_unlock(drv->mutex); *pbuff = buf->data; diff --git a/src/gallium/state_trackers/va/picture.c b/src/gallium/state_trackers/va/picture.c index 89ac02458f4..4bb60f20d9d 100644 --- a/src/gallium/state_trackers/va/picture.c +++ b/src/gallium/state_trackers/va/picture.c @@ -78,7 +78,8 @@ vlVaBeginPicture(VADriverContextP ctx, VAContextID context_id, VASurfaceID rende return VA_STATUS_SUCCESS; } - context->decoder->begin_frame(context->decoder, context->target, &context->desc.base); + if (context->decoder->entrypoint != PIPE_VIDEO_ENTRYPOINT_ENCODE) + context->decoder->begin_frame(context->decoder, context->target, &context->desc.base); return VA_STATUS_SUCCESS; } @@ -278,6 +279,142 @@ handleVASliceDataBufferType(vlVaContext *context, vlVaBuffer *buf) num_buffers, (const void * const*)buffers, sizes); } +static VAStatus +handleVAEncMiscParameterTypeRateControl(vlVaContext *context, VAEncMiscParameterBuffer *misc) +{ + VAEncMiscParameterRateControl *rc = (VAEncMiscParameterRateControl *)misc->data; + if (context->desc.h264enc.rate_ctrl.rate_ctrl_method == + PIPE_H264_ENC_RATE_CONTROL_METHOD_CONSTANT) + context->desc.h264enc.rate_ctrl.target_bitrate = rc->bits_per_second; + else + context->desc.h264enc.rate_ctrl.target_bitrate = rc->bits_per_second * rc->target_percentage; + context->desc.h264enc.rate_ctrl.peak_bitrate = rc->bits_per_second; + if (context->desc.h264enc.rate_ctrl.target_bitrate < 2000000) + context->desc.h264enc.rate_ctrl.vbv_buffer_size = MIN2((context->desc.h264enc.rate_ctrl.target_bitrate * 2.75), 2000000); + else + context->desc.h264enc.rate_ctrl.vbv_buffer_size = context->desc.h264enc.rate_ctrl.target_bitrate; + context->desc.h264enc.rate_ctrl.target_bits_picture = + context->desc.h264enc.rate_ctrl.target_bitrate / context->desc.h264enc.rate_ctrl.frame_rate_num; + context->desc.h264enc.rate_ctrl.peak_bits_picture_integer = + context->desc.h264enc.rate_ctrl.peak_bitrate / context->desc.h264enc.rate_ctrl.frame_rate_num; + context->desc.h264enc.rate_ctrl.peak_bits_picture_fraction = 0; + + return VA_STATUS_SUCCESS; +} + +static VAStatus +handleVAEncSequenceParameterBufferType(vlVaDriver *drv, vlVaContext *context, vlVaBuffer *buf) +{ + VAEncSequenceParameterBufferH264 *h264 = (VAEncSequenceParameterBufferH264 *)buf->data; + if (!context->decoder) { + context->templat.max_references = h264->max_num_ref_frames; + context->templat.level = h264->level_idc; + context->decoder = drv->pipe->create_video_codec(drv->pipe, &context->templat); + if (!context->decoder) + return VA_STATUS_ERROR_ALLOCATION_FAILED; + } + context->desc.h264enc.gop_size = h264->intra_idr_period; + context->desc.h264enc.rate_ctrl.frame_rate_num = h264->time_scale / 2; + context->desc.h264enc.rate_ctrl.frame_rate_den = 1; + return VA_STATUS_SUCCESS; +} + +static VAStatus +handleVAEncMiscParameterBufferType(vlVaContext *context, vlVaBuffer *buf) +{ + VAStatus vaStatus = VA_STATUS_SUCCESS; + VAEncMiscParameterBuffer *misc; + misc = buf->data; + + switch (misc->type) { + case VAEncMiscParameterTypeRateControl: + vaStatus = handleVAEncMiscParameterTypeRateControl(context, misc); + break; + + default: + break; + } + + return vaStatus; +} + +static VAStatus +handleVAEncPictureParameterBufferType(vlVaDriver *drv, vlVaContext *context, vlVaBuffer *buf) +{ + VAEncPictureParameterBufferH264 *h264; + vlVaBuffer *coded_buf; + + h264 = buf->data; + context->desc.h264enc.frame_num = h264->frame_num; + context->desc.h264enc.not_referenced = false; + context->desc.h264enc.is_idr = (h264->pic_fields.bits.idr_pic_flag == 1); + context->desc.h264enc.pic_order_cnt = h264->CurrPic.TopFieldOrderCnt / 2; + if (context->desc.h264enc.is_idr) + context->desc.h264enc.i_remain = 1; + else + context->desc.h264enc.i_remain = 0; + + context->desc.h264enc.p_remain = context->desc.h264enc.gop_size - context->desc.h264enc.gop_cnt - context->desc.h264enc.i_remain; + + coded_buf = handle_table_get(drv->htab, h264->coded_buf); + if (!coded_buf->derived_surface.resource) + coded_buf->derived_surface.resource = pipe_buffer_create(drv->pipe->screen, PIPE_BIND_VERTEX_BUFFER, + PIPE_USAGE_STREAM, coded_buf->size); + context->coded_buf = coded_buf; + + context->desc.h264enc.frame_idx[h264->CurrPic.picture_id] = h264->frame_num; + if (context->desc.h264enc.is_idr) + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_IDR; + else + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_P; + + context->desc.h264enc.quant_i_frames = h264->pic_init_qp; + context->desc.h264enc.quant_b_frames = h264->pic_init_qp; + context->desc.h264enc.quant_p_frames = h264->pic_init_qp; + context->desc.h264enc.frame_num_cnt++; + context->desc.h264enc.gop_cnt++; + if (context->desc.h264enc.gop_cnt == context->desc.h264enc.gop_size) + context->desc.h264enc.gop_cnt = 0; + + return VA_STATUS_SUCCESS; +} + +static VAStatus +handleVAEncSliceParameterBufferType(vlVaDriver *drv, vlVaContext *context, vlVaBuffer *buf) +{ + VAEncSliceParameterBufferH264 *h264; + + h264 = buf->data; + context->desc.h264enc.ref_idx_l0 = VA_INVALID_ID; + context->desc.h264enc.ref_idx_l1 = VA_INVALID_ID; + + for (int i = 0; i < 32; i++) { + if (h264->RefPicList0[i].picture_id != VA_INVALID_ID) { + if (context->desc.h264enc.ref_idx_l0 == VA_INVALID_ID) + context->desc.h264enc.ref_idx_l0 = context->desc.h264enc.frame_idx[h264->RefPicList0[i].picture_id]; + } + if (h264->RefPicList1[i].picture_id != VA_INVALID_ID && h264->slice_type == 1) { + if (context->desc.h264enc.ref_idx_l1 == VA_INVALID_ID) + context->desc.h264enc.ref_idx_l1 = context->desc.h264enc.frame_idx[h264->RefPicList1[i].picture_id]; + } + } + + if (h264->slice_type == 1) + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_B; + else if (h264->slice_type == 0) + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_P; + else if (h264->slice_type == 2) { + if (context->desc.h264enc.is_idr){ + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_IDR; + context->desc.h264enc.idr_pic_id++; + } else + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_I; + } else + context->desc.h264enc.picture_type = PIPE_H264_ENC_PICTURE_TYPE_SKIP; + + return VA_STATUS_SUCCESS; +} + VAStatus vlVaRenderPicture(VADriverContextP ctx, VAContextID context_id, VABufferID *buffers, int num_buffers) { @@ -328,6 +465,22 @@ vlVaRenderPicture(VADriverContextP ctx, VAContextID context_id, VABufferID *buff vaStatus = vlVaHandleVAProcPipelineParameterBufferType(drv, context, buf); break; + case VAEncSequenceParameterBufferType: + vaStatus = handleVAEncSequenceParameterBufferType(drv, context, buf); + break; + + case VAEncMiscParameterBufferType: + vaStatus = handleVAEncMiscParameterBufferType(context, buf); + break; + + case VAEncPictureParameterBufferType: + vaStatus = handleVAEncPictureParameterBufferType(drv, context, buf); + break; + + case VAEncSliceParameterBufferType: + vaStatus = handleVAEncSliceParameterBufferType(drv, context, buf); + break; + default: break; } @@ -342,6 +495,9 @@ vlVaEndPicture(VADriverContextP ctx, VAContextID context_id) { vlVaDriver *drv; vlVaContext *context; + vlVaBuffer *coded_buf; + unsigned int coded_size; + void *feedback; if (!ctx) return VA_STATUS_ERROR_INVALID_CONTEXT; @@ -365,7 +521,19 @@ vlVaEndPicture(VADriverContextP ctx, VAContextID context_id) } context->mpeg4.frame_num++; - context->decoder->end_frame(context->decoder, context->target, &context->desc.base); + + if (context->decoder->entrypoint == PIPE_VIDEO_ENTRYPOINT_ENCODE) { + coded_buf = context->coded_buf; + context->decoder->begin_frame(context->decoder, context->target, &context->desc.base); + context->decoder->encode_bitstream(context->decoder, context->target, + coded_buf->derived_surface.resource, &feedback); + context->decoder->end_frame(context->decoder, context->target, &context->desc.base); + context->decoder->flush(context->decoder); + context->decoder->get_feedback(context->decoder, feedback, &coded_size); + coded_buf->coded_size = coded_size; + } + else + context->decoder->end_frame(context->decoder, context->target, &context->desc.base); return VA_STATUS_SUCCESS; } diff --git a/src/gallium/state_trackers/va/va_private.h b/src/gallium/state_trackers/va/va_private.h index 1addd87303a..dba0d4a7251 100644 --- a/src/gallium/state_trackers/va/va_private.h +++ b/src/gallium/state_trackers/va/va_private.h @@ -242,6 +242,7 @@ typedef struct { } mpeg4; struct vl_deint_filter *deint; + struct vlVaBuffer *coded_buf; } vlVaContext; typedef struct { @@ -261,6 +262,7 @@ typedef struct { } derived_surface; unsigned int export_refcount; VABufferInfo export_state; + unsigned int coded_size; } vlVaBuffer; typedef struct { |