/************************************************************************** * * Copyright 2010 Thomas Balling Sørensen & Orasanu Lucian. * Copyright 2014 Advanced Micro Devices, Inc. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #include "pipe/p_video_codec.h" #include "util/u_handle_table.h" #include "util/u_video.h" #include "vl/vl_vlc.h" #include "vl/vl_winsys.h" #include "va_private.h" VAStatus vlVaBeginPicture(VADriverContextP ctx, VAContextID context_id, VASurfaceID render_target) { vlVaDriver *drv; vlVaContext *context; vlVaSurface *surf; if (!ctx) return VA_STATUS_ERROR_INVALID_CONTEXT; drv = VL_VA_DRIVER(ctx); if (!drv) return VA_STATUS_ERROR_INVALID_CONTEXT; context = handle_table_get(drv->htab, context_id); if (!context) return VA_STATUS_ERROR_INVALID_CONTEXT; surf = handle_table_get(drv->htab, render_target); if (!surf || !surf->buffer) return VA_STATUS_ERROR_INVALID_SURFACE; context->target = surf->buffer; if (!context->decoder) { /* VPP */ if ((context->target->buffer_format != PIPE_FORMAT_B8G8R8A8_UNORM && context->target->buffer_format != PIPE_FORMAT_R8G8B8A8_UNORM && context->target->buffer_format != PIPE_FORMAT_B8G8R8X8_UNORM && context->target->buffer_format != PIPE_FORMAT_R8G8B8X8_UNORM) || context->target->interlaced) return VA_STATUS_ERROR_UNIMPLEMENTED; return VA_STATUS_SUCCESS; } context->decoder->begin_frame(context->decoder, context->target, &context->desc.base); return VA_STATUS_SUCCESS; } static void getReferenceFrame(vlVaDriver *drv, VASurfaceID surface_id, struct pipe_video_buffer **ref_frame) { vlVaSurface *surf = handle_table_get(drv->htab, surface_id); if (surf) *ref_frame = surf->buffer; else *ref_frame = NULL; } static void handlePictureParameterBuffer(vlVaDriver *drv, vlVaContext *context, vlVaBuffer *buf) { VAPictureParameterBufferMPEG2 *mpeg2; VAPictureParameterBufferH264 *h264; VAPictureParameterBufferVC1 * vc1; VAPictureParameterBufferMPEG4 *mpeg4; VAPictureParameterBufferHEVC *hevc; vlVaSurface *surf_forward; vlVaSurface *surf_backward; unsigned int i; static const uint8_t default_intra_quant_matrix[64] = { 0 }; static const uint8_t default_non_intra_quant_matrix[64] = { 0 }; switch (u_reduce_video_profile(context->decoder->profile)) { case PIPE_VIDEO_FORMAT_MPEG12: assert(buf->size >= sizeof(VAPictureParameterBufferMPEG2) && buf->num_elements == 1); mpeg2 = buf->data; /*horizontal_size;*/ /*vertical_size;*/ getReferenceFrame(drv, mpeg2->forward_reference_picture, &context->desc.mpeg12.ref[0]); getReferenceFrame(drv, mpeg2->backward_reference_picture, &context->desc.mpeg12.ref[1]); context->desc.mpeg12.picture_coding_type = mpeg2->picture_coding_type; context->desc.mpeg12.f_code[0][0] = ((mpeg2->f_code >> 12) & 0xf) - 1; context->desc.mpeg12.f_code[0][1] = ((mpeg2->f_code >> 8) & 0xf) - 1; context->desc.mpeg12.f_code[1][0] = ((mpeg2->f_code >> 4) & 0xf) - 1; context->desc.mpeg12.f_code[1][1] = (mpeg2->f_code & 0xf) - 1; context->desc.mpeg12.intra_dc_precision = mpeg2->picture_coding_extension.bits.intra_dc_precision; context->desc.mpeg12.picture_structure = mpeg2->picture_coding_extension.bits.picture_structure; context->desc.mpeg12.top_field_first = mpeg2->picture_coding_extension.bits.top_field_first; context->desc.mpeg12.frame_pred_frame_dct = mpeg2->picture_coding_extension.bits.frame_pred_frame_dct; context->desc.mpeg12.concealment_motion_vectors = mpeg2->picture_coding_extension.bits.concealment_motion_vectors; context->desc.mpeg12.q_scale_type = mpeg2->picture_coding_extension.bits.q_scale_type; context->desc.mpeg12.intra_vlc_format = mpeg2->picture_coding_extension.bits.intra_vlc_format; context->desc.mpeg12.alternate_scan = mpeg2->picture_coding_extension.bits.alternate_scan; /*repeat_first_field*/ /*progressive_frame*/ /*is_first_field*/ break; case PIPE_VIDEO_FORMAT_MPEG4_AVC: assert(buf->size >= sizeof(VAPictureParameterBufferH264) && buf->num_elements == 1); h264 = buf->data; /*CurrPic*/ context->desc.h264.field_order_cnt[0] = h264->CurrPic.TopFieldOrderCnt; context->desc.h264.field_order_cnt[1] = h264->CurrPic.BottomFieldOrderCnt; /*ReferenceFrames[16]*/ /*picture_width_in_mbs_minus1*/ /*picture_height_in_mbs_minus1*/ /*bit_depth_luma_minus8*/ /*bit_depth_chroma_minus8*/ context->desc.h264.num_ref_frames = h264->num_ref_frames; /*chroma_format_idc*/ /*residual_colour_transform_flag*/ /*gaps_in_frame_num_value_allowed_flag*/ context->desc.h264.pps->sps->frame_mbs_only_flag = h264->seq_fields.bits.frame_mbs_only_flag; context->desc.h264.pps->sps->mb_adaptive_frame_field_flag = h264->seq_fields.bits.mb_adaptive_frame_field_flag; context->desc.h264.pps->sps->direct_8x8_inference_flag = h264->seq_fields.bits.direct_8x8_inference_flag; /*MinLumaBiPredSize8x8*/ context->desc.h264.pps->sps->log2_max_frame_num_minus4 = h264->seq_fields.bits.log2_max_frame_num_minus4; context->desc.h264.pps->sps->pic_order_cnt_type = h264->seq_fields.bits.pic_order_cnt_type; context->desc.h264.pps->sps->log2_max_pic_order_cnt_lsb_minus4 = h264->seq_fields.bits.log2_max_pic_order_cnt_lsb_minus4; context->desc.h264.pps->sps->delta_pic_order_always_zero_flag = h264->seq_fields.bits.delta_pic_order_always_zero_flag; /*num_slice_groups_minus1*/ /*slice_group_map_type*/ /*slice_group_change_rate_minus1*/ context->desc.h264.pps->pic_init_qp_minus26 = h264->pic_init_qp_minus26; /*pic_init_qs_minus26*/ context->desc.h264.pps->chroma_qp_index_offset = h264->chroma_qp_index_offset; context->desc.h264.pps->second_chroma_qp_index_offset = h264->second_chroma_qp_index_offset; context->desc.h264.pps->entropy_coding_mode_flag = h264->pic_fields.bits.entropy_coding_mode_flag; context->desc.h264.pps->weighted_pred_flag = h264->pic_fields.bits.weighted_pred_flag; context->desc.h264.pps->weighted_bipred_idc = h264->pic_fields.bits.weighted_bipred_idc; context->desc.h264.pps->transform_8x8_mode_flag = h264->pic_fields.bits.transform_8x8_mode_flag; context->desc.h264.field_pic_flag = h264->pic_fields.bits.field_pic_flag; context->desc.h264.pps->constrained_intra_pred_flag = h264->pic_fields.bits.constrained_intra_pred_flag; context->desc.h264.pps->bottom_field_pic_order_in_frame_present_flag = h264->pic_fields.bits.pic_order_present_flag; context->desc.h264.pps->deblocking_filter_control_present_flag = h264->pic_fields.bits.deblocking_filter_control_present_flag; context->desc.h264.pps->redundant_pic_cnt_present_flag = h264->pic_fields.bits.redundant_pic_cnt_present_flag; /*reference_pic_flag*/ context->desc.h264.frame_num = h264->frame_num; break; case PIPE_VIDEO_FORMAT_VC1: assert(buf->size >= sizeof(VAPictureParameterBufferVC1) && buf->num_elements == 1); vc1 = buf->data; getReferenceFrame(drv, vc1->forward_reference_picture, &context->desc.vc1.ref[0]); getReferenceFrame(drv, vc1->backward_reference_picture, &context->desc.vc1.ref[1]); context->desc.vc1.picture_type = vc1->picture_fields.bits.picture_type; context->desc.vc1.frame_coding_mode = vc1->picture_fields.bits.frame_coding_mode; context->desc.vc1.postprocflag = vc1->post_processing != 0; context->desc.vc1.pulldown = vc1->sequence_fields.bits.pulldown; context->desc.vc1.interlace = vc1->sequence_fields.bits.interlace; context->desc.vc1.tfcntrflag = vc1->sequence_fields.bits.tfcntrflag; context->desc.vc1.finterpflag = vc1->sequence_fields.bits.finterpflag; context->desc.vc1.psf = vc1->sequence_fields.bits.psf; context->desc.vc1.dquant = vc1->pic_quantizer_fields.bits.dquant; context->desc.vc1.panscan_flag = vc1->entrypoint_fields.bits.panscan_flag; context->desc.vc1.refdist_flag = vc1->reference_fields.bits.reference_distance_flag; context->desc.vc1.quantizer = vc1->pic_quantizer_fields.bits.quantizer; context->desc.vc1.extended_mv = vc1->mv_fields.bits.extended_mv_flag; context->desc.vc1.extended_dmv = vc1->mv_fields.bits.extended_dmv_flag; context->desc.vc1.overlap = vc1->sequence_fields.bits.overlap; context->desc.vc1.vstransform = vc1->transform_fields.bits.variable_sized_transform_flag; context->desc.vc1.loopfilter = vc1->entrypoint_fields.bits.loopfilter; context->desc.vc1.fastuvmc = vc1->fast_uvmc_flag; context->desc.vc1.range_mapy_flag = vc1->range_mapping_fields.bits.luma_flag; context->desc.vc1.range_mapy = vc1->range_mapping_fields.bits.luma; context->desc.vc1.range_mapuv_flag = vc1->range_mapping_fields.bits.chroma_flag; context->desc.vc1.range_mapuv = vc1->range_mapping_fields.bits.chroma; context->desc.vc1.multires = vc1->sequence_fields.bits.multires; context->desc.vc1.syncmarker = vc1->sequence_fields.bits.syncmarker; context->desc.vc1.rangered = vc1->sequence_fields.bits.rangered; context->desc.vc1.maxbframes = vc1->sequence_fields.bits.max_b_frames; context->desc.vc1.deblockEnable = vc1->post_processing != 0; context->desc.vc1.pquant = vc1->pic_quantizer_fields.bits.pic_quantizer_scale; break; case PIPE_VIDEO_FORMAT_MPEG4: assert(buf->size >= sizeof(VAPictureParameterBufferMPEG4) && buf->num_elements == 1); mpeg4 = buf->data; context->mpeg4.pps = *mpeg4; /* vop_width */ /* vop_height */ /* forward_reference_picture */ /* backward_reference_picture */ context->desc.mpeg4.short_video_header = mpeg4->vol_fields.bits.short_video_header; /* chroma_format */ context->desc.mpeg4.interlaced = mpeg4->vol_fields.bits.interlaced; /* obmc_disable */ /* sprite_enable */ /* sprite_warping_accuracy */ context->desc.mpeg4.quant_type = mpeg4->vol_fields.bits.quant_type; context->desc.mpeg4.quarter_sample = mpeg4->vol_fields.bits.quarter_sample; /* data_partitioned */ /* reversible_vlc */ context->desc.mpeg4.resync_marker_disable = mpeg4->vol_fields.bits.resync_marker_disable; /* no_of_sprite_warping_points */ /* sprite_trajectory_du */ /* sprite_trajectory_dv */ /* quant_precision */ context->desc.mpeg4.vop_coding_type = mpeg4->vop_fields.bits.vop_coding_type; /* backward_reference_vop_coding_type */ /* vop_rounding_type */ /* intra_dc_vlc_thr */ context->desc.mpeg4.top_field_first = mpeg4->vop_fields.bits.top_field_first; context->desc.mpeg4.alternate_vertical_scan_flag = mpeg4->vop_fields.bits.alternate_vertical_scan_flag; context->desc.mpeg4.vop_fcode_forward = mpeg4->vop_fcode_forward; context->desc.mpeg4.vop_fcode_backward = mpeg4->vop_fcode_backward; context->desc.mpeg4.vop_time_increment_resolution = mpeg4->vop_time_increment_resolution; /* num_gobs_in_vop */ /* num_macroblocks_in_gob */ context->desc.mpeg4.trb[0] = mpeg4->TRB; context->desc.mpeg4.trb[1] = mpeg4->TRB; context->desc.mpeg4.trd[0] = mpeg4->TRD; context->desc.mpeg4.trd[1] = mpeg4->TRD; /* default [non-]intra quant matrix because mpv does not set these matrices */ if (!context->desc.mpeg4.intra_matrix) context->desc.mpeg4.intra_matrix = default_intra_quant_matrix; if (!context->desc.mpeg4.non_intra_matrix) context->desc.mpeg4.non_intra_matrix = default_non_intra_quant_matrix; surf_forward = handle_table_get(drv->htab, mpeg4->forward_reference_picture); if (surf_forward) context->desc.mpeg4.ref[0] = surf_forward->buffer; surf_backward = handle_table_get(drv->htab, mpeg4->backward_reference_picture); if (surf_backward) context->desc.mpeg4.ref[1] = surf_backward->buffer; context->mpeg4.vti_bits = 0; for (i = context->desc.mpeg4.vop_time_increment_resolution; i > 0; i /= 2) ++context->mpeg4.vti_bits; break; case PIPE_VIDEO_FORMAT_HEVC: assert(buf->size >= sizeof(VAPictureParameterBufferHEVC) && buf->num_elements == 1); hevc = buf->data; context->desc.h265.pps->sps->chroma_format_idc = hevc->pic_fields.bits.chroma_format_idc; context->desc.h265.pps->sps->separate_colour_plane_flag = hevc->pic_fields.bits.separate_colour_plane_flag; context->desc.h265.pps->sps->pic_width_in_luma_samples = hevc->pic_width_in_luma_samples; context->desc.h265.pps->sps->pic_height_in_luma_samples = hevc->pic_height_in_luma_samples; context->desc.h265.pps->sps->bit_depth_luma_minus8 = hevc->bit_depth_luma_minus8; context->desc.h265.pps->sps->bit_depth_chroma_minus8 = hevc->bit_depth_chroma_minus8; context->desc.h265.pps->sps->log2_max_pic_order_cnt_lsb_minus4 = hevc->log2_max_pic_order_cnt_lsb_minus4; context->desc.h265.pps->sps->sps_max_dec_pic_buffering_minus1 = hevc->sps_max_dec_pic_buffering_minus1; context->desc.h265.pps->sps->log2_min_luma_coding_block_size_minus3 = hevc->log2_min_luma_coding_block_size_minus3; context->desc.h265.pps->sps->log2_diff_max_min_luma_coding_block_size = hevc->log2_diff_max_min_luma_coding_block_size; context->desc.h265.pps->sps->log2_min_transform_block_size_minus2 = hevc->log2_min_transform_block_size_minus2; context->desc.h265.pps->sps->log2_diff_max_min_transform_block_size = hevc->log2_diff_max_min_transform_block_size; context->desc.h265.pps->sps->max_transform_hierarchy_depth_inter = hevc->max_transform_hierarchy_depth_inter; context->desc.h265.pps->sps->max_transform_hierarchy_depth_intra = hevc->max_transform_hierarchy_depth_intra; context->desc.h265.pps->sps->scaling_list_enabled_flag = hevc->pic_fields.bits.scaling_list_enabled_flag; context->desc.h265.pps->sps->amp_enabled_flag = hevc->pic_fields.bits.amp_enabled_flag; context->desc.h265.pps->sps->sample_adaptive_offset_enabled_flag = hevc->slice_parsing_fields.bits.sample_adaptive_offset_enabled_flag; context->desc.h265.pps->sps->pcm_enabled_flag = hevc->pic_fields.bits.pcm_enabled_flag; if (hevc->pic_fields.bits.pcm_enabled_flag == 1) { context->desc.h265.pps->sps->pcm_sample_bit_depth_luma_minus1 = hevc->pcm_sample_bit_depth_luma_minus1; context->desc.h265.pps->sps->pcm_sample_bit_depth_chroma_minus1 = hevc->pcm_sample_bit_depth_chroma_minus1; context->desc.h265.pps->sps->log2_min_pcm_luma_coding_block_size_minus3 = hevc->log2_min_pcm_luma_coding_block_size_minus3; context->desc.h265.pps->sps->log2_diff_max_min_pcm_luma_coding_block_size = hevc->log2_diff_max_min_pcm_luma_coding_block_size; context->desc.h265.pps->sps->pcm_loop_filter_disabled_flag = hevc->pic_fields.bits.pcm_loop_filter_disabled_flag; } context->desc.h265.pps->sps->num_short_term_ref_pic_sets = hevc->num_short_term_ref_pic_sets; context->desc.h265.pps->sps->long_term_ref_pics_present_flag = hevc->slice_parsing_fields.bits.long_term_ref_pics_present_flag; context->desc.h265.pps->sps->num_long_term_ref_pics_sps = hevc->num_long_term_ref_pic_sps; context->desc.h265.pps->sps->sps_temporal_mvp_enabled_flag = hevc->slice_parsing_fields.bits.sps_temporal_mvp_enabled_flag; context->desc.h265.pps->sps->strong_intra_smoothing_enabled_flag = hevc->pic_fields.bits.strong_intra_smoothing_enabled_flag; context->desc.h265.pps->dependent_slice_segments_enabled_flag = hevc->slice_parsing_fields.bits.dependent_slice_segments_enabled_flag; context->desc.h265.pps->output_flag_present_flag = hevc->slice_parsing_fields.bits.output_flag_present_flag; context->desc.h265.pps->num_extra_slice_header_bits = hevc->num_extra_slice_header_bits; context->desc.h265.pps->sign_data_hiding_enabled_flag = hevc->pic_fields.bits.sign_data_hiding_enabled_flag; context->desc.h265.pps->cabac_init_present_flag = hevc->slice_parsing_fields.bits.cabac_init_present_flag; context->desc.h265.pps->num_ref_idx_l0_default_active_minus1 = hevc->num_ref_idx_l0_default_active_minus1; context->desc.h265.pps->num_ref_idx_l1_default_active_minus1 = hevc->num_ref_idx_l1_default_active_minus1; context->desc.h265.pps->init_qp_minus26 = hevc->init_qp_minus26; context->desc.h265.pps->constrained_intra_pred_flag = hevc->pic_fields.bits.constrained_intra_pred_flag; context->desc.h265.pps->transform_skip_enabled_flag = hevc->pic_fields.bits.transform_skip_enabled_flag; context->desc.h265.pps->cu_qp_delta_enabled_flag = hevc->pic_fields.bits.cu_qp_delta_enabled_flag; context->desc.h265.pps->diff_cu_qp_delta_depth = hevc->diff_cu_qp_delta_depth; context->desc.h265.pps->pps_cb_qp_offset = hevc->pps_cb_qp_offset; context->desc.h265.pps->pps_cr_qp_offset = hevc->pps_cr_qp_offset; context->desc.h265.pps->pps_slice_chroma_qp_offsets_present_flag = hevc->slice_parsing_fields.bits.pps_slice_chroma_qp_offsets_present_flag; context->desc.h265.pps->weighted_pred_flag = hevc->pic_fields.bits.weighted_pred_flag; context->desc.h265.pps->weighted_bipred_flag = hevc->pic_fields.bits.weighted_bipred_flag; context->desc.h265.pps->transquant_bypass_enabled_flag = hevc->pic_fields.bits.transquant_bypass_enabled_flag; context->desc.h265.pps->tiles_enabled_flag = hevc->pic_fields.bits.tiles_enabled_flag; context->desc.h265.pps->entropy_coding_sync_enabled_flag = hevc->pic_fields.bits.entropy_coding_sync_enabled_flag; if (hevc->pic_fields.bits.tiles_enabled_flag == 1) { context->desc.h265.pps->num_tile_columns_minus1 = hevc->num_tile_columns_minus1; context->desc.h265.pps->num_tile_rows_minus1 = hevc->num_tile_rows_minus1; for (i = 0 ; i < 19 ; i++) context->desc.h265.pps->column_width_minus1[i] = hevc->column_width_minus1[i]; for (i = 0 ; i < 21 ; i++) context->desc.h265.pps->row_height_minus1[i] = hevc->row_height_minus1[i]; context->desc.h265.pps->loop_filter_across_tiles_enabled_flag = hevc->pic_fields.bits.loop_filter_across_tiles_enabled_flag; } context->desc.h265.pps->pps_loop_filter_across_slices_enabled_flag = hevc->pic_fields.bits.pps_loop_filter_across_slices_enabled_flag; context->desc.h265.pps->deblocking_filter_override_enabled_flag = hevc->slice_parsing_fields.bits.deblocking_filter_override_enabled_flag; context->desc.h265.pps->pps_deblocking_filter_disabled_flag = hevc->slice_parsing_fields.bits.pps_disable_deblocking_filter_flag; context->desc.h265.pps->pps_beta_offset_div2 = hevc->pps_beta_offset_div2; context->desc.h265.pps->pps_tc_offset_div2 = hevc->pps_tc_offset_div2; context->desc.h265.pps->lists_modification_present_flag = hevc->slice_parsing_fields.bits.lists_modification_present_flag; context->desc.h265.pps->log2_parallel_merge_level_minus2 = hevc->log2_parallel_merge_level_minus2; context->desc.h265.pps->slice_segment_header_extension_present_flag = hevc->slice_parsing_fields.bits.slice_segment_header_extension_present_flag; context->desc.h265.IDRPicFlag = hevc->slice_parsing_fields.bits.IdrPicFlag; context->desc.h265.RAPPicFlag = hevc->slice_parsing_fields.bits.RapPicFlag; context->desc.h265.CurrPicOrderCntVal = hevc->CurrPic.pic_order_cnt; for (i = 0 ; i < 8 ; i++) { context->desc.h265.RefPicSetStCurrBefore[i] = 0xFF; context->desc.h265.RefPicSetStCurrAfter[i] = 0xFF; context->desc.h265.RefPicSetLtCurr[i] = 0xFF; } context->desc.h265.NumPocStCurrBefore = 0; context->desc.h265.NumPocStCurrAfter = 0; context->desc.h265.NumPocLtCurr = 0; unsigned int iBefore = 0; unsigned int iAfter = 0; unsigned int iCurr = 0; for (i = 0 ; i < 15 ; i++) { context->desc.h265.PicOrderCntVal[i] = hevc->ReferenceFrames[i].pic_order_cnt; unsigned int index = hevc->ReferenceFrames[i].picture_id & 0x7F; if (index == 0x7F) continue; getReferenceFrame(drv, hevc->ReferenceFrames[i].picture_id, &context->desc.h265.ref[i]); if ((hevc->ReferenceFrames[i].flags & VA_PICTURE_HEVC_RPS_ST_CURR_BEFORE) && (iBefore < 8)) { context->desc.h265.RefPicSetStCurrBefore[iBefore++] = i; context->desc.h265.NumPocStCurrBefore++; } if ((hevc->ReferenceFrames[i].flags & VA_PICTURE_HEVC_RPS_ST_CURR_AFTER) && (iAfter < 8)) { context->desc.h265.RefPicSetStCurrAfter[iAfter++] = i; context->desc.h265.NumPocStCurrAfter++; } if ((hevc->ReferenceFrames[i].flags & VA_PICTURE_HEVC_RPS_LT_CURR) && (iCurr < 8)) { context->desc.h265.RefPicSetLtCurr[iCurr++] = i; context->desc.h265.NumPocLtCurr++; } } break; default: break; } } static void handleIQMatrixBuffer(vlVaContext *context, vlVaBuffer *buf) { VAIQMatrixBufferMPEG2 *mpeg2; VAIQMatrixBufferH264 *h264; VAIQMatrixBufferMPEG4 *mpeg4; VAIQMatrixBufferHEVC *h265; switch (u_reduce_video_profile(context->decoder->profile)) { case PIPE_VIDEO_FORMAT_MPEG12: assert(buf->size >= sizeof(VAIQMatrixBufferMPEG2) && buf->num_elements == 1); mpeg2 = buf->data; if (mpeg2->load_intra_quantiser_matrix) context->desc.mpeg12.intra_matrix = mpeg2->intra_quantiser_matrix; else context->desc.mpeg12.intra_matrix = NULL; if (mpeg2->load_non_intra_quantiser_matrix) context->desc.mpeg12.non_intra_matrix = mpeg2->non_intra_quantiser_matrix; else context->desc.mpeg12.non_intra_matrix = NULL; break; case PIPE_VIDEO_FORMAT_MPEG4_AVC: assert(buf->size >= sizeof(VAIQMatrixBufferH264) && buf->num_elements == 1); h264 = buf->data; memcpy(&context->desc.h264.pps->ScalingList4x4, h264->ScalingList4x4, 6 * 16); memcpy(&context->desc.h264.pps->ScalingList8x8, h264->ScalingList8x8, 2 * 64); break; case PIPE_VIDEO_FORMAT_HEVC: assert(buf->size >= sizeof(VAIQMatrixBufferH264) && buf->num_elements == 1); h265 = buf->data; memcpy(&context->desc.h265.pps->sps->ScalingList4x4, h265->ScalingList4x4, 6 * 16); memcpy(&context->desc.h265.pps->sps->ScalingList8x8, h265->ScalingList8x8, 6 * 64); memcpy(&context->desc.h265.pps->sps->ScalingList16x16, h265->ScalingList16x16, 6 * 64); memcpy(&context->desc.h265.pps->sps->ScalingList32x32, h265->ScalingList32x32, 2 * 64); memcpy(&context->desc.h265.pps->sps->ScalingListDCCoeff16x16, h265->ScalingListDC16x16, 6); memcpy(&context->desc.h265.pps->sps->ScalingListDCCoeff32x32, h265->ScalingListDC32x32, 2); break; case PIPE_VIDEO_FORMAT_MPEG4: assert(buf->size >= sizeof(VAIQMatrixBufferMPEG4) && buf->num_elements == 1); mpeg4 = buf->data; if (mpeg4->load_intra_quant_mat) context->desc.mpeg4.intra_matrix = mpeg4->intra_quant_mat; else context->desc.mpeg4.intra_matrix = NULL; if (mpeg4->load_non_intra_quant_mat) context->desc.mpeg4.non_intra_matrix = mpeg4->non_intra_quant_mat; else context->desc.mpeg4.non_intra_matrix = NULL; break; default: break; } } static void handleSliceParameterBuffer(vlVaContext *context, vlVaBuffer *buf) { VASliceParameterBufferH264 *h264; VASliceParameterBufferMPEG4 *mpeg4; VASliceParameterBufferHEVC *h265; switch (u_reduce_video_profile(context->decoder->profile)) { case PIPE_VIDEO_FORMAT_MPEG4_AVC: assert(buf->size >= sizeof(VASliceParameterBufferH264) && buf->num_elements == 1); h264 = buf->data; context->desc.h264.num_ref_idx_l0_active_minus1 = h264->num_ref_idx_l0_active_minus1; context->desc.h264.num_ref_idx_l1_active_minus1 = h264->num_ref_idx_l1_active_minus1; break; case PIPE_VIDEO_FORMAT_MPEG4: assert(buf->size >= sizeof(VASliceParameterBufferMPEG4) && buf->num_elements == 1); mpeg4 = buf->data; context->mpeg4.quant_scale = mpeg4->quant_scale; break; case PIPE_VIDEO_FORMAT_HEVC: assert(buf->size >= sizeof(VASliceParameterBufferHEVC) && buf->num_elements == 1); h265 = buf->data; for (int i = 0 ; i < 2 ; i++) { for (int j = 0 ; j < 15 ; j++) context->desc.h265.RefPicList[i][j] = h265->RefPicList[i][j]; } context->desc.h265.UseRefPicList = true; break; default: break; } } struct bit_stream { uint8_t *data; unsigned int length; /* bits */ unsigned int pos; /* bits */ }; static inline void write_bit(struct bit_stream *writer, unsigned int bit) { assert(writer->length > (writer)->pos); writer->data[writer->pos>>3] |= ((bit & 1)<<(7 - (writer->pos & 7))); writer->pos++; } static inline void write_bits(struct bit_stream *writer, unsigned int bits, unsigned int len) { int i; assert(len <= sizeof(bits)*8); for (i = len - 1; i >= 0; i--) write_bit(writer, bits>>i); } static void vlVaDecoderFixMPEG4Startcode(vlVaContext *context) { uint8_t vop[] = { 0x00, 0x00, 0x01, 0xb6, 0x00, 0x00, 0x00, 0x00, 0x00 }; struct bit_stream bs_vop = {vop, sizeof(vop)*8, 32}; unsigned int vop_time_inc; int mod_time; unsigned int vop_size; unsigned int vop_coding_type = context->desc.mpeg4.vop_coding_type; context->mpeg4.start_code_size = 0; memset(context->mpeg4.start_code, 0, sizeof(context->mpeg4.start_code)); if (vop_coding_type+1 == PIPE_MPEG12_PICTURE_CODING_TYPE_I) { unsigned int vop_time = context->mpeg4.frame_num/ context->desc.mpeg4.vop_time_increment_resolution; unsigned int vop_hour = vop_time / 3600; unsigned int vop_minute = (vop_time / 60) % 60; unsigned int vop_second = vop_time % 60; uint8_t group_of_vop[] = { 0x00, 0x00, 0x01, 0xb3, 0x00, 0x00, 0x00 }; struct bit_stream bs_gvop = {group_of_vop, sizeof(group_of_vop)*8, 32}; write_bits(&bs_gvop, vop_hour, 5); write_bits(&bs_gvop, vop_minute, 6); write_bit(&bs_gvop, 1); /* marker_bit */ write_bits(&bs_gvop, vop_second, 6); write_bit(&bs_gvop, 0); /* closed_gov */ /* TODO replace magic */ write_bit(&bs_gvop, 0); /* broken_link */ write_bit(&bs_gvop, 0); /* padding */ write_bits(&bs_gvop, 7, 3); /* padding */ memcpy(context->mpeg4.start_code, group_of_vop, sizeof(group_of_vop)); context->mpeg4.start_code_size += sizeof(group_of_vop); } write_bits(&bs_vop, vop_coding_type, 2); mod_time = context->mpeg4.frame_num % context->desc.mpeg4.vop_time_increment_resolution == 0 && vop_coding_type+1 != PIPE_MPEG12_PICTURE_CODING_TYPE_I; while (mod_time--) write_bit(&bs_vop, 1); /* modulo_time_base */ write_bit(&bs_vop, 0); /* modulo_time_base */ write_bit(&bs_vop, 1); /* marker_bit */ vop_time_inc = context->mpeg4.frame_num % context->desc.mpeg4.vop_time_increment_resolution; write_bits(&bs_vop, vop_time_inc, context->mpeg4.vti_bits); write_bit(&bs_vop, 1); /* marker_bit */ write_bit(&bs_vop, 1); /* vop_coded */ if (vop_coding_type+1 == PIPE_MPEG12_PICTURE_CODING_TYPE_P) write_bit(&bs_vop, context->mpeg4.pps.vop_fields.bits.vop_rounding_type); write_bits(&bs_vop, context->mpeg4.pps.vop_fields.bits.intra_dc_vlc_thr, 3); if (context->mpeg4.pps.vol_fields.bits.interlaced) { write_bit(&bs_vop, context->mpeg4.pps.vop_fields.bits.top_field_first); write_bit(&bs_vop, context->mpeg4.pps.vop_fields.bits.alternate_vertical_scan_flag); } write_bits(&bs_vop, context->mpeg4.quant_scale, context->mpeg4.pps.quant_precision); if (vop_coding_type+1 != PIPE_MPEG12_PICTURE_CODING_TYPE_I) write_bits(&bs_vop, context->desc.mpeg4.vop_fcode_forward, 3); if (vop_coding_type+1 == PIPE_MPEG12_PICTURE_CODING_TYPE_B) write_bits(&bs_vop, context->desc.mpeg4.vop_fcode_backward, 3); vop_size = bs_vop.pos/8; memcpy(context->mpeg4.start_code + context->mpeg4.start_code_size, vop, vop_size); context->mpeg4.start_code_size += vop_size; } static unsigned int bufHasStartcode(vlVaBuffer *buf, unsigned int code, unsigned int bits) { struct vl_vlc vlc = {0}; int i; /* search the first 64 bytes for a startcode */ vl_vlc_init(&vlc, 1, (const void * const*)&buf->data, &buf->size); for (i = 0; i < 64 && vl_vlc_bits_left(&vlc) >= bits; ++i) { if (vl_vlc_peekbits(&vlc, bits) == code) return 1; vl_vlc_eatbits(&vlc, 8); vl_vlc_fillbits(&vlc); } return 0; } static void handleVASliceDataBufferType(vlVaContext *context, vlVaBuffer *buf) { enum pipe_video_format format; unsigned num_buffers = 0; void * const *buffers[2]; unsigned sizes[2]; static const uint8_t start_code_h264[] = { 0x00, 0x00, 0x01 }; static const uint8_t start_code_h265[] = { 0x00, 0x00, 0x01 }; static const uint8_t start_code_vc1[] = { 0x00, 0x00, 0x01, 0x0d }; format = u_reduce_video_profile(context->decoder->profile); switch (format) { case PIPE_VIDEO_FORMAT_MPEG4_AVC: if (bufHasStartcode(buf, 0x000001, 24)) break; buffers[num_buffers] = (void *const)&start_code_h264; sizes[num_buffers++] = sizeof(start_code_h264); break; case PIPE_VIDEO_FORMAT_HEVC: if (bufHasStartcode(buf, 0x000001, 24)) break; buffers[num_buffers] = (void *const)&start_code_h265; sizes[num_buffers++] = sizeof(start_code_h265); break; case PIPE_VIDEO_FORMAT_VC1: if (bufHasStartcode(buf, 0x0000010d, 32) || bufHasStartcode(buf, 0x0000010c, 32) || bufHasStartcode(buf, 0x0000010b, 32)) break; if (context->decoder->profile == PIPE_VIDEO_PROFILE_VC1_ADVANCED) { buffers[num_buffers] = (void *const)&start_code_vc1; sizes[num_buffers++] = sizeof(start_code_vc1); } break; case PIPE_VIDEO_FORMAT_MPEG4: if (bufHasStartcode(buf, 0x000001, 24)) break; vlVaDecoderFixMPEG4Startcode(context); buffers[num_buffers] = (void *)context->mpeg4.start_code; sizes[num_buffers++] = context->mpeg4.start_code_size; default: break; } buffers[num_buffers] = buf->data; sizes[num_buffers] = buf->size; ++num_buffers; context->decoder->decode_bitstream(context->decoder, context->target, &context->desc.base, num_buffers, (const void * const*)buffers, sizes); } static VAStatus handleVAProcPipelineParameterBufferType(vlVaDriver *drv, vlVaContext *context, vlVaBuffer *buf) { struct u_rect src_rect; struct u_rect dst_rect; struct u_rect *dirty_area; vlVaSurface *src_surface; VAProcPipelineParameterBuffer *pipeline_param; struct pipe_surface **surfaces; struct pipe_screen *screen; struct pipe_surface *psurf; if (!drv || !context) return VA_STATUS_ERROR_INVALID_CONTEXT; if (!buf || !buf->data) return VA_STATUS_ERROR_INVALID_BUFFER; if (!context->target) return VA_STATUS_ERROR_INVALID_SURFACE; pipeline_param = (VAProcPipelineParameterBuffer *)buf->data; src_surface = handle_table_get(drv->htab, pipeline_param->surface); if (!src_surface || !src_surface->buffer) return VA_STATUS_ERROR_INVALID_SURFACE; surfaces = context->target->get_surfaces(context->target); if (!surfaces || !surfaces[0]) return VA_STATUS_ERROR_INVALID_SURFACE; screen = drv->pipe->screen; psurf = surfaces[0]; src_rect.x0 = pipeline_param->surface_region->x; src_rect.y0 = pipeline_param->surface_region->y; src_rect.x1 = pipeline_param->surface_region->x + pipeline_param->surface_region->width; src_rect.y1 = pipeline_param->surface_region->y + pipeline_param->surface_region->height; dst_rect.x0 = pipeline_param->output_region->x; dst_rect.y0 = pipeline_param->output_region->y; dst_rect.x1 = pipeline_param->output_region->x + pipeline_param->output_region->width; dst_rect.y1 = pipeline_param->output_region->y + pipeline_param->output_region->height; dirty_area = vl_screen_get_dirty_area(drv->vscreen); vl_compositor_clear_layers(&drv->cstate); vl_compositor_set_buffer_layer(&drv->cstate, &drv->compositor, 0, src_surface->buffer, &src_rect, NULL, VL_COMPOSITOR_WEAVE); vl_compositor_set_layer_dst_area(&drv->cstate, 0, &dst_rect); vl_compositor_render(&drv->cstate, &drv->compositor, psurf, dirty_area, true); screen->fence_reference(screen, &src_surface->fence, NULL); drv->pipe->flush(drv->pipe, &src_surface->fence, 0); return VA_STATUS_SUCCESS; } VAStatus vlVaRenderPicture(VADriverContextP ctx, VAContextID context_id, VABufferID *buffers, int num_buffers) { vlVaDriver *drv; vlVaContext *context; VAStatus vaStatus = VA_STATUS_SUCCESS; unsigned i; if (!ctx) return VA_STATUS_ERROR_INVALID_CONTEXT; drv = VL_VA_DRIVER(ctx); if (!drv) return VA_STATUS_ERROR_INVALID_CONTEXT; context = handle_table_get(drv->htab, context_id); if (!context) return VA_STATUS_ERROR_INVALID_CONTEXT; for (i = 0; i < num_buffers; ++i) { vlVaBuffer *buf = handle_table_get(drv->htab, buffers[i]); if (!buf) return VA_STATUS_ERROR_INVALID_BUFFER; switch (buf->type) { case VAPictureParameterBufferType: handlePictureParameterBuffer(drv, context, buf); break; case VAIQMatrixBufferType: handleIQMatrixBuffer(context, buf); break; case VASliceParameterBufferType: handleSliceParameterBuffer(context, buf); break; case VASliceDataBufferType: handleVASliceDataBufferType(context, buf); break; case VAProcPipelineParameterBufferType: vaStatus = handleVAProcPipelineParameterBufferType(drv, context, buf); break; default: break; } } return vaStatus; } VAStatus vlVaEndPicture(VADriverContextP ctx, VAContextID context_id) { vlVaDriver *drv; vlVaContext *context; if (!ctx) return VA_STATUS_ERROR_INVALID_CONTEXT; drv = VL_VA_DRIVER(ctx); if (!drv) return VA_STATUS_ERROR_INVALID_CONTEXT; context = handle_table_get(drv->htab, context_id); if (!context) return VA_STATUS_ERROR_INVALID_CONTEXT; if (!context->decoder) { /* VPP */ return VA_STATUS_SUCCESS; } context->mpeg4.frame_num++; context->decoder->end_frame(context->decoder, context->target, &context->desc.base); return VA_STATUS_SUCCESS; }