/************************************************************************** * * Copyright 2009 Younes Manton. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #include #include #include "util/u_memory.h" #include "util/u_rect.h" #include "util/u_sampler.h" #include "util/u_video.h" #include "vl_mpeg12_decoder.h" #include "vl_defines.h" #define SCALE_FACTOR_SNORM (32768.0f / 256.0f) #define SCALE_FACTOR_SSCALED (1.0f / 256.0f) struct format_config { enum pipe_format zscan_source_format; enum pipe_format idct_source_format; enum pipe_format mc_source_format; float idct_scale; float mc_scale; }; static const struct format_config bitstream_format_config[] = { // { PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, PIPE_FORMAT_R16G16B16A16_FLOAT, 1.0f, SCALE_FACTOR_SSCALED }, // { PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, 1.0f, SCALE_FACTOR_SSCALED }, { PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, PIPE_FORMAT_R16G16B16A16_FLOAT, 1.0f, SCALE_FACTOR_SNORM }, { PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, 1.0f, SCALE_FACTOR_SNORM } }; static const unsigned num_bitstream_format_configs = sizeof(bitstream_format_config) / sizeof(struct format_config); static const struct format_config idct_format_config[] = { // { PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, PIPE_FORMAT_R16G16B16A16_FLOAT, 1.0f, SCALE_FACTOR_SSCALED }, // { PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, 1.0f, SCALE_FACTOR_SSCALED }, { PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, PIPE_FORMAT_R16G16B16A16_FLOAT, 1.0f, SCALE_FACTOR_SNORM }, { PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, PIPE_FORMAT_R16G16B16A16_SNORM, 1.0f, SCALE_FACTOR_SNORM } }; static const unsigned num_idct_format_configs = sizeof(idct_format_config) / sizeof(struct format_config); static const struct format_config mc_format_config[] = { //{ PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_NONE, PIPE_FORMAT_R16_SSCALED, 0.0f, SCALE_FACTOR_SSCALED }, { PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_NONE, PIPE_FORMAT_R16_SNORM, 0.0f, SCALE_FACTOR_SNORM } }; static const unsigned num_mc_format_configs = sizeof(mc_format_config) / sizeof(struct format_config); static const unsigned const_empty_block_mask_420[3][2][2] = { { { 0x20, 0x10 }, { 0x08, 0x04 } }, { { 0x02, 0x02 }, { 0x02, 0x02 } }, { { 0x01, 0x01 }, { 0x01, 0x01 } } }; static bool init_zscan_buffer(struct vl_mpeg12_decoder *dec, struct vl_mpeg12_buffer *buffer) { struct pipe_resource *res, res_tmpl; struct pipe_sampler_view sv_tmpl; struct pipe_surface **destination; unsigned i; assert(dec && buffer); memset(&res_tmpl, 0, sizeof(res_tmpl)); res_tmpl.target = PIPE_TEXTURE_2D; res_tmpl.format = dec->zscan_source_format; res_tmpl.width0 = dec->blocks_per_line * BLOCK_WIDTH * BLOCK_HEIGHT; res_tmpl.height0 = align(dec->num_blocks, dec->blocks_per_line) / dec->blocks_per_line; res_tmpl.depth0 = 1; res_tmpl.array_size = 1; res_tmpl.usage = PIPE_USAGE_STREAM; res_tmpl.bind = PIPE_BIND_SAMPLER_VIEW; res = dec->base.context->screen->resource_create(dec->base.context->screen, &res_tmpl); if (!res) goto error_source; memset(&sv_tmpl, 0, sizeof(sv_tmpl)); u_sampler_view_default_template(&sv_tmpl, res, res->format); sv_tmpl.swizzle_r = sv_tmpl.swizzle_g = sv_tmpl.swizzle_b = sv_tmpl.swizzle_a = PIPE_SWIZZLE_RED; buffer->zscan_source = dec->base.context->create_sampler_view(dec->base.context, res, &sv_tmpl); pipe_resource_reference(&res, NULL); if (!buffer->zscan_source) goto error_sampler; if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) destination = dec->idct_source->get_surfaces(dec->idct_source); else destination = dec->mc_source->get_surfaces(dec->mc_source); if (!destination) goto error_surface; for (i = 0; i < VL_MAX_PLANES; ++i) if (!vl_zscan_init_buffer(i == 0 ? &dec->zscan_y : &dec->zscan_c, &buffer->zscan[i], buffer->zscan_source, destination[i])) goto error_plane; return true; error_plane: for (; i > 0; --i) vl_zscan_cleanup_buffer(&buffer->zscan[i - 1]); error_surface: error_sampler: pipe_sampler_view_reference(&buffer->zscan_source, NULL); error_source: return false; } static void cleanup_zscan_buffer(struct vl_mpeg12_buffer *buffer) { unsigned i; assert(buffer); for (i = 0; i < VL_MAX_PLANES; ++i) vl_zscan_cleanup_buffer(&buffer->zscan[i]); pipe_sampler_view_reference(&buffer->zscan_source, NULL); } static bool init_idct_buffer(struct vl_mpeg12_decoder *dec, struct vl_mpeg12_buffer *buffer) { struct pipe_sampler_view **idct_source_sv, **mc_source_sv; unsigned i; assert(dec && buffer); idct_source_sv = dec->idct_source->get_sampler_view_planes(dec->idct_source); if (!idct_source_sv) goto error_source_sv; mc_source_sv = dec->mc_source->get_sampler_view_planes(dec->mc_source); if (!mc_source_sv) goto error_mc_source_sv; for (i = 0; i < 3; ++i) if (!vl_idct_init_buffer(i == 0 ? &dec->idct_y : &dec->idct_c, &buffer->idct[i], idct_source_sv[i], mc_source_sv[i])) goto error_plane; return true; error_plane: for (; i > 0; --i) vl_idct_cleanup_buffer(&buffer->idct[i - 1]); error_mc_source_sv: error_source_sv: return false; } static void cleanup_idct_buffer(struct vl_mpeg12_buffer *buf) { unsigned i; assert(buf); for (i = 0; i < 3; ++i) vl_idct_cleanup_buffer(&buf->idct[0]); } static bool init_mc_buffer(struct vl_mpeg12_decoder *dec, struct vl_mpeg12_buffer *buf) { assert(dec && buf); if(!vl_mc_init_buffer(&dec->mc_y, &buf->mc[0])) goto error_mc_y; if(!vl_mc_init_buffer(&dec->mc_c, &buf->mc[1])) goto error_mc_cb; if(!vl_mc_init_buffer(&dec->mc_c, &buf->mc[2])) goto error_mc_cr; return true; error_mc_cr: vl_mc_cleanup_buffer(&buf->mc[1]); error_mc_cb: vl_mc_cleanup_buffer(&buf->mc[0]); error_mc_y: return false; } static void cleanup_mc_buffer(struct vl_mpeg12_buffer *buf) { unsigned i; assert(buf); for (i = 0; i < VL_MAX_PLANES; ++i) vl_mc_cleanup_buffer(&buf->mc[i]); } static INLINE void MacroBlockTypeToPipeWeights(const struct pipe_mpeg12_macroblock *mb, unsigned weights[2]) { assert(mb); switch (mb->macroblock_type & (PIPE_MPEG12_MB_TYPE_MOTION_FORWARD | PIPE_MPEG12_MB_TYPE_MOTION_BACKWARD)) { case PIPE_MPEG12_MB_TYPE_MOTION_FORWARD: weights[0] = PIPE_VIDEO_MV_WEIGHT_MAX; weights[1] = PIPE_VIDEO_MV_WEIGHT_MIN; break; case (PIPE_MPEG12_MB_TYPE_MOTION_FORWARD | PIPE_MPEG12_MB_TYPE_MOTION_BACKWARD): weights[0] = PIPE_VIDEO_MV_WEIGHT_HALF; weights[1] = PIPE_VIDEO_MV_WEIGHT_HALF; break; case PIPE_MPEG12_MB_TYPE_MOTION_BACKWARD: weights[0] = PIPE_VIDEO_MV_WEIGHT_MIN; weights[1] = PIPE_VIDEO_MV_WEIGHT_MAX; break; default: if (mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA) { weights[0] = PIPE_VIDEO_MV_WEIGHT_MIN; weights[1] = PIPE_VIDEO_MV_WEIGHT_MIN; } else { /* no motion vector, but also not intra mb -> just copy the old frame content */ weights[0] = PIPE_VIDEO_MV_WEIGHT_MAX; weights[1] = PIPE_VIDEO_MV_WEIGHT_MIN; } break; } } static INLINE struct vl_motionvector MotionVectorToPipe(const struct pipe_mpeg12_macroblock *mb, unsigned vector, unsigned field_select_mask, unsigned weight) { struct vl_motionvector mv; assert(mb); if (mb->macroblock_type & (PIPE_MPEG12_MB_TYPE_MOTION_FORWARD | PIPE_MPEG12_MB_TYPE_MOTION_BACKWARD)) { switch (mb->macroblock_modes.bits.frame_motion_type) { case PIPE_MPEG12_MO_TYPE_FRAME: mv.top.x = mb->PMV[0][vector][0]; mv.top.y = mb->PMV[0][vector][1]; mv.top.field_select = PIPE_VIDEO_FRAME; mv.top.weight = weight; mv.bottom.x = mb->PMV[0][vector][0]; mv.bottom.y = mb->PMV[0][vector][1]; mv.bottom.weight = weight; mv.bottom.field_select = PIPE_VIDEO_FRAME; break; case PIPE_MPEG12_MO_TYPE_FIELD: mv.top.x = mb->PMV[0][vector][0]; mv.top.y = mb->PMV[0][vector][1]; mv.top.field_select = (mb->motion_vertical_field_select & field_select_mask) ? PIPE_VIDEO_BOTTOM_FIELD : PIPE_VIDEO_TOP_FIELD; mv.top.weight = weight; mv.bottom.x = mb->PMV[1][vector][0]; mv.bottom.y = mb->PMV[1][vector][1]; mv.bottom.field_select = (mb->motion_vertical_field_select & (field_select_mask << 2)) ? PIPE_VIDEO_BOTTOM_FIELD : PIPE_VIDEO_TOP_FIELD; mv.bottom.weight = weight; break; default: // TODO: Support DUALPRIME and 16x8 break; } } else { mv.top.x = mv.top.y = 0; mv.top.field_select = PIPE_VIDEO_FRAME; mv.top.weight = weight; mv.bottom.x = mv.bottom.y = 0; mv.bottom.field_select = PIPE_VIDEO_FRAME; mv.bottom.weight = weight; } return mv; } static INLINE void UploadYcbcrBlocks(struct vl_mpeg12_decoder *dec, struct vl_mpeg12_buffer *buf, const struct pipe_mpeg12_macroblock *mb) { unsigned intra; unsigned tb, x, y, num_blocks = 0; assert(dec && buf); assert(mb); if (!mb->coded_block_pattern) return; intra = mb->macroblock_type & PIPE_MPEG12_MB_TYPE_INTRA ? 1 : 0; for (y = 0; y < 2; ++y) { for (x = 0; x < 2; ++x) { if (mb->coded_block_pattern & const_empty_block_mask_420[0][y][x]) { struct vl_ycbcr_block *stream = buf->ycbcr_stream[0]; stream->x = mb->x * 2 + x; stream->y = mb->y * 2 + y; stream->intra = intra; stream->coding = mb->macroblock_modes.bits.dct_type; stream->block_num = buf->block_num++; buf->num_ycbcr_blocks[0]++; buf->ycbcr_stream[0]++; num_blocks++; } } } /* TODO: Implement 422, 444 */ //assert(ctx->base.chroma_format == PIPE_VIDEO_CHROMA_FORMAT_420); for (tb = 1; tb < 3; ++tb) { if (mb->coded_block_pattern & const_empty_block_mask_420[tb][0][0]) { struct vl_ycbcr_block *stream = buf->ycbcr_stream[tb]; stream->x = mb->x; stream->y = mb->y; stream->intra = intra; stream->coding = 0; stream->block_num = buf->block_num++; buf->num_ycbcr_blocks[tb]++; buf->ycbcr_stream[tb]++; num_blocks++; } } memcpy(buf->texels, mb->blocks, 64 * sizeof(short) * num_blocks); buf->texels += 64 * num_blocks; } static void vl_mpeg12_destroy_buffer(void *buffer) { struct vl_mpeg12_buffer *buf = buffer; assert(buf); cleanup_zscan_buffer(buf); cleanup_idct_buffer(buf); cleanup_mc_buffer(buf); vl_vb_cleanup(&buf->vertex_stream); FREE(buf); } static void vl_mpeg12_destroy(struct pipe_video_decoder *decoder) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder*)decoder; unsigned i; assert(decoder); /* Asserted in softpipe_delete_fs_state() for some reason */ dec->base.context->bind_vs_state(dec->base.context, NULL); dec->base.context->bind_fs_state(dec->base.context, NULL); dec->base.context->delete_depth_stencil_alpha_state(dec->base.context, dec->dsa); dec->base.context->delete_sampler_state(dec->base.context, dec->sampler_ycbcr); vl_mc_cleanup(&dec->mc_y); vl_mc_cleanup(&dec->mc_c); dec->mc_source->destroy(dec->mc_source); if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) { vl_idct_cleanup(&dec->idct_y); vl_idct_cleanup(&dec->idct_c); dec->idct_source->destroy(dec->idct_source); } vl_zscan_cleanup(&dec->zscan_y); vl_zscan_cleanup(&dec->zscan_c); dec->base.context->delete_vertex_elements_state(dec->base.context, dec->ves_ycbcr); dec->base.context->delete_vertex_elements_state(dec->base.context, dec->ves_mv); pipe_resource_reference(&dec->quads.buffer, NULL); pipe_resource_reference(&dec->pos.buffer, NULL); pipe_sampler_view_reference(&dec->zscan_linear, NULL); pipe_sampler_view_reference(&dec->zscan_normal, NULL); pipe_sampler_view_reference(&dec->zscan_alternate, NULL); for (i = 0; i < 4; ++i) if (dec->dec_buffers[i]) vl_mpeg12_destroy_buffer(dec->dec_buffers[i]); FREE(dec); } static struct vl_mpeg12_buffer * vl_mpeg12_get_decode_buffer(struct vl_mpeg12_decoder *dec) { struct vl_mpeg12_buffer *buffer; assert(dec); buffer = vl_video_buffer_get_associated_data(dec->target, &dec->base); if (buffer) return buffer; buffer = dec->dec_buffers[dec->current_buffer]; if (buffer) return buffer; buffer = CALLOC_STRUCT(vl_mpeg12_buffer); if (buffer == NULL) return NULL; if (!vl_vb_init(&buffer->vertex_stream, dec->base.context, dec->base.width / MACROBLOCK_WIDTH, dec->base.height / MACROBLOCK_HEIGHT)) goto error_vertex_buffer; if (!init_mc_buffer(dec, buffer)) goto error_mc; if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) if (!init_idct_buffer(dec, buffer)) goto error_idct; if (!init_zscan_buffer(dec, buffer)) goto error_zscan; if (dec->base.entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) vl_mpg12_bs_init(&buffer->bs, &dec->base); if (dec->expect_chunked_decode) vl_video_buffer_set_associated_data(dec->target, &dec->base, buffer, vl_mpeg12_destroy_buffer); else dec->dec_buffers[dec->current_buffer] = buffer; return buffer; error_zscan: cleanup_idct_buffer(buffer); error_idct: cleanup_mc_buffer(buffer); error_mc: vl_vb_cleanup(&buffer->vertex_stream); error_vertex_buffer: FREE(buffer); return NULL; } static void vl_mpeg12_set_picture_parameters(struct pipe_video_decoder *decoder, struct pipe_picture_desc *picture) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct pipe_mpeg12_picture_desc *pic = (struct pipe_mpeg12_picture_desc *)picture; assert(dec && pic); dec->picture_desc = *pic; } static void vl_mpeg12_set_quant_matrix(struct pipe_video_decoder *decoder, const struct pipe_quant_matrix *matrix) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; const struct pipe_mpeg12_quant_matrix *m = (const struct pipe_mpeg12_quant_matrix *)matrix; assert(dec); assert(matrix->codec == PIPE_VIDEO_CODEC_MPEG12); memcpy(dec->intra_matrix, m->intra_matrix, 64); memcpy(dec->non_intra_matrix, m->non_intra_matrix, 64); } static void vl_mpeg12_set_decode_target(struct pipe_video_decoder *decoder, struct pipe_video_buffer *target) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct pipe_surface **surfaces; unsigned i; assert(dec); dec->target = target; surfaces = target->get_surfaces(target); for (i = 0; i < VL_MAX_PLANES; ++i) pipe_surface_reference(&dec->target_surfaces[i], surfaces[i]); } static void vl_mpeg12_set_reference_frames(struct pipe_video_decoder *decoder, struct pipe_video_buffer **ref_frames, unsigned num_ref_frames) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct pipe_sampler_view **sv; unsigned i,j; assert(dec); assert(num_ref_frames <= VL_MAX_REF_FRAMES); for (i = 0; i < num_ref_frames; ++i) { sv = ref_frames[i]->get_sampler_view_planes(ref_frames[i]); for (j = 0; j < VL_MAX_PLANES; ++j) pipe_sampler_view_reference(&dec->ref_frames[i][j], sv[j]); } for (; i < VL_MAX_REF_FRAMES; ++i) for (j = 0; j < VL_MAX_PLANES; ++j) pipe_sampler_view_reference(&dec->ref_frames[i][j], NULL); } static void vl_mpeg12_begin_frame(struct pipe_video_decoder *decoder) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct vl_mpeg12_buffer *buf; struct pipe_resource *tex; struct pipe_box rect = { 0, 0, 0, 1, 1, 1 }; unsigned i; assert(dec && dec->target); buf = vl_mpeg12_get_decode_buffer(dec); assert(buf); if (dec->base.entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) dec->intra_matrix[0] = 1 << (7 - dec->picture_desc.intra_dc_precision); for (i = 0; i < VL_MAX_PLANES; ++i) { struct vl_zscan *zscan = i == 0 ? &dec->zscan_y : &dec->zscan_c; vl_zscan_upload_quant(zscan, &buf->zscan[i], dec->intra_matrix, true); vl_zscan_upload_quant(zscan, &buf->zscan[i], dec->non_intra_matrix, false); } vl_vb_map(&buf->vertex_stream, dec->base.context); tex = buf->zscan_source->texture; rect.width = tex->width0; rect.height = tex->height0; buf->tex_transfer = dec->base.context->get_transfer ( dec->base.context, tex, 0, PIPE_TRANSFER_WRITE | PIPE_TRANSFER_DISCARD_RANGE, &rect ); buf->block_num = 0; buf->texels = dec->base.context->transfer_map(dec->base.context, buf->tex_transfer); for (i = 0; i < VL_MAX_PLANES; ++i) { buf->ycbcr_stream[i] = vl_vb_get_ycbcr_stream(&buf->vertex_stream, i); buf->num_ycbcr_blocks[i] = 0; } for (i = 0; i < VL_MAX_REF_FRAMES; ++i) buf->mv_stream[i] = vl_vb_get_mv_stream(&buf->vertex_stream, i); if (dec->base.entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) { vl_mpg12_bs_set_picture_desc(&buf->bs, &dec->picture_desc); } else { for (i = 0; i < VL_MAX_PLANES; ++i) vl_zscan_set_layout(&buf->zscan[i], dec->zscan_linear); } } static void vl_mpeg12_decode_macroblock(struct pipe_video_decoder *decoder, const struct pipe_macroblock *macroblocks, unsigned num_macroblocks) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; const struct pipe_mpeg12_macroblock *mb = (const struct pipe_mpeg12_macroblock *)macroblocks; struct vl_mpeg12_buffer *buf; unsigned i, j, mv_weights[2]; assert(dec && dec->target); assert(macroblocks && macroblocks->codec == PIPE_VIDEO_CODEC_MPEG12); buf = vl_mpeg12_get_decode_buffer(dec); assert(buf); for (; num_macroblocks > 0; --num_macroblocks) { unsigned mb_addr = mb->y * dec->width_in_macroblocks + mb->x; if (mb->macroblock_type & (PIPE_MPEG12_MB_TYPE_PATTERN | PIPE_MPEG12_MB_TYPE_INTRA)) UploadYcbcrBlocks(dec, buf, mb); MacroBlockTypeToPipeWeights(mb, mv_weights); for (i = 0; i < 2; ++i) { if (!dec->ref_frames[i][0]) continue; buf->mv_stream[i][mb_addr] = MotionVectorToPipe ( mb, i, i ? PIPE_MPEG12_FS_FIRST_BACKWARD : PIPE_MPEG12_FS_FIRST_FORWARD, mv_weights[i] ); } /* see section 7.6.6 of the spec */ if (mb->num_skipped_macroblocks > 0) { struct vl_motionvector skipped_mv[2]; if (dec->ref_frames[0][0] && !dec->ref_frames[1][0]) { skipped_mv[0].top.x = skipped_mv[0].top.y = 0; skipped_mv[0].top.weight = PIPE_VIDEO_MV_WEIGHT_MAX; } else { skipped_mv[0] = buf->mv_stream[0][mb_addr]; skipped_mv[1] = buf->mv_stream[1][mb_addr]; } skipped_mv[0].top.field_select = PIPE_VIDEO_FRAME; skipped_mv[1].top.field_select = PIPE_VIDEO_FRAME; skipped_mv[0].bottom = skipped_mv[0].top; skipped_mv[1].bottom = skipped_mv[1].top; ++mb_addr; for (i = 0; i < mb->num_skipped_macroblocks; ++i, ++mb_addr) { for (j = 0; j < 2; ++j) { if (!dec->ref_frames[j][0]) continue; buf->mv_stream[j][mb_addr] = skipped_mv[j]; } } } ++mb; } } static void vl_mpeg12_decode_bitstream(struct pipe_video_decoder *decoder, unsigned num_buffers, const void * const *buffers, const unsigned *sizes) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct vl_mpeg12_buffer *buf; unsigned i; assert(dec && dec->target); buf = vl_mpeg12_get_decode_buffer(dec); assert(buf); for (i = 0; i < VL_MAX_PLANES; ++i) vl_zscan_set_layout(&buf->zscan[i], dec->picture_desc.alternate_scan ? dec->zscan_alternate : dec->zscan_normal); vl_mpg12_bs_decode(&buf->bs, num_buffers, buffers, sizes); } static void vl_mpeg12_end_frame(struct pipe_video_decoder *decoder) { struct vl_mpeg12_decoder *dec = (struct vl_mpeg12_decoder *)decoder; struct pipe_sampler_view **mc_source_sv; struct pipe_vertex_buffer vb[3]; struct vl_mpeg12_buffer *buf; unsigned i, j, component; unsigned nr_components; assert(dec && dec->target); buf = vl_mpeg12_get_decode_buffer(dec); vl_vb_unmap(&buf->vertex_stream, dec->base.context); dec->base.context->transfer_unmap(dec->base.context, buf->tex_transfer); dec->base.context->transfer_destroy(dec->base.context, buf->tex_transfer); vb[0] = dec->quads; vb[1] = dec->pos; dec->base.context->bind_vertex_elements_state(dec->base.context, dec->ves_mv); for (i = 0; i < VL_MAX_PLANES; ++i) { if (!dec->target_surfaces[i]) continue; vl_mc_set_surface(&buf->mc[i], dec->target_surfaces[i]); for (j = 0; j < VL_MAX_REF_FRAMES; ++j) { if (!dec->ref_frames[j][i]) continue; vb[2] = vl_vb_get_mv(&buf->vertex_stream, j);; dec->base.context->set_vertex_buffers(dec->base.context, 3, vb); vl_mc_render_ref(i ? &dec->mc_c : &dec->mc_y, &buf->mc[i], dec->ref_frames[j][i]); } } dec->base.context->bind_vertex_elements_state(dec->base.context, dec->ves_ycbcr); for (i = 0; i < VL_MAX_PLANES; ++i) { if (!buf->num_ycbcr_blocks[i]) continue; vb[1] = vl_vb_get_ycbcr(&buf->vertex_stream, i); dec->base.context->set_vertex_buffers(dec->base.context, 2, vb); vl_zscan_render(i ? &dec->zscan_c : & dec->zscan_y, &buf->zscan[i] , buf->num_ycbcr_blocks[i]); if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) vl_idct_flush(i ? &dec->idct_c : &dec->idct_y, &buf->idct[i], buf->num_ycbcr_blocks[i]); } mc_source_sv = dec->mc_source->get_sampler_view_planes(dec->mc_source); for (i = 0, component = 0; i < VL_MAX_PLANES; ++i) { if (!dec->target_surfaces[i]) continue; nr_components = util_format_get_nr_components(dec->target_surfaces[i]->texture->format); for (j = 0; j < nr_components; ++j, ++component) { if (!buf->num_ycbcr_blocks[i]) continue; vb[1] = vl_vb_get_ycbcr(&buf->vertex_stream, component); dec->base.context->set_vertex_buffers(dec->base.context, 2, vb); if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) vl_idct_prepare_stage2(i ? &dec->idct_c : &dec->idct_y, &buf->idct[component]); else { dec->base.context->set_fragment_sampler_views(dec->base.context, 1, &mc_source_sv[component]); dec->base.context->bind_fragment_sampler_states(dec->base.context, 1, &dec->sampler_ycbcr); } vl_mc_render_ycbcr(i ? &dec->mc_c : &dec->mc_y, &buf->mc[i], j, buf->num_ycbcr_blocks[component]); } } ++dec->current_buffer; dec->current_buffer %= 4; } static void vl_mpeg12_flush(struct pipe_video_decoder *decoder) { assert(decoder); //Noop, for shaders it is much faster to flush everything in end_frame } static bool init_pipe_state(struct vl_mpeg12_decoder *dec) { struct pipe_depth_stencil_alpha_state dsa; struct pipe_sampler_state sampler; unsigned i; assert(dec); memset(&dsa, 0, sizeof dsa); dsa.depth.enabled = 0; dsa.depth.writemask = 0; dsa.depth.func = PIPE_FUNC_ALWAYS; for (i = 0; i < 2; ++i) { dsa.stencil[i].enabled = 0; dsa.stencil[i].func = PIPE_FUNC_ALWAYS; dsa.stencil[i].fail_op = PIPE_STENCIL_OP_KEEP; dsa.stencil[i].zpass_op = PIPE_STENCIL_OP_KEEP; dsa.stencil[i].zfail_op = PIPE_STENCIL_OP_KEEP; dsa.stencil[i].valuemask = 0; dsa.stencil[i].writemask = 0; } dsa.alpha.enabled = 0; dsa.alpha.func = PIPE_FUNC_ALWAYS; dsa.alpha.ref_value = 0; dec->dsa = dec->base.context->create_depth_stencil_alpha_state(dec->base.context, &dsa); dec->base.context->bind_depth_stencil_alpha_state(dec->base.context, dec->dsa); memset(&sampler, 0, sizeof(sampler)); sampler.wrap_s = PIPE_TEX_WRAP_CLAMP_TO_EDGE; sampler.wrap_t = PIPE_TEX_WRAP_CLAMP_TO_EDGE; sampler.wrap_r = PIPE_TEX_WRAP_CLAMP_TO_BORDER; sampler.min_img_filter = PIPE_TEX_FILTER_NEAREST; sampler.min_mip_filter = PIPE_TEX_MIPFILTER_NONE; sampler.mag_img_filter = PIPE_TEX_FILTER_NEAREST; sampler.compare_mode = PIPE_TEX_COMPARE_NONE; sampler.compare_func = PIPE_FUNC_ALWAYS; sampler.normalized_coords = 1; dec->sampler_ycbcr = dec->base.context->create_sampler_state(dec->base.context, &sampler); if (!dec->sampler_ycbcr) return false; return true; } static const struct format_config* find_format_config(struct vl_mpeg12_decoder *dec, const struct format_config configs[], unsigned num_configs) { struct pipe_screen *screen; unsigned i; assert(dec); screen = dec->base.context->screen; for (i = 0; i < num_configs; ++i) { if (!screen->is_format_supported(screen, configs[i].zscan_source_format, PIPE_TEXTURE_2D, 1, PIPE_BIND_SAMPLER_VIEW)) continue; if (configs[i].idct_source_format != PIPE_FORMAT_NONE) { if (!screen->is_format_supported(screen, configs[i].idct_source_format, PIPE_TEXTURE_2D, 1, PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET)) continue; if (!screen->is_format_supported(screen, configs[i].mc_source_format, PIPE_TEXTURE_3D, 1, PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET)) continue; } else { if (!screen->is_format_supported(screen, configs[i].mc_source_format, PIPE_TEXTURE_2D, 1, PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_RENDER_TARGET)) continue; } return &configs[i]; } return NULL; } static bool init_zscan(struct vl_mpeg12_decoder *dec, const struct format_config* format_config) { unsigned num_channels; assert(dec); dec->zscan_source_format = format_config->zscan_source_format; dec->zscan_linear = vl_zscan_layout(dec->base.context, vl_zscan_linear, dec->blocks_per_line); dec->zscan_normal = vl_zscan_layout(dec->base.context, vl_zscan_normal, dec->blocks_per_line); dec->zscan_alternate = vl_zscan_layout(dec->base.context, vl_zscan_alternate, dec->blocks_per_line); num_channels = dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT ? 4 : 1; if (!vl_zscan_init(&dec->zscan_y, dec->base.context, dec->base.width, dec->base.height, dec->blocks_per_line, dec->num_blocks, num_channels)) return false; if (!vl_zscan_init(&dec->zscan_c, dec->base.context, dec->chroma_width, dec->chroma_height, dec->blocks_per_line, dec->num_blocks, num_channels)) return false; return true; } static bool init_idct(struct vl_mpeg12_decoder *dec, const struct format_config* format_config) { unsigned nr_of_idct_render_targets, max_inst; enum pipe_format formats[3]; struct pipe_sampler_view *matrix = NULL; nr_of_idct_render_targets = dec->base.context->screen->get_param ( dec->base.context->screen, PIPE_CAP_MAX_RENDER_TARGETS ); max_inst = dec->base.context->screen->get_shader_param ( dec->base.context->screen, PIPE_SHADER_FRAGMENT, PIPE_SHADER_CAP_MAX_INSTRUCTIONS ); // Just assume we need 32 inst per render target, not 100% true, but should work in most cases if (nr_of_idct_render_targets >= 4 && max_inst >= 32*4) // more than 4 render targets usually doesn't makes any seens nr_of_idct_render_targets = 4; else nr_of_idct_render_targets = 1; formats[0] = formats[1] = formats[2] = format_config->idct_source_format; dec->idct_source = vl_video_buffer_create_ex ( dec->base.context, dec->base.width / 4, dec->base.height, 1, dec->base.chroma_format, formats, PIPE_USAGE_STATIC ); if (!dec->idct_source) goto error_idct_source; formats[0] = formats[1] = formats[2] = format_config->mc_source_format; dec->mc_source = vl_video_buffer_create_ex ( dec->base.context, dec->base.width / nr_of_idct_render_targets, dec->base.height / 4, nr_of_idct_render_targets, dec->base.chroma_format, formats, PIPE_USAGE_STATIC ); if (!dec->mc_source) goto error_mc_source; if (!(matrix = vl_idct_upload_matrix(dec->base.context, format_config->idct_scale))) goto error_matrix; if (!vl_idct_init(&dec->idct_y, dec->base.context, dec->base.width, dec->base.height, nr_of_idct_render_targets, matrix, matrix)) goto error_y; if(!vl_idct_init(&dec->idct_c, dec->base.context, dec->chroma_width, dec->chroma_height, nr_of_idct_render_targets, matrix, matrix)) goto error_c; pipe_sampler_view_reference(&matrix, NULL); return true; error_c: vl_idct_cleanup(&dec->idct_y); error_y: pipe_sampler_view_reference(&matrix, NULL); error_matrix: dec->mc_source->destroy(dec->mc_source); error_mc_source: dec->idct_source->destroy(dec->idct_source); error_idct_source: return false; } static bool init_mc_source_widthout_idct(struct vl_mpeg12_decoder *dec, const struct format_config* format_config) { enum pipe_format formats[3]; formats[0] = formats[1] = formats[2] = format_config->mc_source_format; dec->mc_source = vl_video_buffer_create_ex ( dec->base.context, dec->base.width, dec->base.height, 1, dec->base.chroma_format, formats, PIPE_USAGE_STATIC ); return dec->mc_source != NULL; } static void mc_vert_shader_callback(void *priv, struct vl_mc *mc, struct ureg_program *shader, unsigned first_output, struct ureg_dst tex) { struct vl_mpeg12_decoder *dec = priv; struct ureg_dst o_vtex; assert(priv && mc); assert(shader); if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) { struct vl_idct *idct = mc == &dec->mc_y ? &dec->idct_y : &dec->idct_c; vl_idct_stage2_vert_shader(idct, shader, first_output, tex); } else { o_vtex = ureg_DECL_output(shader, TGSI_SEMANTIC_GENERIC, first_output); ureg_MOV(shader, ureg_writemask(o_vtex, TGSI_WRITEMASK_XY), ureg_src(tex)); } } static void mc_frag_shader_callback(void *priv, struct vl_mc *mc, struct ureg_program *shader, unsigned first_input, struct ureg_dst dst) { struct vl_mpeg12_decoder *dec = priv; struct ureg_src src, sampler; assert(priv && mc); assert(shader); if (dec->base.entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) { struct vl_idct *idct = mc == &dec->mc_y ? &dec->idct_y : &dec->idct_c; vl_idct_stage2_frag_shader(idct, shader, first_input, dst); } else { src = ureg_DECL_fs_input(shader, TGSI_SEMANTIC_GENERIC, first_input, TGSI_INTERPOLATE_LINEAR); sampler = ureg_DECL_sampler(shader, 0); ureg_TEX(shader, dst, TGSI_TEXTURE_2D, src, sampler); } } struct pipe_video_decoder * vl_create_mpeg12_decoder(struct pipe_context *context, enum pipe_video_profile profile, enum pipe_video_entrypoint entrypoint, enum pipe_video_chroma_format chroma_format, unsigned width, unsigned height, unsigned max_references, bool expect_chunked_decode) { const unsigned block_size_pixels = BLOCK_WIDTH * BLOCK_HEIGHT; const struct format_config *format_config; struct vl_mpeg12_decoder *dec; assert(u_reduce_video_profile(profile) == PIPE_VIDEO_CODEC_MPEG12); dec = CALLOC_STRUCT(vl_mpeg12_decoder); if (!dec) return NULL; dec->base.context = context; dec->base.profile = profile; dec->base.entrypoint = entrypoint; dec->base.chroma_format = chroma_format; dec->base.width = width; dec->base.height = height; dec->base.max_references = max_references; dec->base.destroy = vl_mpeg12_destroy; dec->base.set_picture_parameters = vl_mpeg12_set_picture_parameters; dec->base.set_quant_matrix = vl_mpeg12_set_quant_matrix; dec->base.set_decode_target = vl_mpeg12_set_decode_target; dec->base.set_reference_frames = vl_mpeg12_set_reference_frames; dec->base.begin_frame = vl_mpeg12_begin_frame; dec->base.decode_macroblock = vl_mpeg12_decode_macroblock; dec->base.decode_bitstream = vl_mpeg12_decode_bitstream; dec->base.end_frame = vl_mpeg12_end_frame; dec->base.flush = vl_mpeg12_flush; dec->blocks_per_line = MAX2(util_next_power_of_two(dec->base.width) / block_size_pixels, 4); dec->num_blocks = (dec->base.width * dec->base.height) / block_size_pixels; dec->width_in_macroblocks = align(dec->base.width, MACROBLOCK_WIDTH) / MACROBLOCK_WIDTH; dec->expect_chunked_decode = expect_chunked_decode; /* TODO: Implement 422, 444 */ assert(dec->base.chroma_format == PIPE_VIDEO_CHROMA_FORMAT_420); if (dec->base.chroma_format == PIPE_VIDEO_CHROMA_FORMAT_420) { dec->chroma_width = dec->base.width / 2; dec->chroma_height = dec->base.height / 2; dec->num_blocks = dec->num_blocks * 2; } else if (dec->base.chroma_format == PIPE_VIDEO_CHROMA_FORMAT_422) { dec->chroma_width = dec->base.width; dec->chroma_height = dec->base.height / 2; dec->num_blocks = dec->num_blocks * 2 + dec->num_blocks; } else { dec->chroma_width = dec->base.width; dec->chroma_height = dec->base.height; dec->num_blocks = dec->num_blocks * 3; } dec->quads = vl_vb_upload_quads(dec->base.context); dec->pos = vl_vb_upload_pos( dec->base.context, dec->base.width / MACROBLOCK_WIDTH, dec->base.height / MACROBLOCK_HEIGHT ); dec->ves_ycbcr = vl_vb_get_ves_ycbcr(dec->base.context); dec->ves_mv = vl_vb_get_ves_mv(dec->base.context); switch (entrypoint) { case PIPE_VIDEO_ENTRYPOINT_BITSTREAM: format_config = find_format_config(dec, bitstream_format_config, num_bitstream_format_configs); break; case PIPE_VIDEO_ENTRYPOINT_IDCT: format_config = find_format_config(dec, idct_format_config, num_idct_format_configs); break; case PIPE_VIDEO_ENTRYPOINT_MC: format_config = find_format_config(dec, mc_format_config, num_mc_format_configs); break; default: assert(0); FREE(dec); return NULL; } if (!format_config) { FREE(dec); return NULL; } if (!init_zscan(dec, format_config)) goto error_zscan; if (entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) { if (!init_idct(dec, format_config)) goto error_sources; } else { if (!init_mc_source_widthout_idct(dec, format_config)) goto error_sources; } if (!vl_mc_init(&dec->mc_y, dec->base.context, dec->base.width, dec->base.height, MACROBLOCK_HEIGHT, format_config->mc_scale, mc_vert_shader_callback, mc_frag_shader_callback, dec)) goto error_mc_y; // TODO if (!vl_mc_init(&dec->mc_c, dec->base.context, dec->base.width, dec->base.height, BLOCK_HEIGHT, format_config->mc_scale, mc_vert_shader_callback, mc_frag_shader_callback, dec)) goto error_mc_c; if (!init_pipe_state(dec)) goto error_pipe_state; memset(dec->intra_matrix, 0x10, 64); memset(dec->non_intra_matrix, 0x10, 64); return &dec->base; error_pipe_state: vl_mc_cleanup(&dec->mc_c); error_mc_c: vl_mc_cleanup(&dec->mc_y); error_mc_y: if (entrypoint <= PIPE_VIDEO_ENTRYPOINT_IDCT) { vl_idct_cleanup(&dec->idct_y); vl_idct_cleanup(&dec->idct_c); dec->idct_source->destroy(dec->idct_source); } dec->mc_source->destroy(dec->mc_source); error_sources: vl_zscan_cleanup(&dec->zscan_y); vl_zscan_cleanup(&dec->zscan_c); error_zscan: FREE(dec); return NULL; }