diff options
Diffstat (limited to 'contrib')
-rw-r--r-- | contrib/ffmpeg/A01-multi-object-pgs.patch | 671 |
1 files changed, 671 insertions, 0 deletions
diff --git a/contrib/ffmpeg/A01-multi-object-pgs.patch b/contrib/ffmpeg/A01-multi-object-pgs.patch new file mode 100644 index 000000000..1a177aee0 --- /dev/null +++ b/contrib/ffmpeg/A01-multi-object-pgs.patch @@ -0,0 +1,671 @@ +From 638cb29d7579ad76ea358ee30ab5e6069ee796c3 Mon Sep 17 00:00:00 2001 +From: John Stebbins <[email protected]> +Date: Thu, 12 Jun 2014 15:24:26 -0700 +Subject: [PATCH] pgssubdec: handle more complex PGS scenarios + +Add ability to handle multiple palettes and objects simultaneously. +Each simultaneous object is given its own AVSubtitleRect. +Note that there can be up to 64 currently valid objects, but only +2 at any one time can be "presented". +--- + libavcodec/pgssubdec.c | 424 ++++++++++++++++++++++++++++++++++--------------- + 1 file changed, 294 insertions(+), 130 deletions(-) + +diff --git a/libavcodec/pgssubdec.c b/libavcodec/pgssubdec.c +index 4f175a3..1cafd9f 100644 +--- a/libavcodec/pgssubdec.c ++++ b/libavcodec/pgssubdec.c +@@ -33,38 +33,104 @@ + #include "libavutil/imgutils.h" + + #define RGBA(r,g,b,a) (((a) << 24) | ((r) << 16) | ((g) << 8) | (b)) ++#define MAX_EPOCH_PALETTES 8 // Max 8 allowed per PGS epoch ++#define MAX_EPOCH_OBJECTS 64 // Max 64 allowed per PGS epoch ++#define MAX_OBJECT_REFS 2 // Max objects per display set + + enum SegmentType { + PALETTE_SEGMENT = 0x14, +- PICTURE_SEGMENT = 0x15, ++ OBJECT_SEGMENT = 0x15, + PRESENTATION_SEGMENT = 0x16, + WINDOW_SEGMENT = 0x17, + DISPLAY_SEGMENT = 0x80, + }; + ++typedef struct PGSSubObjectRef { ++ int id; ++ int window_id; ++ uint8_t composition_flag; ++ int x; ++ int y; ++ int crop_x; ++ int crop_y; ++ int crop_w; ++ int crop_h; ++} PGSSubObjectRef; ++ + typedef struct PGSSubPresentation { +- int x; +- int y; + int id_number; +- int object_number; +- uint8_t composition_flag; ++ int palette_id; ++ int object_count; ++ PGSSubObjectRef objects[MAX_OBJECT_REFS]; + int64_t pts; + } PGSSubPresentation; + +-typedef struct PGSSubPicture { ++typedef struct PGSSubObject { ++ int id; + int w; + int h; + uint8_t *rle; + unsigned int rle_buffer_size, rle_data_len; + unsigned int rle_remaining_len; +-} PGSSubPicture; ++} PGSSubObject; ++ ++typedef struct PGSSubObjects { ++ int count; ++ PGSSubObject object[MAX_EPOCH_OBJECTS]; ++} PGSSubObjects; ++ ++typedef struct PGSSubPalette { ++ int id; ++ uint32_t clut[256]; ++} PGSSubPalette; ++ ++typedef struct PGSSubPalettes { ++ int count; ++ PGSSubPalette palette[MAX_EPOCH_PALETTES]; ++} PGSSubPalettes; + + typedef struct PGSSubContext { + PGSSubPresentation presentation; +- uint32_t clut[256]; +- PGSSubPicture picture; ++ PGSSubPalettes palettes; ++ PGSSubObjects objects; + } PGSSubContext; + ++static void flush_cache(AVCodecContext *avctx) ++{ ++ PGSSubContext *ctx = avctx->priv_data; ++ int i; ++ ++ for (i = 0; i < ctx->objects.count; i++) { ++ av_freep(&ctx->objects.object[i].rle); ++ ctx->objects.object[i].rle_buffer_size = 0; ++ ctx->objects.object[i].rle_remaining_len = 0; ++ } ++ ctx->objects.count = 0; ++ ctx->palettes.count = 0; ++} ++ ++static PGSSubObject * find_object(int id, PGSSubObjects *objects) ++{ ++ int i; ++ ++ for (i = 0; i < objects->count; i++) { ++ if (objects->object[i].id == id) ++ return &objects->object[i]; ++ } ++ return NULL; ++} ++ ++static PGSSubPalette * find_palette(int id, PGSSubPalettes *palettes) ++{ ++ int i; ++ ++ for (i = 0; i < palettes->count; i++) { ++ if (palettes->palette[i].id == id) ++ return &palettes->palette[i]; ++ } ++ return NULL; ++} ++ + static av_cold int init_decoder(AVCodecContext *avctx) + { + avctx->pix_fmt = AV_PIX_FMT_PAL8; +@@ -74,10 +140,7 @@ static av_cold int init_decoder(AVCodecContext *avctx) + + static av_cold int close_decoder(AVCodecContext *avctx) + { +- PGSSubContext *ctx = avctx->priv_data; +- +- av_freep(&ctx->picture.rle); +- ctx->picture.rle_buffer_size = 0; ++ flush_cache(avctx); + + return 0; + } +@@ -92,7 +155,7 @@ static av_cold int close_decoder(AVCodecContext *avctx) + * @param buf pointer to the RLE data to process + * @param buf_size size of the RLE data to process + */ +-static int decode_rle(AVCodecContext *avctx, AVSubtitle *sub, ++static int decode_rle(AVCodecContext *avctx, AVSubtitleRect *rect, + const uint8_t *buf, unsigned int buf_size) + { + const uint8_t *rle_bitmap_end; +@@ -100,15 +163,15 @@ static int decode_rle(AVCodecContext *avctx, AVSubtitle *sub, + + rle_bitmap_end = buf + buf_size; + +- sub->rects[0]->pict.data[0] = av_malloc(sub->rects[0]->w * sub->rects[0]->h); ++ rect->pict.data[0] = av_malloc(rect->w * rect->h); + +- if (!sub->rects[0]->pict.data[0]) +- return -1; ++ if (!rect->pict.data[0]) ++ return AVERROR(ENOMEM); + + pixel_count = 0; + line_count = 0; + +- while (buf < rle_bitmap_end && line_count < sub->rects[0]->h) { ++ while (buf < rle_bitmap_end && line_count < rect->h) { + uint8_t flags, color; + int run; + +@@ -123,27 +186,31 @@ static int decode_rle(AVCodecContext *avctx, AVSubtitle *sub, + color = flags & 0x80 ? bytestream_get_byte(&buf) : 0; + } + +- if (run > 0 && pixel_count + run <= sub->rects[0]->w * sub->rects[0]->h) { +- memset(sub->rects[0]->pict.data[0] + pixel_count, color, run); ++ if (run > 0 && pixel_count + run <= rect->w * rect->h) { ++ memset(rect->pict.data[0] + pixel_count, color, run); + pixel_count += run; + } else if (!run) { + /* + * New Line. Check if correct pixels decoded, if not display warning + * and adjust bitmap pointer to correct new line position. + */ +- if (pixel_count % sub->rects[0]->w > 0) ++ if (pixel_count % rect->w > 0) { + av_log(avctx, AV_LOG_ERROR, "Decoded %d pixels, when line should be %d pixels\n", +- pixel_count % sub->rects[0]->w, sub->rects[0]->w); ++ pixel_count % rect->w, rect->w); ++ if (avctx->err_recognition & AV_EF_EXPLODE) { ++ return AVERROR_INVALIDDATA; ++ } ++ } + line_count++; + } + } + +- if (pixel_count < sub->rects[0]->w * sub->rects[0]->h) { ++ if (pixel_count < rect->w * rect->h) { + av_log(avctx, AV_LOG_ERROR, "Insufficient RLE data for subtitle\n"); +- return -1; ++ return AVERROR_INVALIDDATA; + } + +- av_dlog(avctx, "Pixel Count = %d, Area = %d\n", pixel_count, sub->rects[0]->w * sub->rects[0]->h); ++ av_dlog(avctx, "Pixel Count = %d, Area = %d\n", pixel_count, rect->w * rect->h); + + return 0; + } +@@ -157,40 +224,52 @@ static int decode_rle(AVCodecContext *avctx, AVSubtitle *sub, + * @param avctx contains the current codec context + * @param buf pointer to the packet to process + * @param buf_size size of packet to process +- * @todo TODO: Enable support for RLE data over multiple packets + */ +-static int parse_picture_segment(AVCodecContext *avctx, ++static int parse_object_segment(AVCodecContext *avctx, + const uint8_t *buf, int buf_size) + { + PGSSubContext *ctx = avctx->priv_data; ++ PGSSubObject *object; + + uint8_t sequence_desc; + unsigned int rle_bitmap_len, width, height; ++ int id; + + if (buf_size <= 4) +- return -1; ++ return AVERROR_INVALIDDATA; + buf_size -= 4; + +- /* skip 3 unknown bytes: Object ID (2 bytes), Version Number */ +- buf += 3; ++ id = bytestream_get_be16(&buf); ++ object = find_object(id, &ctx->objects); ++ if (!object) { ++ if (ctx->objects.count >= MAX_EPOCH_OBJECTS) { ++ av_log(avctx, AV_LOG_ERROR, "Too many objects in epoch\n"); ++ return AVERROR_INVALIDDATA; ++ } ++ object = &ctx->objects.object[ctx->objects.count++]; ++ object->id = id; ++ } ++ ++ /* skip object version number */ ++ buf += 1; + + /* Read the Sequence Description to determine if start of RLE data or appended to previous RLE */ + sequence_desc = bytestream_get_byte(&buf); + + if (!(sequence_desc & 0x80)) { + /* Additional RLE data */ +- if (buf_size > ctx->picture.rle_remaining_len) +- return -1; ++ if (buf_size > object->rle_remaining_len) ++ return AVERROR_INVALIDDATA; + +- memcpy(ctx->picture.rle + ctx->picture.rle_data_len, buf, buf_size); +- ctx->picture.rle_data_len += buf_size; +- ctx->picture.rle_remaining_len -= buf_size; ++ memcpy(object->rle + object->rle_data_len, buf, buf_size); ++ object->rle_data_len += buf_size; ++ object->rle_remaining_len -= buf_size; + + return 0; + } + + if (buf_size <= 7) +- return -1; ++ return AVERROR_INVALIDDATA; + buf_size -= 7; + + /* Decode rle bitmap length, stored size includes width/height data */ +@@ -203,20 +282,20 @@ static int parse_picture_segment(AVCodecContext *avctx, + /* Make sure the bitmap is not too large */ + if (avctx->width < width || avctx->height < height) { + av_log(avctx, AV_LOG_ERROR, "Bitmap dimensions larger than video.\n"); +- return -1; ++ return AVERROR_INVALIDDATA; + } + +- ctx->picture.w = width; +- ctx->picture.h = height; ++ object->w = width; ++ object->h = height; + +- av_fast_malloc(&ctx->picture.rle, &ctx->picture.rle_buffer_size, rle_bitmap_len); ++ av_fast_malloc(&object->rle, &object->rle_buffer_size, rle_bitmap_len); + +- if (!ctx->picture.rle) +- return -1; ++ if (!object->rle) ++ return AVERROR(ENOMEM); + +- memcpy(ctx->picture.rle, buf, buf_size); +- ctx->picture.rle_data_len = buf_size; +- ctx->picture.rle_remaining_len = rle_bitmap_len - buf_size; ++ memcpy(object->rle, buf, buf_size); ++ object->rle_data_len = buf_size; ++ object->rle_remaining_len = rle_bitmap_len - buf_size; + + return 0; + } +@@ -231,19 +310,32 @@ static int parse_picture_segment(AVCodecContext *avctx, + * @param buf pointer to the packet to process + * @param buf_size size of packet to process + */ +-static void parse_palette_segment(AVCodecContext *avctx, ++static int parse_palette_segment(AVCodecContext *avctx, + const uint8_t *buf, int buf_size) + { + PGSSubContext *ctx = avctx->priv_data; ++ PGSSubPalette *palette; + + const uint8_t *buf_end = buf + buf_size; + const uint8_t *cm = ff_cropTbl + MAX_NEG_CROP; + int color_id; + int y, cb, cr, alpha; + int r, g, b, r_add, g_add, b_add; ++ int id; ++ ++ id = bytestream_get_byte(&buf); ++ palette = find_palette(id, &ctx->palettes); ++ if (!palette) { ++ if (ctx->palettes.count >= MAX_EPOCH_PALETTES) { ++ av_log(avctx, AV_LOG_ERROR, "Too many palettes in epoch\n"); ++ return AVERROR_INVALIDDATA; ++ } ++ palette = &ctx->palettes.palette[ctx->palettes.count++]; ++ palette->id = id; ++ } + +- /* Skip two null bytes */ +- buf += 2; ++ /* Skip palette version */ ++ buf += 1; + + while (buf < buf_end) { + color_id = bytestream_get_byte(&buf); +@@ -258,8 +350,9 @@ static void parse_palette_segment(AVCodecContext *avctx, + av_dlog(avctx, "Color %d := (%d,%d,%d,%d)\n", color_id, r, g, b, alpha); + + /* Store color in palette */ +- ctx->clut[color_id] = RGBA(r,g,b,alpha); ++ palette->clut[color_id] = RGBA(r,g,b,alpha); + } ++ return 0; + } + + /** +@@ -272,7 +365,6 @@ static void parse_palette_segment(AVCodecContext *avctx, + * @param buf pointer to the packet to process + * @param buf_size size of packet to process + * @todo TODO: Implement cropping +- * @todo TODO: Implement forcing of subtitles + */ + static int parse_presentation_segment(AVCodecContext *avctx, + const uint8_t *buf, int buf_size, +@@ -280,8 +372,9 @@ static int parse_presentation_segment(AVCodecContext *avctx, + { + PGSSubContext *ctx = avctx->priv_data; + +- int x, y, ret; ++ int i, state, ret; + ++ // Video descriptor + int w = bytestream_get_be16(&buf); + int h = bytestream_get_be16(&buf); + +@@ -293,49 +386,75 @@ static int parse_presentation_segment(AVCodecContext *avctx, + if (ret < 0) + return ret; + +- /* Skip 1 bytes of unknown, frame rate? */ ++ /* Skip 1 bytes of unknown, frame rate */ + buf++; + ++ // Composition descriptor + ctx->presentation.id_number = bytestream_get_be16(&buf); +- + /* +- * Skip 3 bytes of unknown: +- * state +- * palette_update_flag (0x80), +- * palette_id_to_use, ++ * state is a 2 bit field that defines pgs epoch boundaries ++ * 00 - Normal, previously defined objects and palettes are still valid ++ * 01 - Acquisition point, previous objects and palettes can be released ++ * 10 - Epoch start, previous objects and palettes can be released ++ * 11 - Epoch continue, previous objects and palettes can be released ++ * ++ * reserved 6 bits discarded + */ +- buf += 3; +- +- ctx->presentation.object_number = bytestream_get_byte(&buf); +- ctx->presentation.composition_flag = 0; +- if (!ctx->presentation.object_number) +- return 0; ++ state = bytestream_get_byte(&buf) >> 6; ++ if (state != 0) { ++ flush_cache(avctx); ++ } + + /* +- * Skip 3 bytes of unknown: +- * object_id_ref (2 bytes), +- * window_id_ref, ++ * skip palette_update_flag (0x80), + */ +- buf += 3; +- ctx->presentation.composition_flag = bytestream_get_byte(&buf); +- +- x = bytestream_get_be16(&buf); +- y = bytestream_get_be16(&buf); +- +- /* TODO If cropping, cropping_x, cropping_y, cropping_width, cropping_height (all 2 bytes).*/ ++ buf += 1; ++ ctx->presentation.palette_id = bytestream_get_byte(&buf); ++ ctx->presentation.object_count = bytestream_get_byte(&buf); ++ if (ctx->presentation.object_count > MAX_OBJECT_REFS) { ++ av_log(avctx, AV_LOG_ERROR, ++ "Invalid number of presentation objects %d\n", ++ ctx->presentation.object_count); ++ ctx->presentation.object_count = 2; ++ if (avctx->err_recognition & AV_EF_EXPLODE) { ++ return AVERROR_INVALIDDATA; ++ } ++ } + +- av_dlog(avctx, "Subtitle Placement x=%d, y=%d\n", x, y); ++ for (i = 0; i < ctx->presentation.object_count; i++) ++ { ++ ctx->presentation.objects[i].id = bytestream_get_be16(&buf); ++ ctx->presentation.objects[i].window_id = bytestream_get_byte(&buf); ++ ctx->presentation.objects[i].composition_flag = bytestream_get_byte(&buf); ++ ++ ctx->presentation.objects[i].x = bytestream_get_be16(&buf); ++ ctx->presentation.objects[i].y = bytestream_get_be16(&buf); ++ ++ // If cropping ++ if (ctx->presentation.objects[i].composition_flag & 0x80) { ++ ctx->presentation.objects[i].crop_x = bytestream_get_be16(&buf); ++ ctx->presentation.objects[i].crop_y = bytestream_get_be16(&buf); ++ ctx->presentation.objects[i].crop_w = bytestream_get_be16(&buf); ++ ctx->presentation.objects[i].crop_h = bytestream_get_be16(&buf); ++ } + +- if (x > avctx->width || y > avctx->height) { +- av_log(avctx, AV_LOG_ERROR, "Subtitle out of video bounds. x = %d, y = %d, video width = %d, video height = %d.\n", +- x, y, avctx->width, avctx->height); +- x = 0; y = 0; ++ av_dlog(avctx, "Subtitle Placement x=%d, y=%d\n", ++ ctx->presentation.objects[i].x, ctx->presentation.objects[i].y); ++ ++ if (ctx->presentation.objects[i].x > avctx->width || ++ ctx->presentation.objects[i].y > avctx->height) { ++ av_log(avctx, AV_LOG_ERROR, "Subtitle out of video bounds. x = %d, y = %d, video width = %d, video height = %d.\n", ++ ctx->presentation.objects[i].x, ++ ctx->presentation.objects[i].y, ++ avctx->width, avctx->height); ++ ctx->presentation.objects[i].x = 0; ++ ctx->presentation.objects[i].y = 0; ++ if (avctx->err_recognition & AV_EF_EXPLODE) { ++ return AVERROR_INVALIDDATA; ++ } ++ } + } + +- /* Fill in dimensions */ +- ctx->presentation.x = x; +- ctx->presentation.y = y; +- + return 0; + } + +@@ -348,64 +467,105 @@ static int parse_presentation_segment(AVCodecContext *avctx, + * @param data pointer to the data pertaining the subtitle to display + * @param buf pointer to the packet to process + * @param buf_size size of packet to process +- * @todo TODO: Fix start time, relies on correct PTS, currently too late +- * +- * @todo TODO: Fix end time, normally cleared by a second display +- * @todo segment, which is currently ignored as it clears +- * @todo the subtitle too early. + */ + static int display_end_segment(AVCodecContext *avctx, void *data, + const uint8_t *buf, int buf_size) + { + AVSubtitle *sub = data; + PGSSubContext *ctx = avctx->priv_data; +- +- /* +- * The end display time is a timeout value and is only reached +- * if the next subtitle is later then timeout or subtitle has +- * not been cleared by a subsequent empty display command. +- */ ++ PGSSubPalette *palette; ++ int i, ret; + + memset(sub, 0, sizeof(*sub)); + sub->pts = ctx->presentation.pts; +- +- // Blank if last object_number was 0. +- // Note that this may be wrong for more complex subtitles. +- if (!ctx->presentation.object_number) +- return 1; + sub->start_display_time = 0; +- sub->end_display_time = 20000; ++ // There is no explicit end time for PGS subtitles. The end time ++ // is defined by the start of the next sub which may contain no ++ // objects (i.e. clears the previous sub) ++ sub->end_display_time = UINT32_MAX; + sub->format = 0; + +- sub->rects = av_mallocz(sizeof(*sub->rects)); +- sub->rects[0] = av_mallocz(sizeof(*sub->rects[0])); +- sub->num_rects = 1; +- +- if (ctx->presentation.composition_flag & 0x40) +- sub->rects[0]->flags |= AV_SUBTITLE_FLAG_FORCED; ++ // Blank if last object_count was 0. ++ if (!ctx->presentation.object_count) ++ return 1; ++ sub->rects = av_mallocz(sizeof(*sub->rects) * ctx->presentation.object_count); ++ if (!sub->rects) { ++ return AVERROR(ENOMEM); ++ } ++ palette = find_palette(ctx->presentation.palette_id, &ctx->palettes); ++ if (!palette) { ++ // Missing palette. Should only happen with damaged streams. ++ av_log(avctx, AV_LOG_ERROR, "Invalid palette id %d\n", ++ ctx->presentation.palette_id); ++ avsubtitle_free(sub); ++ return AVERROR_INVALIDDATA; ++ } ++ for (i = 0; i < ctx->presentation.object_count; i++) { ++ PGSSubObject *object; + +- sub->rects[0]->x = ctx->presentation.x; +- sub->rects[0]->y = ctx->presentation.y; +- sub->rects[0]->w = ctx->picture.w; +- sub->rects[0]->h = ctx->picture.h; +- sub->rects[0]->type = SUBTITLE_BITMAP; ++ sub->rects[i] = av_mallocz(sizeof(*sub->rects[0])); ++ if (!sub->rects[i]) { ++ avsubtitle_free(sub); ++ return AVERROR(ENOMEM); ++ } ++ sub->num_rects++; ++ sub->rects[i]->type = SUBTITLE_BITMAP; ++ ++ /* Process bitmap */ ++ object = find_object(ctx->presentation.objects[i].id, &ctx->objects); ++ if (!object) { ++ // Missing object. Should only happen with damaged streams. ++ av_log(avctx, AV_LOG_ERROR, "Invalid object id %d\n", ++ ctx->presentation.objects[i].id); ++ if (avctx->err_recognition & AV_EF_EXPLODE) { ++ avsubtitle_free(sub); ++ return AVERROR_INVALIDDATA; ++ } ++ // Leaves rect empty with 0 width and height. ++ continue; ++ } ++ if (ctx->presentation.objects[i].composition_flag & 0x40) ++ sub->rects[i]->flags |= AV_SUBTITLE_FLAG_FORCED; ++ ++ sub->rects[i]->x = ctx->presentation.objects[i].x; ++ sub->rects[i]->y = ctx->presentation.objects[i].y; ++ sub->rects[i]->w = object->w; ++ sub->rects[i]->h = object->h; ++ ++ sub->rects[i]->pict.linesize[0] = object->w; ++ ++ if (object->rle) { ++ if (object->rle_remaining_len) { ++ av_log(avctx, AV_LOG_ERROR, "RLE data length %u is %u bytes shorter than expected\n", ++ object->rle_data_len, object->rle_remaining_len); ++ if (avctx->err_recognition & AV_EF_EXPLODE) { ++ avsubtitle_free(sub); ++ return AVERROR_INVALIDDATA; ++ } ++ } ++ ret = decode_rle(avctx, sub->rects[i], object->rle, object->rle_data_len); ++ if (ret < 0) { ++ if ((avctx->err_recognition & AV_EF_EXPLODE) || ++ ret == AVERROR(ENOMEM)) { ++ avsubtitle_free(sub); ++ return ret; ++ } ++ sub->rects[i]->w = 0; ++ sub->rects[i]->h = 0; ++ continue; ++ } ++ } ++ /* Allocate memory for colors */ ++ sub->rects[i]->nb_colors = 256; ++ sub->rects[i]->pict.data[1] = av_mallocz(AVPALETTE_SIZE); ++ if (!sub->rects[i]->pict.data[1]) { ++ avsubtitle_free(sub); ++ return AVERROR(ENOMEM); ++ } + +- /* Process bitmap */ +- sub->rects[0]->pict.linesize[0] = ctx->picture.w; ++ memcpy(sub->rects[i]->pict.data[1], palette->clut, sub->rects[i]->nb_colors * sizeof(uint32_t)); + +- if (ctx->picture.rle) { +- if (ctx->picture.rle_remaining_len) +- av_log(avctx, AV_LOG_ERROR, "RLE data length %u is %u bytes shorter than expected\n", +- ctx->picture.rle_data_len, ctx->picture.rle_remaining_len); +- if(decode_rle(avctx, sub, ctx->picture.rle, ctx->picture.rle_data_len) < 0) +- return 0; + } +- /* Allocate memory for colors */ +- sub->rects[0]->nb_colors = 256; +- sub->rects[0]->pict.data[1] = av_mallocz(AVPALETTE_SIZE); +- +- memcpy(sub->rects[0]->pict.data[1], ctx->clut, sub->rects[0]->nb_colors * sizeof(uint32_t)); +- + return 1; + } + +@@ -449,17 +609,16 @@ static int decode(AVCodecContext *avctx, void *data, int *data_size, + if (segment_type != DISPLAY_SEGMENT && segment_length > buf_end - buf) + break; + ++ ret = 0; + switch (segment_type) { + case PALETTE_SEGMENT: +- parse_palette_segment(avctx, buf, segment_length); ++ ret = parse_palette_segment(avctx, buf, segment_length); + break; +- case PICTURE_SEGMENT: +- parse_picture_segment(avctx, buf, segment_length); ++ case OBJECT_SEGMENT: ++ ret = parse_object_segment(avctx, buf, segment_length); + break; + case PRESENTATION_SEGMENT: + ret = parse_presentation_segment(avctx, buf, segment_length, avpkt->pts); +- if (ret < 0) +- return ret; + break; + case WINDOW_SEGMENT: + /* +@@ -472,13 +631,18 @@ static int decode(AVCodecContext *avctx, void *data, int *data_size, + */ + break; + case DISPLAY_SEGMENT: +- *data_size = display_end_segment(avctx, data, buf, segment_length); ++ ret = display_end_segment(avctx, data, buf, segment_length); ++ if (ret >= 0) ++ *data_size = ret; + break; + default: + av_log(avctx, AV_LOG_ERROR, "Unknown subtitle segment type 0x%x, length %d\n", + segment_type, segment_length); ++ ret = AVERROR_INVALIDDATA; + break; + } ++ if (ret < 0 && (avctx->err_recognition & AV_EF_EXPLODE)) ++ return ret; + + buf += segment_length; + } +-- +1.9.3 + |