/* $Id: sync.c,v 1.38 2005/04/14 21:57:58 titer Exp $ This file is part of the HandBrake source code. Homepage: . It may be used under the terms of the GNU General Public License. */ #include "hb.h" #include "hbffmpeg.h" #include #include "samplerate.h" #ifdef INT64_MIN #undef INT64_MIN /* Because it isn't defined correctly in Zeta */ #endif #define INT64_MIN (-9223372036854775807LL-1) #define AC3_SAMPLES_PER_FRAME 1536 typedef struct { hb_lock_t * mutex; int ref; /* Reference count to tell us when it's unused */ int count_frames; int64_t audio_pts_slip; int64_t video_pts_slip; int64_t pts_offset; /* Frame based point-to-point support */ int64_t audio_pts_thresh; int start_found; hb_cond_t * next_frame; int pts_count; int64_t * first_pts; } hb_sync_common_t; typedef struct { int index; double next_start; /* start time of next output frame */ int64_t first_drop; /* PTS of first 'went backwards' frame dropped */ int drop_count; /* count of 'time went backwards' drops */ /* Raw */ SRC_STATE * state; SRC_DATA data; /* AC-3 */ int ac3_size; uint8_t * ac3_buf; } hb_sync_audio_t; typedef struct { /* Video */ int first_frame; int64_t pts_skip; int64_t next_start; /* start time of next output frame */ int64_t first_drop; /* PTS of first 'went backwards' frame dropped */ int drop_count; /* count of 'time went backwards' drops */ int drops; /* frames dropped to make a cbr video stream */ int dups; /* frames duplicated to make a cbr video stream */ int video_sequence; int count_frames_max; int chap_mark; /* to propagate chapter mark across a drop */ hb_buffer_t * cur; /* The next picture to process */ /* Statistics */ uint64_t st_counts[4]; uint64_t st_dates[4]; uint64_t st_first; /* Subtitles */ hb_buffer_t * sub_list; /* list of subtitles to be passed thru or rendered */ } hb_sync_video_t; struct hb_work_private_s { hb_job_t * job; hb_sync_common_t * common; union { hb_sync_video_t video; hb_sync_audio_t audio; } type; }; /*********************************************************************** * Local prototypes **********************************************************************/ static void getPtsOffset( hb_work_object_t * w ); static int checkPtsOffset( hb_work_object_t * w ); static void InitAudio( hb_job_t * job, hb_sync_common_t * common, int i ); static void InsertSilence( hb_work_object_t * w, int64_t d ); static void UpdateState( hb_work_object_t * w ); static void UpdateSearchState( hb_work_object_t * w, int64_t start ); static hb_buffer_t * OutputAudioFrame( hb_audio_t *audio, hb_buffer_t *buf, hb_sync_audio_t *sync ); /*********************************************************************** * hb_work_sync_init *********************************************************************** * Initialize the work object **********************************************************************/ hb_work_object_t * hb_sync_init( hb_job_t * job ) { hb_title_t * title = job->title; hb_chapter_t * chapter; int i; uint64_t duration; hb_work_private_t * pv; hb_sync_video_t * sync; hb_work_object_t * w; hb_work_object_t * ret = NULL; pv = calloc( 1, sizeof( hb_work_private_t ) ); sync = &pv->type.video; pv->common = calloc( 1, sizeof( hb_sync_common_t ) ); pv->common->ref++; pv->common->mutex = hb_lock_init(); pv->common->audio_pts_thresh = -1; pv->common->next_frame = hb_cond_init(); pv->common->pts_count = 1; if ( job->frame_to_start || job->pts_to_start ) { pv->common->start_found = 0; } else { pv->common->start_found = 1; } ret = w = hb_get_work( WORK_SYNC_VIDEO ); w->private_data = pv; w->fifo_in = job->fifo_raw; w->fifo_out = job->fifo_sync; pv->job = job; pv->common->pts_offset = INT64_MIN; sync->first_frame = 1; if( job->pass == 2 ) { /* We already have an accurate frame count from pass 1 */ hb_interjob_t * interjob = hb_interjob_get( job->h ); sync->count_frames_max = interjob->frame_count; } else { /* Calculate how many video frames we are expecting */ if ( job->pts_to_stop ) { duration = job->pts_to_stop + 90000; } else if( job->frame_to_stop ) { /* Set the duration to a rough estimate */ duration = ( job->frame_to_stop / ( title->rate / title->rate_base ) ) * 90000; } else { duration = 0; for( i = job->chapter_start; i <= job->chapter_end; i++ ) { chapter = hb_list_item( title->list_chapter, i - 1 ); duration += chapter->duration; } duration += 90000; /* 1 second safety so we're sure we won't miss anything */ } sync->count_frames_max = duration * title->rate / title->rate_base / 90000; } hb_log( "sync: expecting %d video frames", sync->count_frames_max ); /* Initialize libsamplerate for every audio track we have */ if ( ! job->indepth_scan ) { for( i = 0; i < hb_list_count( title->list_audio ); i++ ) { InitAudio( job, pv->common, i ); } } pv->common->first_pts = malloc( sizeof(int64_t) * pv->common->pts_count ); for ( i = 0; i < pv->common->pts_count; i++ ) pv->common->first_pts[i] = INT64_MAX; return ret; } /*********************************************************************** * Close Video *********************************************************************** * **********************************************************************/ void syncVideoClose( hb_work_object_t * w ) { hb_work_private_t * pv = w->private_data; hb_job_t * job = pv->job; hb_sync_video_t * sync = &pv->type.video; // Wake up audio sync if it's still waiting on condition. pv->common->pts_offset = 0; pv->common->start_found = 1; hb_cond_broadcast( pv->common->next_frame ); if( sync->cur ) { hb_buffer_close( &sync->cur ); } hb_log( "sync: got %d frames, %d expected", pv->common->count_frames, sync->count_frames_max ); /* save data for second pass */ if( job->pass == 1 ) { /* Preserve frame count for better accuracy in pass 2 */ hb_interjob_t * interjob = hb_interjob_get( job->h ); interjob->frame_count = pv->common->count_frames; interjob->last_job = job->sequence_id; } if (sync->drops || sync->dups ) { hb_log( "sync: %d frames dropped, %d duplicated", sync->drops, sync->dups ); } hb_lock( pv->common->mutex ); if ( --pv->common->ref == 0 ) { hb_unlock( pv->common->mutex ); hb_lock_close( &pv->common->mutex ); free( pv->common ); } else { hb_unlock( pv->common->mutex ); } free( pv ); w->private_data = NULL; } /*********************************************************************** * syncVideoWork *********************************************************************** * **********************************************************************/ static hb_buffer_t * copy_subtitle( hb_buffer_t * src ); int syncVideoWork( hb_work_object_t * w, hb_buffer_t ** buf_in, hb_buffer_t ** buf_out ) { hb_buffer_t * cur, * next, * sub = NULL; hb_work_private_t * pv = w->private_data; hb_job_t * job = pv->job; hb_subtitle_t * subtitle; hb_sync_video_t * sync = &pv->type.video; int i; int64_t start, next_start; *buf_out = NULL; next = *buf_in; *buf_in = NULL; /* Wait till we can determine the initial pts of all streams */ if( next->size != 0 && pv->common->pts_offset == INT64_MIN ) { pv->common->first_pts[0] = next->start; hb_lock( pv->common->mutex ); while( pv->common->pts_offset == INT64_MIN ) { // Full fifos will make us wait forever, so get the // pts offset from the available streams if full if ( hb_fifo_is_full( job->fifo_raw ) ) { getPtsOffset( w ); hb_cond_broadcast( pv->common->next_frame ); } else if ( checkPtsOffset( w ) ) hb_cond_broadcast( pv->common->next_frame ); else hb_cond_timedwait( pv->common->next_frame, pv->common->mutex, 200 ); } hb_unlock( pv->common->mutex ); } hb_lock( pv->common->mutex ); next_start = next->start - pv->common->video_pts_slip; hb_unlock( pv->common->mutex ); /* Wait for start of point-to-point encoding */ if( !pv->common->start_found ) { hb_sync_video_t * sync = &pv->type.video; if( next->size == 0 ) { *buf_out = next; pv->common->start_found = 1; pv->common->first_pts[0] = INT64_MAX - 1; hb_cond_broadcast( pv->common->next_frame ); /* * Push through any subtitle EOFs in case they * were not synced through. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); if( subtitle->config.dest == PASSTHRUSUB ) { if( subtitle->source == VOBSUB ) hb_fifo_push( subtitle->fifo_sync, hb_buffer_init( 0 ) ); else hb_fifo_push( subtitle->fifo_out, hb_buffer_init( 0 ) ); } } return HB_WORK_DONE; } if ( pv->common->count_frames < job->frame_to_start || next->start < job->pts_to_start ) { // Flush any subtitles that have pts prior to the // current frame for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); while( ( sub = hb_fifo_see( subtitle->fifo_raw ) ) ) { if ( sub->start > next->start ) break; sub = hb_fifo_get( subtitle->fifo_raw ); hb_buffer_close( &sub ); } } hb_lock( pv->common->mutex ); // Tell the audio threads what must be dropped pv->common->audio_pts_thresh = next_start + pv->common->video_pts_slip; hb_cond_broadcast( pv->common->next_frame ); hb_unlock( pv->common->mutex ); UpdateSearchState( w, next_start ); hb_buffer_close( &next ); return HB_WORK_OK; } hb_lock( pv->common->mutex ); pv->common->audio_pts_thresh = 0; pv->common->audio_pts_slip += next_start; pv->common->video_pts_slip += next_start; next_start = 0; pv->common->start_found = 1; pv->common->count_frames = 0; hb_cond_broadcast( pv->common->next_frame ); hb_unlock( pv->common->mutex ); sync->st_first = 0; } if( !sync->cur ) { sync->cur = next; if( sync->cur->size == 0 ) { /* we got an end-of-stream as our first video packet? * Feed it downstream & signal that we're done. */ *buf_out = hb_buffer_init( 0 ); pv->common->start_found = 1; pv->common->first_pts[0] = INT64_MAX - 1; hb_cond_broadcast( pv->common->next_frame ); /* * Push through any subtitle EOFs in case they * were not synced through. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); if( subtitle->config.dest == PASSTHRUSUB ) { if( subtitle->source == VOBSUB ) hb_fifo_push( subtitle->fifo_sync, hb_buffer_init( 0 ) ); else hb_fifo_push( subtitle->fifo_out, hb_buffer_init( 0 ) ); } } return HB_WORK_DONE; } return HB_WORK_OK; } cur = sync->cur; /* At this point we have a frame to process. Let's check 1) if we will be able to push into the fifo ahead 2) if the next frame is there already, since we need it to compute the duration of the current frame*/ if( next->size == 0 ) { hb_buffer_close( &next ); pv->common->first_pts[0] = INT64_MAX - 1; cur->start = sync->next_start; cur->stop = cur->start + 90000. / ((double)job->vrate / (double)job->vrate_base); sync->next_start += cur->stop - cur->start;; /* Make sure last frame is reflected in frame count */ pv->common->count_frames++; /* Push the frame to the renderer */ hb_fifo_push( job->fifo_sync, cur ); sync->cur = NULL; /* we got an end-of-stream. Feed it downstream & signal that * we're done. Note that this means we drop the final frame of * video (we don't know its duration). On DVDs the final frame * is often strange and dropping it seems to be a good idea. */ *buf_out = hb_buffer_init( 0 ); /* * Push through any subtitle EOFs in case they were not synced through. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); if( subtitle->config.dest == PASSTHRUSUB ) { if( subtitle->source == VOBSUB ) hb_fifo_push( subtitle->fifo_sync, hb_buffer_init( 0 ) ); else hb_fifo_push( subtitle->fifo_out, hb_buffer_init( 0 ) ); } } pv->common->start_found = 1; hb_cond_broadcast( pv->common->next_frame ); return HB_WORK_DONE; } /* Check for end of point-to-point frame encoding */ if( job->frame_to_stop && pv->common->count_frames > job->frame_to_stop ) { // Drop an empty buffer into our output to ensure that things // get flushed all the way out. hb_buffer_close( &sync->cur ); hb_buffer_close( &next ); *buf_out = hb_buffer_init( 0 ); hb_log( "sync: reached %d frames, exiting early", pv->common->count_frames ); /* * Push through any subtitle EOFs in case they were not synced through. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); if( subtitle->config.dest == PASSTHRUSUB ) { if( subtitle->source == VOBSUB ) hb_fifo_push( subtitle->fifo_sync, hb_buffer_init( 0 ) ); else hb_fifo_push( subtitle->fifo_out, hb_buffer_init( 0 ) ); } } return HB_WORK_DONE; } hb_lock( pv->common->mutex ); start = cur->start - pv->common->video_pts_slip; hb_unlock( pv->common->mutex ); /* Check for end of point-to-point pts encoding */ if( job->pts_to_stop && sync->next_start >= job->pts_to_stop ) { // Drop an empty buffer into our output to ensure that things // get flushed all the way out. hb_log( "sync: reached pts %"PRId64", exiting early", start ); hb_buffer_close( &sync->cur ); hb_buffer_close( &next ); *buf_out = hb_buffer_init( 0 ); /* * Push through any subtitle EOFs in case they were not synced through. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); if( subtitle->config.dest == PASSTHRUSUB ) { if( subtitle->source == VOBSUB ) hb_fifo_push( subtitle->fifo_sync, hb_buffer_init( 0 ) ); else hb_fifo_push( subtitle->fifo_out, hb_buffer_init( 0 ) ); } } return HB_WORK_DONE; } if( sync->first_frame ) { /* This is our first frame */ if ( start > 0 ) { /* * The first pts from a dvd should always be zero but * can be non-zero with a transport or program stream since * we're not guaranteed to start on an IDR frame. If we get * a non-zero initial PTS extend its duration so it behaves * as if it started at zero so that our audio timing will * be in sync. */ hb_log( "sync: first pts is %"PRId64, start ); start = 0; } sync->first_frame = 0; } /* * since the first frame is always 0 and the upstream reader code * is taking care of adjusting for pts discontinuities, we just have * to deal with the next frame's start being in the past. This can * happen when the PTS is adjusted after data loss but video frame * reordering causes some frames with the old clock to appear after * the clock change. This creates frames that overlap in time which * looks to us like time going backward. The downstream muxing code * can deal with overlaps of up to a frame time but anything larger * we handle by dropping frames here. */ if ( next_start - start <= 0 ) { if ( sync->first_drop == 0 ) { sync->first_drop = next_start; } ++sync->drop_count; if ( next->new_chap ) { // don't drop a chapter mark when we drop the buffer sync->chap_mark = next->new_chap; } hb_buffer_close( &next ); return HB_WORK_OK; } if ( sync->first_drop ) { hb_log( "sync: video time didn't advance - dropped %d frames " "(delta %d ms, current %"PRId64", next %"PRId64", dur %d)", sync->drop_count, (int)( start - sync->first_drop ) / 90, start, next_start, (int)( next_start - start ) ); sync->first_drop = 0; sync->drop_count = 0; } /* * Track the video sequence number locally so that we can sync the audio * to it using the sequence number as well as the PTS. */ sync->video_sequence = cur->sequence; /* Process subtitles that apply to this video frame */ // NOTE: There is no logic in either subtitle-sync algorithm that waits for the // subtitle-decoder if it is lagging behind the video-decoder. // // Therefore there is the implicit assumption that the subtitle-decoder // is always faster than the video-decoder. This assumption is definitely // incorrect in some cases where the SSA subtitle decoder is used. // Enable the SUBSYNC_VERBOSE_TIMING flag below to debug. #define SUBSYNC_ALGORITHM_SIMULTANEOUS 1 #define SUBSYNC_ALGORITHM_CLASSIC 0 /* * Enables logging of three kinds of events: * SUB***: Subtitle received by sync object * SUB+++: Subtitle now shown * SUB---: Subtitle now hidden and disposed * * Lead times on SUB*** events should be positive. * Negative lead times lead to lag times on SUB+++ or the complete drop of a subtitle. * Lag times on SUB+++ and SUB--- should be small positive values in the 0-40ms range. */ #define SUBSYNC_VERBOSE_TIMING 0 #if SUBSYNC_ALGORITHM_SIMULTANEOUS #define sub_list sync->sub_list /* * 1. Find all subtitles that need to be burned into the current video frame * and attach them to the frame. * 2. Find all subtitles that need to be passed thru and do so immediately. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { subtitle = hb_list_item( job->list_subtitle, i ); // If this subtitle track's packets are to be passed thru, do so immediately if( subtitle->config.dest == PASSTHRUSUB ) { while ( ( sub = hb_fifo_get( subtitle->fifo_raw ) ) != NULL ) { if ( subtitle->source == VOBSUB ) { hb_fifo_push( subtitle->fifo_sync, sub ); } else { hb_fifo_push( subtitle->fifo_out, sub ); } } } // If this subtitle track's packets are to be rendered, identify the // packets that need to be rendered on the current video frame else if( subtitle->config.dest == RENDERSUB ) { // Migrate subtitles from 'subtitle->fifo_raw' to 'sub_list' immediately. // Note that the size of 'sub_list' is unbounded. while ( ( sub = hb_fifo_see( subtitle->fifo_raw ) ) != NULL ) { sub = hb_fifo_get( subtitle->fifo_raw ); // pop #if SUBSYNC_VERBOSE_TIMING printf( "\nSUB*** (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lead by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (sub->start - cur->start)/90); if (pv->common->video_pts_slip) { printf( " VIDEO-LAG: %"PRId64"\n", pv->common->video_pts_slip ); } #endif // Prepend to sub_list hb_buffer_t *sub_list_next = sub_list; sub_list = sub; sub_list->next = sub_list_next; } hb_buffer_t *last_sub = NULL; for ( sub = sub_list; sub != NULL; ) { // NOTE: Strictly speaking this sequence check is probably unnecessary. // It is a holdover behavior inherited from the classic subsync algorithm. if ( sub->sequence > cur->sequence ) { // Subtitle sequence in the future // (Keep the subtitle in the stream) last_sub = sub; sub = sub->next; continue; } if ( cur->start < sub->start ) { // Subtitle starts in the future // (Keep the subtitle in the stream) last_sub = sub; sub = sub->next; continue; } else { // Subtitle starts in the past... if ( cur->start < sub->stop ) { // Subtitle starts in the past and finishes in the future // Attach a copy of the subtitle packet to the current video packet // to be burned in by the 'render' work-object. // (Can't just alias it because we don't know when the 'render' // work-object will dispose of it.) hb_buffer_t * old_sublist_head = cur->sub; cur->sub = copy_subtitle( sub ); cur->sub->next = old_sublist_head; #if SUBSYNC_VERBOSE_TIMING if (!(sub->new_chap & 0x01)) { printf( "\nSUB+++ (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lag by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (cur->start - sub->start)/90 ); if (pv->common->video_pts_slip) { printf( " VIDEO-LAG: %"PRId64"\n", pv->common->video_pts_slip ); } sub->new_chap |= 0x01; } #endif // (Keep the subtitle in the stream) last_sub = sub; sub = sub->next; continue; } else { // Subtitle starts in the past and has already finished #if SUBSYNC_VERBOSE_TIMING printf( "\nSUB--- (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lag by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (cur->start - sub->stop)/90 ); if (pv->common->video_pts_slip) { printf( " VIDEO-LAG: %"PRId64"\n", pv->common->video_pts_slip ); } #endif // Remove it from the stream... if (last_sub != NULL) { last_sub->next = sub->next; } if (sub_list == sub) { sub_list = sub->next; } // ...and trash it hb_buffer_t *next_sub = sub->next; // XXX: Prevent hb_buffer_close from killing the whole list // before we finish iterating over it sub->next = NULL; hb_buffer_t * subpicture_list = sub; hb_buffer_t * subpicture; hb_buffer_t * subpicture_next; for ( subpicture = subpicture_list; subpicture; subpicture = subpicture_next ) { subpicture_next = subpicture->next_subpicture; hb_buffer_close( &subpicture ); } // (last_sub remains the same) sub = next_sub; continue; } } } } } // end subtitles #undef sub_list #elif SUBSYNC_ALGORITHM_CLASSIC // NOTE: This algorithm does not correctly support the simultaneous display of temporally overlapping subtitles. /* * Look for a subtitle for this frame. * * If found then it will be tagged onto a video buffer of the correct time and * sent in to the render pipeline. This only needs to be done for VOBSUBs which * get rendered, other types of subtitles can just sit in their raw_queue until * delt with at muxing. */ for( i = 0; i < hb_list_count( job->list_subtitle ); i++) { int64_t sub_start, sub_stop, duration; subtitle = hb_list_item( job->list_subtitle, i ); /* * Rewrite timestamps on subtitles that need it (on raw queue). */ // NOTE: It's probably fine to use this logic for passthru VOBSUBs as well, // but I am currently preserving backwards compatibility with the old // VOBSUB behavior, which uses the more complex logic following this if-statement. if( subtitle->config.dest == PASSTHRUSUB && subtitle->source != VOBSUB ) { /* * Rewrite timestamps on subtitles that came from Closed Captions * since they are using the MPEG2 timestamps. */ while( ( sub = hb_fifo_see( subtitle->fifo_raw ) ) ) { hb_lock( pv->common->mutex ); sub_start = sub->start - pv->common->video_pts_slip; hb_unlock( pv->common->mutex ); duration = sub->stop - sub->start; sub_stop = sub_start + duration; /* * Rewrite the timestamps as and when the video * (cur->start) reaches the same timestamp as a * closed caption (sub->start). * * What about discontinuity boundaries - not delt * with here - Van? * * Bypass the sync fifo altogether. */ if( sub->size <= 0 ) { sub = hb_fifo_get( subtitle->fifo_raw ); hb_fifo_push( subtitle->fifo_out, sub ); sub = NULL; break; } else { /* * Sync the subtitles to the incoming video, and use * the matching converted video timestamp. * * Note that it doesn't appear that we need to convert * timestamps, I guess that they were already correct, * so just push them through for rendering. * */ if( sub_start <= start ) { sub = hb_fifo_get( subtitle->fifo_raw ); sub->start = sub_start; sub->stop = sub_stop; hb_fifo_push( subtitle->fifo_out, sub ); } else { // sub too early. Leave it in the fifo. sub = NULL; break; } } } continue; } // For rendered subtitles (and, for backward compatibility, passthru VOBSUBs), // delay pushing subtitle packets through the pipeline until the video catches up if( subtitle->config.dest == RENDERSUB || subtitle->source == VOBSUB ) { hb_buffer_t * sub2; while( ( sub = hb_fifo_see( subtitle->fifo_raw ) ) ) { if( sub->size == 0 ) { /* * EOF, pass it through immediately. */ break; } hb_lock( pv->common->mutex ); sub_start = sub->start - pv->common->video_pts_slip; hb_unlock( pv->common->mutex ); duration = sub->stop - sub->start; sub_stop = sub_start + duration; /* If two DVD subtitles overlap, make the first one stop when the second one starts */ // TODO: Consider removing this entirely. Currently retained // to preserve old DVD subtitle behavior. if ( subtitle->source == VOBSUB ) { sub2 = hb_fifo_see2( subtitle->fifo_raw ); if( sub2 && sub->stop > sub2->start ) { sub->stop = sub2->start; } } // hb_log("0x%x: video seq: %"PRId64" subtitle sequence: %"PRId64, // sub, cur->sequence, sub->sequence); if( sub->sequence > cur->sequence ) { /* * The video is behind where we are, so wait until * it catches up to the same reader point on the * DVD. Then our PTS should be in the same region * as the video. */ sub = NULL; break; } #if SUBSYNC_VERBOSE_TIMING if (!(sub->new_chap & 0x02)) { printf( "\nSUB*** (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lead by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (sub->start - cur->start)/90); sub->new_chap |= 0x02; } #endif if( sub_stop > start ) { // CONDITION: cur->start < sub->stop /* * The stop time is in the future, so fall through * and we'll deal with it in the next block of * code. */ /* * There is a valid subtitle, is it time to display it? */ if( sub_stop > sub_start) { // CONDITION: {cur->start, sub->start} < sub->stop /* * Normal subtitle which ends after it starts, * check to see that the current video is between * the start and end. */ if( start > sub_start && start < sub_stop ) { // CONDITION: sub->start < cur->start < sub->stop /* * We should be playing this, so leave the * subtitle in place. * * fall through to display */ } else { // CONDITION: cur->start < sub->start < sub->stop /* * Defer until the play point is within * the subtitle */ sub = NULL; } } else { // CONDITION: cur->start < sub->stop < sub->start /* * The end of the subtitle is less than the start, * this is a sign of a PTS discontinuity. */ if( sub_start > start ) { // CONDITION: cur->start < sub->stop < sub->start /* * we haven't reached the start time yet, or * we have jumped backwards after having * already started this subtitle. */ if( start < sub_stop ) { // CONDITION: cur->start < sub->stop < sub->start /* * We have jumped backwards and so should * continue displaying this subtitle. * * fall through to display. */ } else { // CONDITION: Mathematically impossible to get here /* * Defer until the play point is * within the subtitle */ sub = NULL; } } else { // CONDITION: Mathematically impossible to get here /* * Play this subtitle as the start is * greater than our video point. * * fall through to display/ */ } } break; } else { // CONDITION: sub->stop < cur->start #if SUBSYNC_VERBOSE_TIMING printf( "\nSUB--- (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lag by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (cur->start - sub->stop)/90 ); #endif /* * The subtitle is older than this picture, trash it */ sub = hb_fifo_get( subtitle->fifo_raw ); hb_buffer_close( &sub ); } } /* If we have a subtitle for this picture, copy it */ if( sub ) { #if SUBSYNC_VERBOSE_TIMING if (!(sub->new_chap & 0x01)) { printf( "\nSUB+++ (%"PRId64"/%"PRId64":%"PRId64") @ %"PRId64"/%"PRId64":%"PRId64" (lag by %"PRId64"ms)\n", sub->start/90, sub->start/90/1000/60, sub->start/90/1000%60, cur->start/90, cur->start/90/1000/60, cur->start/90/1000%60, (cur->start - sub->start)/90 ); sub->new_chap |= 0x01; } #endif if( sub->size > 0 ) { if( subtitle->config.dest == RENDERSUB ) { /* * Tack onto the video buffer for rendering. * * Note that there may be multiple subtitles * whose time intervals overlap which must display * on the same frame. */ hb_buffer_t * old_sublist_head = cur->sub; /* FIXME: we should avoid this memcpy */ cur->sub = copy_subtitle( sub ); cur->sub->next = old_sublist_head; cur->sub->start = sub_start; cur->sub->stop = sub_stop; // Leave the subtitle on the raw queue // (until it no longer needs to be displayed) } else { /* * Pass-Through, pop it off of the raw queue, */ sub = hb_fifo_get( subtitle->fifo_raw ); sub->start = sub_start; sub->stop = sub_stop; hb_fifo_push( subtitle->fifo_sync, sub ); } } else { /* * EOF - consume for rendered, else pass through */ if( subtitle->config.dest == RENDERSUB ) { sub = hb_fifo_get( subtitle->fifo_raw ); hb_buffer_close( &sub ); } else { sub = hb_fifo_get( subtitle->fifo_raw ); sub->start = sub_start; sub->stop = sub_stop; hb_fifo_push( subtitle->fifo_sync, sub ); } } } } } // end subtitles #else #error "Must select a subtitle sync algorithm." #endif /* * Adjust the pts of the current frame so that it's contiguous * with the previous frame. The start time of the current frame * has to be the end time of the previous frame and the stop * time has to be the start of the next frame. We don't * make any adjustments to the source timestamps other than removing * the clock offsets (which also removes pts discontinuities). * This means we automatically encode at the source's frame rate. * MP2 uses an implicit duration (frames end when the next frame * starts) but more advanced containers like MP4 use an explicit * duration. Since we're looking ahead one frame we set the * explicit stop time from the start time of the next frame. */ *buf_out = cur; sync->cur = cur = next; cur->sub = NULL; int64_t duration = next_start - start; sync->pts_skip = 0; if ( duration <= 0 ) { hb_log( "sync: invalid video duration %"PRId64", start %"PRId64", next %"PRId64"", duration, start, next_start ); } (*buf_out)->start = sync->next_start; sync->next_start += duration; (*buf_out)->stop = sync->next_start; if ( sync->chap_mark ) { // we have a pending chapter mark from a recent drop - put it on this // buffer (this may make it one frame late but we can't do any better). (*buf_out)->new_chap = sync->chap_mark; sync->chap_mark = 0; } /* Update UI */ UpdateState( w ); return HB_WORK_OK; } static hb_buffer_t * copy_subtitle( hb_buffer_t * src_list ) { hb_buffer_t * dst_list = NULL; hb_buffer_t * src; hb_buffer_t * dst; hb_buffer_t ** dst_ptr = &dst_list; for ( src = src_list, dst_ptr = &dst_list; src; src = src->next_subpicture, dst_ptr = &dst->next_subpicture ) { (*dst_ptr) = hb_buffer_init( src->size ); dst = (*dst_ptr); dst->x = src->x; dst->y = src->y; dst->width = src->width; dst->height = src->height; memcpy( dst->data, src->data, src->size ); } return dst_list; } // sync*Init does nothing because sync has a special initializer // that takes care of initializing video and all audio tracks int syncVideoInit( hb_work_object_t * w, hb_job_t * job) { return 0; } hb_work_object_t hb_sync_video = { WORK_SYNC_VIDEO, "Video Synchronization", syncVideoInit, syncVideoWork, syncVideoClose }; /*********************************************************************** * Close Audio *********************************************************************** * **********************************************************************/ void syncAudioClose( hb_work_object_t * w ) { hb_work_private_t * pv = w->private_data; hb_sync_audio_t * sync = &pv->type.audio; if( w->audio->config.out.codec == HB_ACODEC_AC3_PASS ) { free( sync->ac3_buf ); } else { src_delete( sync->state ); } hb_lock( pv->common->mutex ); if ( --pv->common->ref == 0 ) { hb_unlock( pv->common->mutex ); hb_lock_close( &pv->common->mutex ); free( pv->common ); } else { hb_unlock( pv->common->mutex ); } free( pv ); w->private_data = NULL; } int syncAudioInit( hb_work_object_t * w, hb_job_t * job) { return 0; } /*********************************************************************** * SyncAudio *********************************************************************** * **********************************************************************/ static int syncAudioWork( hb_work_object_t * w, hb_buffer_t ** buf_in, hb_buffer_t ** buf_out ) { hb_work_private_t * pv = w->private_data; hb_job_t * job = pv->job; hb_sync_audio_t * sync = &pv->type.audio; hb_buffer_t * buf; int64_t start; *buf_out = NULL; buf = *buf_in; *buf_in = NULL; /* if the next buffer is an eof send it downstream */ if ( buf->size <= 0 ) { hb_buffer_close( &buf ); *buf_out = hb_buffer_init( 0 ); pv->common->first_pts[sync->index+1] = INT64_MAX - 1; return HB_WORK_DONE; } /* Wait till we can determine the initial pts of all streams */ if( pv->common->pts_offset == INT64_MIN ) { pv->common->first_pts[sync->index+1] = buf->start; hb_lock( pv->common->mutex ); while( pv->common->pts_offset == INT64_MIN ) { // Full fifos will make us wait forever, so get the // pts offset from the available streams if full if (hb_fifo_is_full(w->fifo_in)) { getPtsOffset( w ); hb_cond_broadcast( pv->common->next_frame ); } else if ( checkPtsOffset( w ) ) hb_cond_broadcast( pv->common->next_frame ); else hb_cond_timedwait( pv->common->next_frame, pv->common->mutex, 200 ); } hb_unlock( pv->common->mutex ); } /* Wait for start frame if doing point-to-point */ hb_lock( pv->common->mutex ); start = buf->start - pv->common->audio_pts_slip; while ( !pv->common->start_found ) { if ( pv->common->audio_pts_thresh < 0 ) { // I would initialize this in hb_sync_init, but // job->pts_to_start can be modified by reader // after hb_sync_init is called. pv->common->audio_pts_thresh = job->pts_to_start; } if ( buf->start < pv->common->audio_pts_thresh ) { hb_buffer_close( &buf ); hb_unlock( pv->common->mutex ); return HB_WORK_OK; } while ( !pv->common->start_found && buf->start >= pv->common->audio_pts_thresh ) { hb_cond_timedwait( pv->common->next_frame, pv->common->mutex, 200 ); } start = buf->start - pv->common->audio_pts_slip; } if ( start < 0 ) { hb_buffer_close( &buf ); hb_unlock( pv->common->mutex ); return HB_WORK_OK; } hb_unlock( pv->common->mutex ); if( job->frame_to_stop && pv->common->count_frames >= job->frame_to_stop ) { hb_buffer_close( &buf ); *buf_out = hb_buffer_init( 0 ); return HB_WORK_DONE; } if( job->pts_to_stop && sync->next_start >= job->pts_to_stop ) { hb_buffer_close( &buf ); *buf_out = hb_buffer_init( 0 ); return HB_WORK_DONE; } if ( start - sync->next_start < 0 ) { // audio time went backwards. // If our output clock is more than a half frame ahead of the // input clock drop this frame to move closer to sync. // Otherwise drop frames until the input clock matches the output clock. if ( sync->first_drop || sync->next_start - start > 90*15 ) { // Discard data that's in the past. if ( sync->first_drop == 0 ) { sync->first_drop = sync->next_start; } ++sync->drop_count; hb_buffer_close( &buf ); return HB_WORK_OK; } } if ( sync->first_drop ) { // we were dropping old data but input buf time is now current hb_log( "sync: audio %d time went backwards %d ms, dropped %d frames " "(next %"PRId64", current %"PRId64")", w->audio->id, (int)( sync->next_start - sync->first_drop ) / 90, sync->drop_count, sync->first_drop, (int64_t)sync->next_start ); sync->first_drop = 0; sync->drop_count = 0; } if ( start - sync->next_start >= (90 * 70) ) { if ( start - sync->next_start > (90000LL * 60) ) { // there's a gap of more than a minute between the last // frame and this. assume we got a corrupted timestamp // and just drop the next buf. hb_log( "sync: %d minute time gap in audio %d - dropping buf" " start %"PRId64", next %"PRId64, (int)((start - sync->next_start) / (90000*60)), w->audio->id, start, (int64_t)sync->next_start ); hb_buffer_close( &buf ); return HB_WORK_OK; } /* * there's a gap of at least 70ms between the last * frame we processed & the next. Fill it with silence. * Or in the case of DCA, skip some frames from the * other streams. */ if( w->audio->config.out.codec == HB_ACODEC_DCA_PASS ) { hb_log( "sync: audio gap %d ms. Skipping frames. Audio %d" " start %"PRId64", next %"PRId64, (int)((start - sync->next_start) / 90), w->audio->id, start, (int64_t)sync->next_start ); hb_lock( pv->common->mutex ); pv->common->audio_pts_slip += (start - sync->next_start); pv->common->video_pts_slip += (start - sync->next_start); hb_unlock( pv->common->mutex ); *buf_out = buf; return HB_WORK_OK; } hb_log( "sync: adding %d ms of silence to audio %d" " start %"PRId64", next %"PRId64, (int)((start - sync->next_start) / 90), w->audio->id, start, (int64_t)sync->next_start ); InsertSilence( w, start - sync->next_start ); } /* * When we get here we've taken care of all the dups and gaps in the * audio stream and are ready to inject the next input frame into * the output stream. */ *buf_out = OutputAudioFrame( w->audio, buf, sync ); return HB_WORK_OK; } hb_work_object_t hb_sync_audio = { WORK_SYNC_AUDIO, "AudioSynchronization", syncAudioInit, syncAudioWork, syncAudioClose }; static void InitAudio( hb_job_t * job, hb_sync_common_t * common, int i ) { hb_work_object_t * w; hb_work_private_t * pv; hb_title_t * title = job->title; hb_sync_audio_t * sync; pv = calloc( 1, sizeof( hb_work_private_t ) ); sync = &pv->type.audio; sync->index = i; pv->job = job; pv->common = common; pv->common->ref++; pv->common->pts_count++; w = hb_get_work( WORK_SYNC_AUDIO ); w->private_data = pv; w->audio = hb_list_item( title->list_audio, i ); w->fifo_in = w->audio->priv.fifo_raw; if( w->audio->config.out.codec == HB_ACODEC_AC3_PASS || w->audio->config.out.codec == HB_ACODEC_DCA_PASS ) { w->fifo_out = w->audio->priv.fifo_out; } else { w->fifo_out = w->audio->priv.fifo_sync; } if( w->audio->config.out.codec == HB_ACODEC_AC3_PASS ) { /* Have a silent AC-3 frame ready in case we have to fill a gap */ AVCodec * codec; AVCodecContext * c; short * zeros; codec = avcodec_find_encoder( CODEC_ID_AC3 ); c = avcodec_alloc_context(); c->bit_rate = w->audio->config.in.bitrate; c->sample_rate = w->audio->config.in.samplerate; c->channels = HB_INPUT_CH_LAYOUT_GET_DISCRETE_COUNT( w->audio->config.in.channel_layout ); c->sample_fmt = AV_SAMPLE_FMT_FLT; if( hb_avcodec_open( c, codec ) < 0 ) { hb_log( "sync: avcodec_open failed" ); return; } zeros = calloc( AC3_SAMPLES_PER_FRAME * sizeof( float ) * c->channels, 1 ); sync->ac3_size = w->audio->config.in.bitrate * AC3_SAMPLES_PER_FRAME / w->audio->config.in.samplerate / 8; sync->ac3_buf = malloc( sync->ac3_size ); if( avcodec_encode_audio( c, sync->ac3_buf, sync->ac3_size, zeros ) != sync->ac3_size ) { hb_log( "sync: avcodec_encode_audio failed" ); } free( zeros ); hb_avcodec_close( c ); av_free( c ); } else { /* Initialize libsamplerate */ int error; sync->state = src_new( SRC_SINC_MEDIUM_QUALITY, HB_AMIXDOWN_GET_DISCRETE_CHANNEL_COUNT( w->audio->config.out.mixdown), &error ); sync->data.end_of_input = 0; } hb_list_add( job->list_work, w ); } static hb_buffer_t * OutputAudioFrame( hb_audio_t *audio, hb_buffer_t *buf, hb_sync_audio_t *sync ) { int64_t start = (int64_t)sync->next_start; double duration = buf->stop - buf->start; if( audio->config.in.samplerate == audio->config.out.samplerate || audio->config.out.codec == HB_ACODEC_AC3_PASS || audio->config.out.codec == HB_ACODEC_DCA_PASS ) { /* * If we don't have to do sample rate conversion or this audio is * pass-thru just send the input buffer downstream after adjusting * its timestamps to make the output stream continuous. */ } else { /* Not pass-thru - do sample rate conversion */ int count_in, count_out; hb_buffer_t * buf_raw = buf; int channel_count = HB_AMIXDOWN_GET_DISCRETE_CHANNEL_COUNT(audio->config.out.mixdown) * sizeof( float ); count_in = buf_raw->size / channel_count; /* * When using stupid rates like 44.1 there will always be some * truncation error. E.g., a 1536 sample AC3 frame will turn into a * 1536*44.1/48.0 = 1411.2 sample frame. If we just truncate the .2 * the error will build up over time and eventually the audio will * substantially lag the video. libsamplerate will keep track of the * fractional sample & give it to us when appropriate if we give it * an extra sample of space in the output buffer. */ count_out = ( duration * audio->config.out.samplerate ) / 90000 + 1; sync->data.input_frames = count_in; sync->data.output_frames = count_out; sync->data.src_ratio = (double)audio->config.out.samplerate / (double)audio->config.in.samplerate; buf = hb_buffer_init( count_out * channel_count ); sync->data.data_in = (float *) buf_raw->data; sync->data.data_out = (float *) buf->data; if( src_process( sync->state, &sync->data ) ) { /* XXX If this happens, we're screwed */ hb_log( "sync: audio %d src_process failed", audio->id ); } hb_buffer_close( &buf_raw ); buf->size = sync->data.output_frames_gen * channel_count; duration = (double)( sync->data.output_frames_gen * 90000 ) / audio->config.out.samplerate; } buf->frametype = HB_FRAME_AUDIO; buf->start = start; sync->next_start += duration; buf->stop = (int64_t)sync->next_start; return buf; } static void InsertSilence( hb_work_object_t * w, int64_t duration ) { hb_work_private_t * pv = w->private_data; hb_sync_audio_t *sync = &pv->type.audio; hb_buffer_t *buf; hb_fifo_t *fifo; // to keep pass-thru and regular audio in sync we generate silence in // AC3 frame-sized units. If the silence duration isn't an integer multiple // of the AC3 frame duration we will truncate or round up depending on // which minimizes the timing error. const int frame_dur = ( 90000 * AC3_SAMPLES_PER_FRAME ) / w->audio->config.in.samplerate; int frame_count = ( duration + (frame_dur >> 1) ) / frame_dur; while ( --frame_count >= 0 ) { if( w->audio->config.out.codec == HB_ACODEC_AC3_PASS ) { buf = hb_buffer_init( sync->ac3_size ); buf->start = sync->next_start; buf->stop = buf->start + frame_dur; memcpy( buf->data, sync->ac3_buf, buf->size ); fifo = w->audio->priv.fifo_out; } else { buf = hb_buffer_init( AC3_SAMPLES_PER_FRAME * sizeof( float ) * HB_AMIXDOWN_GET_DISCRETE_CHANNEL_COUNT( w->audio->config.out.mixdown) ); buf->start = sync->next_start; buf->stop = buf->start + frame_dur; memset( buf->data, 0, buf->size ); fifo = w->audio->priv.fifo_sync; } buf = OutputAudioFrame( w->audio, buf, sync ); hb_fifo_push( fifo, buf ); } } static void UpdateState( hb_work_object_t * w ) { hb_work_private_t * pv = w->private_data; hb_sync_video_t * sync = &pv->type.video; hb_state_t state; if( !pv->common->count_frames ) { sync->st_first = hb_get_date(); pv->job->st_pause_date = -1; pv->job->st_paused = 0; } pv->common->count_frames++; if( hb_get_date() > sync->st_dates[3] + 1000 ) { memmove( &sync->st_dates[0], &sync->st_dates[1], 3 * sizeof( uint64_t ) ); memmove( &sync->st_counts[0], &sync->st_counts[1], 3 * sizeof( uint64_t ) ); sync->st_dates[3] = hb_get_date(); sync->st_counts[3] = pv->common->count_frames; } #define p state.param.working state.state = HB_STATE_WORKING; p.progress = (float) pv->common->count_frames / (float) sync->count_frames_max; if( p.progress > 1.0 ) { p.progress = 1.0; } p.rate_cur = 1000.0 * (float) ( sync->st_counts[3] - sync->st_counts[0] ) / (float) ( sync->st_dates[3] - sync->st_dates[0] ); if( hb_get_date() > sync->st_first + 4000 ) { int eta; p.rate_avg = 1000.0 * (float) sync->st_counts[3] / (float) ( sync->st_dates[3] - sync->st_first - pv->job->st_paused); eta = (float) ( sync->count_frames_max - sync->st_counts[3] ) / p.rate_avg; p.hours = eta / 3600; p.minutes = ( eta % 3600 ) / 60; p.seconds = eta % 60; } else { p.rate_avg = 0.0; p.hours = -1; p.minutes = -1; p.seconds = -1; } #undef p hb_set_state( pv->job->h, &state ); } static void UpdateSearchState( hb_work_object_t * w, int64_t start ) { hb_work_private_t * pv = w->private_data; hb_sync_video_t * sync = &pv->type.video; hb_state_t state; uint64_t now; double avg; now = hb_get_date(); if( !pv->common->count_frames ) { sync->st_first = now; pv->job->st_pause_date = -1; pv->job->st_paused = 0; } pv->common->count_frames++; #define p state.param.working state.state = HB_STATE_SEARCHING; if ( pv->job->frame_to_start ) p.progress = (float) pv->common->count_frames / (float) pv->job->frame_to_start; else if ( pv->job->pts_to_start ) p.progress = (float) start / (float) pv->job->pts_to_start; else p.progress = 0; if( p.progress > 1.0 ) { p.progress = 1.0; } if (now > sync->st_first) { int eta; if ( pv->job->frame_to_start ) { avg = 1000.0 * (double)pv->common->count_frames / (now - sync->st_first); eta = ( pv->job->frame_to_start - pv->common->count_frames ) / avg; } else if ( pv->job->pts_to_start ) { avg = 1000.0 * (double)start / (now - sync->st_first); eta = ( pv->job->pts_to_start - start ) / avg; } p.hours = eta / 3600; p.minutes = ( eta % 3600 ) / 60; p.seconds = eta % 60; } else { p.rate_avg = 0.0; p.hours = -1; p.minutes = -1; p.seconds = -1; } #undef p hb_set_state( pv->job->h, &state ); } static void getPtsOffset( hb_work_object_t * w ) { hb_work_private_t * pv = w->private_data; int i ; int64_t first_pts = INT64_MAX; for( i = 0; i < pv->common->pts_count; i++ ) { if ( pv->common->first_pts[i] < first_pts ) first_pts = pv->common->first_pts[i]; } pv->common->video_pts_slip = pv->common->audio_pts_slip = pv->common->pts_offset = first_pts; return; } static int checkPtsOffset( hb_work_object_t * w ) { hb_work_private_t * pv = w->private_data; int i ; for( i = 0; i < pv->common->pts_count; i++ ) { if ( pv->common->first_pts[i] == INT64_MAX ) return 0; } getPtsOffset( w ); return 1; }