diff options
author | Rodeo <[email protected]> | 2013-08-22 20:34:44 +0000 |
---|---|---|
committer | Rodeo <[email protected]> | 2013-08-22 20:34:44 +0000 |
commit | 3326f988806a5decae025727784a19c8cc223833 (patch) | |
tree | cd75adb1975d223d7a0fd43a31030a78939e73d3 /contrib/ffmpeg/A00-qsv.patch | |
parent | d41905d539046445e1b81499ff7bd04d170c91d4 (diff) |
Big merge, QSV to trunk: part 2 (new files).
git-svn-id: svn://svn.handbrake.fr/HandBrake/trunk@5738 b64f7644-9d1e-0410-96f1-a4d463321fa5
Diffstat (limited to 'contrib/ffmpeg/A00-qsv.patch')
-rw-r--r-- | contrib/ffmpeg/A00-qsv.patch | 2285 |
1 files changed, 2285 insertions, 0 deletions
diff --git a/contrib/ffmpeg/A00-qsv.patch b/contrib/ffmpeg/A00-qsv.patch new file mode 100644 index 000000000..908033617 --- /dev/null +++ b/contrib/ffmpeg/A00-qsv.patch @@ -0,0 +1,2285 @@ +diff -Naur ../../libav-v9.6/configure ./configure +--- ../../libav-v9.6/configure 2013-05-12 08:39:07.000000000 +0200 ++++ ./configure 2013-08-14 10:48:00.520497159 +0200 +@@ -133,6 +133,7 @@ + --enable-vaapi enable VAAPI code + --enable-vda enable VDA code + --enable-vdpau enable VDPAU code ++ --enable-qsv enable QSV code + + Individual component options: + --disable-everything disable all components listed below +@@ -1076,6 +1077,7 @@ + vaapi + vda + vdpau ++ qsv + version3 + xmm_clobber_test + x11grab +@@ -1629,6 +1631,7 @@ + wmv3_dxva2_hwaccel_select="vc1_dxva2_hwaccel" + wmv3_vaapi_hwaccel_select="vc1_vaapi_hwaccel" + wmv3_vdpau_decoder_select="vc1_vdpau_decoder" ++h264_qsv_decoder_select="qsv h264_decoder" + + # parsers + h264_parser_select="error_resilience golomb h264dsp h264pred mpegvideo" +@@ -3584,6 +3587,12 @@ + check_cpp_condition vdpau/vdpau.h "defined VDP_DECODER_PROFILE_MPEG4_PART2_ASP" || + { echolog "Please upgrade to libvdpau >= 0.2 if you would like vdpau support." && disable vdpau; } + fi ++if enabled qsv; then ++ disable qsv ++ check_header msdk/mfxvideo.h && enable qsv ++else ++ disable qsv ++fi + + enabled debug && add_cflags -g"$debuglevel" && add_asflags -g"$debuglevel" + +@@ -3795,6 +3804,7 @@ + echo "libdxva2 enabled ${dxva2-no}" + echo "libva enabled ${vaapi-no}" + echo "libvdpau enabled ${vdpau-no}" ++echo "libqsv enabled ${qsv-no}" + echo "AVISynth enabled ${avisynth-no}" + echo "frei0r enabled ${frei0r-no}" + echo "gnutls enabled ${gnutls-no}" +diff -Naur ../../libav-v9.6/libavcodec/allcodecs.c ./libavcodec/allcodecs.c +--- ../../libav-v9.6/libavcodec/allcodecs.c 2013-05-12 08:39:07.000000000 +0200 ++++ ./libavcodec/allcodecs.c 2013-08-14 10:48:00.520497159 +0200 +@@ -143,6 +143,7 @@ + REGISTER_DECODER(H263I, h263i); + REGISTER_ENCODER(H263P, h263p); + REGISTER_DECODER(H264, h264); ++ REGISTER_DECODER(H264_QSV, h264_qsv); + REGISTER_DECODER(H264_VDPAU, h264_vdpau); + REGISTER_ENCDEC (HUFFYUV, huffyuv); + REGISTER_DECODER(IDCIN, idcin); +diff -Naur ../../libav-v9.6/libavcodec/Makefile ./libavcodec/Makefile +--- ../../libav-v9.6/libavcodec/Makefile 2013-05-12 08:39:07.000000000 +0200 ++++ ./libavcodec/Makefile 2013-08-14 10:48:00.521497282 +0200 +@@ -10,6 +10,7 @@ + vdpau.h \ + version.h \ + xvmc.h \ ++ qsv.h \ + + OBJS = allcodecs.o \ + audioconvert.o \ +@@ -196,6 +197,7 @@ + h264_loopfilter.o h264_direct.o \ + cabac.o h264_sei.o h264_ps.o \ + h264_refs.o h264_cavlc.o h264_cabac.o ++OBJS-$(CONFIG_H264_QSV_DECODER) += qsv_h264.o qsv.o + OBJS-$(CONFIG_H264_DXVA2_HWACCEL) += dxva2_h264.o + OBJS-$(CONFIG_H264_VAAPI_HWACCEL) += vaapi_h264.o + OBJS-$(CONFIG_H264_VDA_HWACCEL) += vda_h264.o +diff -Naur ../../libav-v9.6/libavcodec/qsv.c ./libavcodec/qsv.c +--- ../../libav-v9.6/libavcodec/qsv.c 1970-01-01 01:00:00.000000000 +0100 ++++ ./libavcodec/qsv.c 2013-08-19 21:32:01.704244071 +0200 +@@ -0,0 +1,646 @@ ++/* ********************************************************************* *\ ++ ++Copyright (C) 2013 Intel Corporation. All rights reserved. ++ ++Redistribution and use in source and binary forms, with or without ++modification, are permitted provided that the following conditions are met: ++- Redistributions of source code must retain the above copyright notice, ++this list of conditions and the following disclaimer. ++- Redistributions in binary form must reproduce the above copyright notice, ++this list of conditions and the following disclaimer in the documentation ++and/or other materials provided with the distribution. ++- Neither the name of Intel Corporation nor the names of its contributors ++may be used to endorse or promote products derived from this software ++without specific prior written permission. ++ ++THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION "AS IS" AND ANY EXPRESS OR ++IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES ++OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. ++IN NO EVENT SHALL INTEL CORPORATION BE LIABLE FOR ANY DIRECT, INDIRECT, ++INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT ++NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, ++DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY ++THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF ++THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ ++\* ********************************************************************* */ ++ ++#include "qsv.h" ++ ++#include "avcodec.h" ++#include "internal.h" ++ ++int av_qsv_get_free_encode_task(av_qsv_list * tasks) ++{ ++ int ret = MFX_ERR_NOT_FOUND; ++ int i = 0; ++ if (tasks) ++ for (i = 0; i < av_qsv_list_count(tasks); i++) { ++ av_qsv_task *task = av_qsv_list_item(tasks, i); ++ if (task->stage && task->stage->out.sync) ++ if (!(*task->stage->out.sync->p_sync)) { ++ ret = i; ++ break; ++ } ++ } ++ return ret; ++} ++ ++int av_qsv_get_free_sync(av_qsv_space * space, av_qsv_context * qsv) ++{ ++ int ret = -1; ++ int counter = 0; ++ ++ while (1) { ++ for (int i = 0; i < space->sync_num; i++) { ++ if (!(*(space->p_syncp[i]->p_sync)) && ++ 0 == space->p_syncp[i]->in_use ) { ++ if (i > space->sync_num_max_used) ++ space->sync_num_max_used = i; ++ ff_qsv_atomic_inc(&space->p_syncp[i]->in_use); ++ return i; ++ } ++ } ++#if HAVE_THREADS ++ if (++counter >= AV_QSV_REPEAT_NUM_DEFAULT) { ++#endif ++ av_log(NULL, AV_LOG_FATAL, "not enough to have %d sync point(s) allocated\n", ++ space->sync_num); ++ break; ++#if HAVE_THREADS ++ } ++ av_qsv_sleep(5); ++#endif ++ } ++ return ret; ++} ++ ++int av_qsv_get_free_surface(av_qsv_space * space, av_qsv_context * qsv, ++ mfxFrameInfo * info, av_qsv_split part) ++{ ++ int ret = -1; ++ int from = 0; ++ int up = space->surface_num; ++ int counter = 0; ++ ++ while (1) { ++ from = 0; ++ up = space->surface_num; ++ if (part == QSV_PART_LOWER) ++ up /= 2; ++ if (part == QSV_PART_UPPER) ++ from = up / 2; ++ ++ for (int i = from; i < up; i++) { ++ if (0 == space->p_surfaces[i]->Data.Locked) { ++ memcpy(&(space->p_surfaces[i]->Info), info, ++ sizeof(mfxFrameInfo)); ++ if (i > space->surface_num_max_used) ++ space->surface_num_max_used = i; ++ return i; ++ } ++ } ++#if HAVE_THREADS ++ if (++counter >= AV_QSV_REPEAT_NUM_DEFAULT) { ++#endif ++ av_log(NULL, AV_LOG_FATAL, ++ "not enough to have %d surface(s) allocated\n", up); ++ break; ++#if HAVE_THREADS ++ } ++ av_qsv_sleep(5); ++#endif ++ } ++ return ret; ++} ++ ++int ff_qsv_is_surface_in_pipe(mfxFrameSurface1 * p_surface, av_qsv_context * qsv) ++{ ++ int ret = 0; ++ int a, b,i; ++ av_qsv_list *list = 0; ++ av_qsv_stage *stage = 0; ++ ++ if (!p_surface) ++ return ret; ++ if (!qsv->pipes) ++ return ret; ++ ++ for (a = 0; a < av_qsv_list_count(qsv->pipes); a++) { ++ list = av_qsv_list_item(qsv->pipes, a); ++ for (b = 0; b < av_qsv_list_count(list); b++) { ++ stage = av_qsv_list_item(list, b); ++ if (p_surface == stage->out.p_surface) ++ return (stage->type << 16) | 2; ++ if (p_surface == stage->in.p_surface) ++ return (stage->type << 16) | 1; ++ } ++ } ++ return ret; ++} ++ ++int ff_qsv_is_sync_in_pipe(mfxSyncPoint * sync, av_qsv_context * qsv) ++{ ++ int ret = 0; ++ int a, b; ++ av_qsv_list *list = 0; ++ av_qsv_stage *stage = 0; ++ ++ if (!sync) ++ return ret; ++ if (!qsv->pipes) ++ return ret; ++ ++ for (a = 0; a < av_qsv_list_count(qsv->pipes); a++) { ++ list = av_qsv_list_item(qsv->pipes, a); ++ for (b = 0; b < av_qsv_list_count(list); b++) { ++ stage = av_qsv_list_item(list, b); ++ if (sync == stage->out.sync->p_sync) { ++ return 1; ++ } ++ } ++ } ++ return ret; ++} ++ ++av_qsv_stage *av_qsv_stage_init(void) ++{ ++ av_qsv_stage *stage = av_mallocz(sizeof(av_qsv_stage)); ++ return stage; ++} ++ ++void av_qsv_stage_clean(av_qsv_stage ** stage) ++{ ++ if ((*stage)->out.sync) { ++ if ((*stage)->out.sync->p_sync) ++ *(*stage)->out.sync->p_sync = 0; ++ if ((*stage)->out.sync->in_use > 0) ++ ff_qsv_atomic_dec(&(*stage)->out.sync->in_use); ++ (*stage)->out.sync = 0; ++ } ++ if ((*stage)->out.p_surface) { ++ (*stage)->out.p_surface = 0; ++ ++ } ++ if ((*stage)->in.p_surface) { ++ (*stage)->in.p_surface = 0; ++ } ++ ++ av_freep(stage); ++} ++ ++void av_qsv_add_context_usage(av_qsv_context * qsv, int is_threaded) ++{ ++ int is_active = 0; ++#if HAVE_THREADS ++ int mut_ret = 0; ++#endif ++ ++ is_active = ff_qsv_atomic_inc(&qsv->is_context_active); ++ if (is_active == 1) { ++ memset(&qsv->mfx_session, 0, sizeof(mfxSession)); ++ av_qsv_pipe_list_create(&qsv->pipes, is_threaded); ++ ++ qsv->dts_seq = av_qsv_list_init(is_threaded); ++ ++#if HAVE_THREADS ++ if (is_threaded) { ++ qsv->qts_seq_mutex = av_mallocz(sizeof(pthread_mutex_t)); ++ if (qsv->qts_seq_mutex){ ++ mut_ret = pthread_mutex_init(qsv->qts_seq_mutex, NULL); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_init issue[%d] at %s\n",mut_ret,__FUNCTION__); ++ } ++ ++ } else ++#endif ++ qsv->qts_seq_mutex = 0; ++ } ++} ++ ++int av_qsv_context_clean(av_qsv_context * qsv) ++{ ++ int is_active = 0; ++ mfxStatus sts = MFX_ERR_NONE; ++#if HAVE_THREADS ++ int mut_ret = 0; ++#endif ++ ++ is_active = ff_qsv_atomic_dec(&qsv->is_context_active); ++ ++ // spaces would have to be cleaned on the own, ++ // here we care about the rest, common stuff ++ if (is_active == 0) { ++ ++ if (qsv->dts_seq) { ++ while (av_qsv_list_count(qsv->dts_seq)) ++ av_qsv_dts_pop(qsv); ++ ++ av_qsv_list_close(&qsv->dts_seq); ++ } ++#if HAVE_THREADS ++ if (qsv->qts_seq_mutex) { ++ mut_ret = pthread_mutex_destroy(qsv->qts_seq_mutex); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_destroy issue[%d] at %s\n", mut_ret,__FUNCTION__); ++#endif ++ qsv->qts_seq_mutex = 0; ++#if HAVE_THREADS ++ } ++#endif ++ ++ if (qsv->pipes) ++ av_qsv_pipe_list_clean(&qsv->pipes); ++ ++ if (qsv->mfx_session) { ++ sts = MFXClose(qsv->mfx_session); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ qsv->mfx_session = 0; ++ } ++ } ++ return 0; ++} ++ ++void av_qsv_pipe_list_create(av_qsv_list ** list, int is_threaded) ++{ ++ if (!*list) ++ *list = av_qsv_list_init(is_threaded); ++} ++ ++void av_qsv_pipe_list_clean(av_qsv_list ** list) ++{ ++ av_qsv_list *stage; ++ int i = 0; ++ if (*list) { ++ for (i = av_qsv_list_count(*list); i > 0; i--) { ++ stage = av_qsv_list_item(*list, i - 1); ++ av_qsv_flush_stages(*list, &stage); ++ } ++ av_qsv_list_close(list); ++ } ++} ++ ++void av_qsv_add_stagee(av_qsv_list ** list, av_qsv_stage * stage, int is_threaded) ++{ ++ if (!*list) ++ *list = av_qsv_list_init(is_threaded); ++ av_qsv_list_add(*list, stage); ++} ++ ++av_qsv_stage *av_qsv_get_last_stage(av_qsv_list * list) ++{ ++ av_qsv_stage *stage = 0; ++ int size = 0; ++ ++ av_qsv_list_lock(list); ++ size = av_qsv_list_count(list); ++ if (size > 0) ++ stage = av_qsv_list_item(list, size - 1); ++ av_qsv_list_unlock(list); ++ ++ return stage; ++} ++ ++void av_qsv_flush_stages(av_qsv_list * list, av_qsv_list ** item) ++{ ++ int i = 0; ++ int x = 0; ++ av_qsv_stage *stage = 0; ++ av_qsv_list *to_remove_list = 0; ++ av_qsv_list *to_remove_atom_list = 0; ++ av_qsv_list *to_remove_atom = 0; ++ ++ for (i = 0; i < av_qsv_list_count(*item); i++) { ++ stage = av_qsv_list_item(*item, i); ++ if(stage->pending){ ++ if(!to_remove_list) ++ to_remove_list = av_qsv_list_init(0); ++ av_qsv_list_add(to_remove_list, stage->pending); ++ } ++ av_qsv_stage_clean(&stage); ++ // should actually remove from the list but ok... ++ } ++ av_qsv_list_rem(list, *item); ++ av_qsv_list_close(item); ++ ++ if(to_remove_list){ ++ for (i = av_qsv_list_count(to_remove_list); i > 0; i--){ ++ to_remove_atom_list = av_qsv_list_item(to_remove_list, i-1); ++ for (x = av_qsv_list_count(to_remove_atom_list); x > 0; x--){ ++ to_remove_atom = av_qsv_list_item(to_remove_atom_list, x-1); ++ av_qsv_flush_stages(list,&to_remove_atom); ++ } ++ } ++ av_qsv_list_close(&to_remove_list); ++ } ++} ++ ++av_qsv_list *av_qsv_pipe_by_stage(av_qsv_list * list, av_qsv_stage * stage) ++{ ++ av_qsv_list *item = 0; ++ av_qsv_stage *cur_stage = 0; ++ int i = 0; ++ int a = 0; ++ for (i = 0; i < av_qsv_list_count(list); i++) { ++ item = av_qsv_list_item(list, i); ++ for (a = 0; a < av_qsv_list_count(item); a++) { ++ cur_stage = av_qsv_list_item(item, a); ++ if (cur_stage == stage) ++ return item; ++ } ++ } ++ return 0; ++} ++ ++// no duplicate of the same value, if end == 0 : working over full length ++void av_qsv_dts_ordered_insert(av_qsv_context * qsv, int start, int end, ++ int64_t dts, int iter) ++{ ++ av_qsv_dts *cur_dts = 0; ++ av_qsv_dts *new_dts = 0; ++ int i = 0; ++#if HAVE_THREADS ++ int mut_ret = 0; ++#endif ++ ++ ++#if HAVE_THREADS ++ if (iter == 0 && qsv->qts_seq_mutex){ ++ mut_ret = pthread_mutex_lock(qsv->qts_seq_mutex); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_lock issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ } ++#endif ++ ++ if (end == 0) ++ end = av_qsv_list_count(qsv->dts_seq); ++ ++ if (end <= start) { ++ new_dts = av_mallocz(sizeof(av_qsv_dts)); ++ if( new_dts ) { ++ new_dts->dts = dts; ++ av_qsv_list_add(qsv->dts_seq, new_dts); ++ } ++ } else ++ for (i = end; i > start; i--) { ++ cur_dts = av_qsv_list_item(qsv->dts_seq, i - 1); ++ if (cur_dts->dts < dts) { ++ new_dts = av_mallocz(sizeof(av_qsv_dts)); ++ if( new_dts ) { ++ new_dts->dts = dts; ++ av_qsv_list_insert(qsv->dts_seq, i, new_dts); ++ } ++ break; ++ } else if (cur_dts->dts == dts) ++ break; ++ } ++#if HAVE_THREADS ++ if (iter == 0 && qsv->qts_seq_mutex){ ++ mut_ret = pthread_mutex_unlock(qsv->qts_seq_mutex); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_unlock issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ } ++#endif ++} ++ ++void av_qsv_dts_pop(av_qsv_context * qsv) ++{ ++ av_qsv_dts *item = 0; ++#if HAVE_THREADS ++ int mut_ret = 0; ++#endif ++ ++#if HAVE_THREADS ++ if (qsv && qsv->qts_seq_mutex){ ++ mut_ret = pthread_mutex_lock(qsv->qts_seq_mutex); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_lock issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ } ++#endif ++ ++ if (av_qsv_list_count(qsv->dts_seq)) { ++ item = av_qsv_list_item(qsv->dts_seq, 0); ++ av_qsv_list_rem(qsv->dts_seq, item); ++ av_free(item); ++ } ++#if HAVE_THREADS ++ if (qsv && qsv->qts_seq_mutex){ ++ mut_ret = pthread_mutex_unlock(qsv->qts_seq_mutex); ++ if(mut_ret) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_lock issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ } ++#endif ++} ++ ++ ++av_qsv_list *av_qsv_list_init(int is_threaded) ++{ ++ av_qsv_list *l; ++#if HAVE_THREADS ++ int mut_ret; ++#endif ++ ++ l = av_mallocz(sizeof(av_qsv_list)); ++ if (!l) ++ return 0; ++ l->items = av_mallocz(AV_QSV_JOB_SIZE_DEFAULT * sizeof(void *)); ++ if (!l->items) ++ return 0; ++ l->items_alloc = AV_QSV_JOB_SIZE_DEFAULT; ++ ++#if HAVE_THREADS ++ if (is_threaded) { ++ l->mutex = av_mallocz(sizeof(pthread_mutex_t)); ++ if (l->mutex){ ++ mut_ret = pthread_mutexattr_init(&l->mta); ++ if( mut_ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutexattr_init issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ mut_ret = pthread_mutexattr_settype(&l->mta, PTHREAD_MUTEX_RECURSIVE /*PTHREAD_MUTEX_ERRORCHECK*/); ++ if( mut_ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutexattr_settype issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ mut_ret = pthread_mutex_init(l->mutex, &l->mta); ++ if( mut_ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_init issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ } ++ } else ++#endif ++ l->mutex = 0; ++ return l; ++} ++ ++int av_qsv_list_count(av_qsv_list * l) ++{ ++ int count; ++ ++ av_qsv_list_lock(l); ++ count = l->items_count; ++ av_qsv_list_unlock(l); ++ return count; ++} ++ ++int av_qsv_list_add(av_qsv_list * l, void *p) ++{ ++ int pos = -1; ++ ++ if (!p) { ++ return pos; ++ } ++ ++ av_qsv_list_lock(l); ++ ++ if (l->items_count == l->items_alloc) { ++ /* We need a bigger boat */ ++ l->items_alloc += AV_QSV_JOB_SIZE_DEFAULT; ++ l->items = av_realloc(l->items, l->items_alloc * sizeof(void *)); ++ } ++ ++ l->items[l->items_count] = p; ++ pos = (l->items_count); ++ l->items_count++; ++ ++ av_qsv_list_unlock(l); ++ ++ return pos; ++} ++ ++void av_qsv_list_rem(av_qsv_list * l, void *p) ++{ ++ int i; ++ ++ av_qsv_list_lock(l); ++ ++ /* Find the item in the list */ ++ for (i = 0; i < l->items_count; i++) { ++ if (l->items[i] == p) { ++ /* Shift all items after it sizeof( void * ) bytes earlier */ ++ memmove(&l->items[i], &l->items[i + 1], ++ (l->items_count - i - 1) * sizeof(void *)); ++ ++ l->items_count--; ++ break; ++ } ++ } ++ ++ av_qsv_list_unlock(l); ++} ++ ++void *av_qsv_list_item(av_qsv_list * l, int i) ++{ ++ void *ret = NULL; ++ ++ if (i < 0) ++ return NULL; ++ ++ av_qsv_list_lock(l); ++ if( i < l->items_count) ++ ret = l->items[i]; ++ av_qsv_list_unlock(l); ++ return ret; ++} ++ ++void av_qsv_list_insert(av_qsv_list * l, int pos, void *p) ++{ ++ ++ if (!p) ++ return; ++ ++ av_qsv_list_lock(l); ++ ++ if (l->items_count == l->items_alloc) { ++ l->items_alloc += AV_QSV_JOB_SIZE_DEFAULT; ++ l->items = av_realloc(l->items, l->items_alloc * sizeof(void *)); ++ } ++ ++ if (l->items_count != pos) { ++ memmove(&l->items[pos + 1], &l->items[pos], ++ (l->items_count - pos) * sizeof(void *)); ++ } ++ ++ l->items[pos] = p; ++ l->items_count--; ++ ++ av_qsv_list_unlock(l); ++} ++ ++void av_qsv_list_close(av_qsv_list ** _l) ++{ ++ av_qsv_list *l = *_l; ++#if HAVE_THREADS ++ int mut_ret; ++#endif ++ ++ av_qsv_list_lock(l); ++ ++ av_free(l->items); ++ ++#if HAVE_THREADS ++ if (l->mutex){ ++ mut_ret = pthread_mutex_unlock(l->mutex); ++ if( mut_ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_unlock issue[%d] at %s\n",mut_ret, __FUNCTION__); ++ mut_ret = pthread_mutex_destroy(&l->mutex); ++ mut_ret = pthread_mutexattr_destroy(&l->mta); ++ } ++#endif ++ av_freep(_l); ++} ++ ++int av_qsv_list_lock(av_qsv_list *l){ ++ int ret = 0; ++#if HAVE_THREADS ++ if (l->mutex){ ++ ret = pthread_mutex_lock(l->mutex); ++ if( ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_lock issue[%d] at %s\n",ret, __FUNCTION__); ++ } ++#endif ++ return ret; ++} ++ ++int av_qsv_list_unlock(av_qsv_list *l){ ++ int ret = 0; ++#if HAVE_THREADS ++ if (l->mutex){ ++ ret = pthread_mutex_unlock(l->mutex); ++ if( ret ) ++ av_log(NULL, AV_LOG_ERROR, "pthread_mutex_unlock issue[%d] at %s\n",ret, __FUNCTION__); ++ } ++#endif ++ return ret; ++} ++ ++int av_is_qsv_available(mfxIMPL impl, mfxVersion * ver) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ mfxSession mfx_session; ++ ++ memset(&mfx_session, 0, sizeof(mfxSession)); ++ sts = MFXInit(impl, ver, &mfx_session); ++ if (sts >= 0) ++ MFXClose(mfx_session); ++ return sts; ++} ++ ++void av_qsv_wait_on_sync(av_qsv_context *qsv, av_qsv_stage *stage) ++{ ++ int iter = 0; ++ mfxStatus sts = MFX_ERR_NONE; ++ if( stage ) ++ if(*stage->out.sync->p_sync){ ++ while(1){ ++ iter++; ++ sts = MFXVideoCORE_SyncOperation(qsv->mfx_session,*stage->out.sync->p_sync, AV_QSV_SYNC_TIME_DEFAULT); ++ if(MFX_WRN_IN_EXECUTION == sts){ ++ ++ if(iter>20) ++ AV_QSV_DEBUG_ASSERT(1, "Sync failed"); ++ ++ av_qsv_sleep(10); ++ continue; ++ } ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ break; ++ } ++ } ++} +\ No newline at end of file +diff -Naur ../../libav-v9.6/libavcodec/qsv.h ./libavcodec/qsv.h +--- ../../libav-v9.6/libavcodec/qsv.h 1970-01-01 01:00:00.000000000 +0100 ++++ ./libavcodec/qsv.h 2013-08-19 21:32:01.709244686 +0200 +@@ -0,0 +1,494 @@ ++/* ********************************************************************* *\ ++ ++Copyright (C) 2013 Intel Corporation. All rights reserved. ++ ++Redistribution and use in source and binary forms, with or without ++modification, are permitted provided that the following conditions are met: ++- Redistributions of source code must retain the above copyright notice, ++this list of conditions and the following disclaimer. ++- Redistributions in binary form must reproduce the above copyright notice, ++this list of conditions and the following disclaimer in the documentation ++and/or other materials provided with the distribution. ++- Neither the name of Intel Corporation nor the names of its contributors ++may be used to endorse or promote products derived from this software ++without specific prior written permission. ++ ++THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION "AS IS" AND ANY EXPRESS OR ++IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES ++OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. ++IN NO EVENT SHALL INTEL CORPORATION BE LIABLE FOR ANY DIRECT, INDIRECT, ++INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT ++NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, ++DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY ++THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF ++THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ ++\* ********************************************************************* */ ++ ++#ifndef AVCODEC_QSV_H ++#define AVCODEC_QSV_H ++ ++/** ++ * @file ++ * @ingroup lavc_codec_hwaccel_qsv ++ * Common header for QSV/MediaSDK acceleration ++ */ ++ ++/** ++ * @defgroup lavc_codec_hwaccel_qsv QSV/MediaSDK based Decode/Encode and VPP ++ * @ingroup lavc_codec_hwaccel ++ * ++ * As Intel Quick Sync Video (QSV) can decode/preprocess/encode with HW ++ * acceleration. ++ * ++ * Supported features: ++ * - access: ++ * - format AV_PIX_FMT_QSV_H264, AVCodec decoder based implementation ++ * - name "h264_qsv", avcodec_find_decoder_by_name( "h264_qsv") ++ * - IO Pattern: ++ * - Opaque memory: MFX_IOPATTERN_OUT_OPAQUE_MEMORY // Video memory is ++ * MFX_IMPL_HARDWARE or MFX_IMPL_AUTO and runtime support, ++ * otherwise: System Memory ++ * - System memory: MFX_IOPATTERN_OUT_SYSTEM_MEMORY ++ * - Allocators: ++ * - default allocator for System memory: MFX_MEMTYPE_SYSTEM_MEMORY ++ * - details: ++ * implementation as "per frame" ++ * ++ * TODO list: ++ * - access: ++ * - format AV_PIX_FMT_QSV_MPEG2 ++ * - format AV_PIX_FMT_QSV_VC1 ++ * - format AV_PIX_FMT_QSV, see "details" below ++ * - IO Pattern: ++ * - VIDEO_MEMORY // MFX_IOPATTERN_OUT_VIDEO_MEMORY ++ * - Allocators: ++ * - Video memory: MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET / ++ * MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET ++ * - details: ++ * "per slice" support: AV_PIX_FMT_QSV with AVHWAccel based implementation ++ * ++ * Note av_qsv_config struct required to fill in via ++ * AVCodecContext.hwaccel_context ++ * ++ * As per frame, note AVFrame.data[2] (qsv_atom) used for frame atom id, ++ * data/linesize should be used together with SYSTEM_MEMORY and tested ++ * ++ * Note: Compilation would require: ++ * - Intel MediaSDK headers, Full SDK is avaialble from the original web site: ++ * http://software.intel.com/en-us/vcsource/tools/media-SDK ++ * Will be referenced as msdk/*.h (mfxdefs.h, mfxstructures.h, ... ) ++ * and ++ * - Final application has to link against Intel MediaSDK dispatcher, available ++ * at MediaSDK as well ++ * ++ * Target OS: as per available dispatcher and driver support ++ * ++ * Implementation details: ++ * Provided struct av_qsv_context contain several struct av_qsv_space(s) for decode, ++ * VPP and encode. ++ * av_qsv_space just contain needed environment for the appropriate action. ++ * Based on this - pipeline (see pipes) will be build to pass details such as ++ * mfxFrameSurface1* and mfxSyncPoint* from one action to the next. ++ * ++ * Resources re-usage (av_qsv_flush_stages): ++ * av_qsv_context *qsv = (av_qsv_context *)video_codec_ctx->priv_data; ++ * av_qsv_list *pipe = (av_qsv_list *)video_frame->data[2]; ++ * av_qsv_flush_stages( qsv->pipes, &pipe ); ++ * ++ * DTS re-usage: ++ * av_qsv_dts_pop(qsv); ++ * ++ * for video,DX9/11 memory it has to be Unlock'ed as well ++ * ++ * Implementation is thread aware and uses synchronization point(s) from MediaSDK ++ * as per configuration. ++ * ++ * For the details of MediaSDK usage and options available - please refer to the ++ * available documentation at MediaSDK. ++ * ++ * Feature set used from MSDK is defined by AV_QSV_MSDK_VERSION_MAJOR and ++ * AV_QSV_MSDK_VERSION_MINOR ++ * ++ * @{ ++ */ ++ ++#include <stdint.h> ++#include <string.h> ++#include "msdk/mfxvideo.h" ++#include "libavutil/mem.h" ++#include "libavutil/time.h" ++ ++#ifdef HAVE_AV_CONFIG_H ++#include "config.h" ++#endif ++ ++#if HAVE_THREADS ++#if defined (__GNUC__) ++#include <pthread.h> ++#define ff_qsv_atomic_inc(ptr) __sync_add_and_fetch(ptr,1) ++#define ff_qsv_atomic_dec(ptr) __sync_sub_and_fetch (ptr,1) ++#elif HAVE_WINDOWS_H // MSVC case ++#include <windows.h> ++#if HAVE_PTHREADS ++#include <pthread.h> ++#elif HAVE_W32THREADS ++#include "w32pthreads.h" ++#endif ++#define ff_qsv_atomic_inc(ptr) InterlockedIncrement(ptr) ++#define ff_qsv_atomic_dec(ptr) InterlockedDecrement (ptr) ++#else ++// targeting only for MinGW or MSVC ++#endif ++ ++#else ++#define ff_qsv_atomic_inc(ptr) ((*ptr)++) ++#define ff_qsv_atomic_dec(ptr) ((*ptr)--) ++#endif ++ ++ ++// sleep is defined in milliseconds ++#define av_qsv_sleep(x) av_usleep((x)*1000) ++ ++#define AV_QSV_ZERO_MEMORY(VAR) {memset(&VAR, 0, sizeof(VAR));} ++#define AV_QSV_ALIGN32(X) (((mfxU32)((X)+31)) & (~ (mfxU32)31)) ++#define AV_QSV_ALIGN16(value) (((value + 15) >> 4) << 4) ++#ifndef AV_QSV_PRINT_RET_MSG ++#define AV_QSV_PRINT_RET_MSG(ERR) { av_log(NULL, AV_LOG_FATAL,"Error code %d,\t%s\t%d\n", ERR, __FUNCTION__, __LINE__); } ++#endif ++ ++#ifndef AV_QSV_DEBUG_ASSERT ++#define AV_QSV_DEBUG_ASSERT(x,y) {if ((x)) {av_log(NULL, AV_LOG_FATAL,"\nASSERT: %s\n",y);};} ++#endif ++ ++#define AV_QSV_CHECK_RESULT(P, X, ERR) {if ((X) > (P)) {AV_QSV_PRINT_RET_MSG(ERR); return ERR;}} ++#define AV_QSV_CHECK_POINTER(P, ERR) {if (!(P)) {AV_QSV_PRINT_RET_MSG(ERR); return ERR;}} ++#define AV_QSV_IGNORE_MFX_STS(P, X) {if ((X) == (P)) {P = MFX_ERR_NONE;}} ++ ++#define AV_QSV_ID_BUFFER MFX_MAKEFOURCC('B','U','F','F') ++#define AV_QSV_ID_FRAME MFX_MAKEFOURCC('F','R','M','E') ++ ++#define AV_QSV_SURFACE_NUM 80 ++#define AV_QSV_SYNC_NUM AV_QSV_SURFACE_NUM*3/4 ++#define AV_QSV_BUF_SIZE_DEFAULT 4096*2160*10 ++#define AV_QSV_JOB_SIZE_DEFAULT 10 ++#define AV_QSV_SYNC_TIME_DEFAULT 10000 ++// see av_qsv_get_free_sync, av_qsv_get_free_surface , 100 if usleep(10*1000)(10ms) == 1 sec ++#define AV_QSV_REPEAT_NUM_DEFAULT 100 ++#define AV_QSV_ASYNC_DEPTH_DEFAULT 4 ++ ++// version of MSDK/QSV API currently used ++#define AV_QSV_MSDK_VERSION_MAJOR 1 ++#define AV_QSV_MSDK_VERSION_MINOR 3 ++ ++typedef enum AV_QSV_STAGE_TYPE { ++ ++#define AV_QSV_DECODE_MASK 0x001 ++ AV_QSV_DECODE = 0x001, ++ ++#define AV_QSV_VPP_MASK 0x0F0 ++ // "Mandatory VPP filter" , might be with "Hint-based VPP filters" ++ AV_QSV_VPP_DEFAULT = 0x010, ++ // "User Modules" etc ++ AV_QSV_VPP_USER = 0x020, ++ ++#define av_QSV_ENCODE_MASK 0x100 ++ AV_QSV_ENCODE = 0x100 ++#define AV_QSV_ANY_MASK 0xFFF ++} AV_QSV_STAGE_TYPE; ++ ++ ++typedef struct av_qsv_list { ++ // practically pthread_mutex_t ++ void *mutex; ++#if HAVE_THREADS ++ pthread_mutexattr_t mta; ++#endif ++ ++ void **items; ++ int items_alloc; ++ ++ int items_count; ++} av_qsv_list; ++ ++typedef struct av_qsv_sync { ++ mfxSyncPoint* p_sync; ++ int in_use; ++} av_qsv_sync; ++ ++typedef struct av_qsv_stage { ++ AV_QSV_STAGE_TYPE type; ++ struct { ++ mfxBitstream *p_bs; ++ mfxFrameSurface1 *p_surface; ++ } in; ++ struct { ++ mfxBitstream *p_bs; ++ mfxFrameSurface1 *p_surface; ++ av_qsv_sync *sync; ++ } out; ++ av_qsv_list *pending; ++} av_qsv_stage; ++ ++typedef struct av_qsv_task { ++ mfxBitstream *bs; ++ av_qsv_stage *stage; ++} av_qsv_task; ++ ++ ++typedef struct av_qsv_space { ++ ++ uint8_t is_init_done; ++ ++ AV_QSV_STAGE_TYPE type; ++ ++ mfxVideoParam m_mfxVideoParam; ++ ++ mfxFrameAllocResponse response; ++ mfxFrameAllocRequest request[2]; // [0] - in, [1] - out, if needed ++ ++ mfxExtOpaqueSurfaceAlloc ext_opaque_alloc; ++ mfxExtBuffer **p_ext_params; ++ uint16_t p_ext_param_num; ++ ++ uint16_t surface_num_max_used; ++ uint16_t surface_num; ++ mfxFrameSurface1 *p_surfaces[AV_QSV_SURFACE_NUM]; ++ ++ uint16_t sync_num_max_used; ++ uint16_t sync_num; ++ av_qsv_sync *p_syncp[AV_QSV_SYNC_NUM]; ++ ++ mfxBitstream bs; ++ uint8_t *p_buf; ++ size_t p_buf_max_size; ++ ++ // only for encode and tasks ++ av_qsv_list *tasks; ++ ++ av_qsv_list *pending; ++ ++ // storage for allocations/mfxMemId* ++ mfxMemId *mids; ++} av_qsv_space; ++ ++typedef struct av_qsv_context { ++ volatile int is_context_active; ++ ++ mfxIMPL impl; ++ mfxSession mfx_session; ++ mfxVersion ver; ++ ++ // decode ++ av_qsv_space *dec_space; ++ // encode ++ av_qsv_space *enc_space; ++ // vpp ++ av_qsv_list *vpp_space; ++ ++ av_qsv_list *pipes; ++ ++ // MediaSDK starting from API version 1.6 includes DecodeTimeStamp ++ // in addition to TimeStamp ++ // see also AV_QSV_MSDK_VERSION_MINOR , AV_QSV_MSDK_VERSION_MAJOR ++ av_qsv_list *dts_seq; ++ ++ // practically pthread_mutex_t ++ void *qts_seq_mutex; ++ ++ int is_anex; ++ ++ void *qsv_config; ++ ++} av_qsv_context; ++ ++typedef enum { ++ QSV_PART_ANY = 0, ++ QSV_PART_LOWER, ++ QSV_PART_UPPER ++} av_qsv_split; ++ ++typedef struct { ++ int64_t dts; ++} av_qsv_dts; ++ ++typedef struct av_qsv_alloc_frame { ++ mfxU32 id; ++ mfxFrameInfo info; ++} av_qsv_alloc_frame; ++ ++typedef struct av_qsv_alloc_buffer { ++ mfxU32 id; ++ mfxU32 nbytes; ++ mfxU16 type; ++} av_qsv_alloc_buffer; ++ ++typedef struct av_qsv_allocators_space { ++ av_qsv_space *space; ++ mfxFrameAllocator frame_alloc; ++ mfxBufferAllocator buffer_alloc; ++} av_qsv_allocators_space; ++ ++typedef struct av_qsv_config { ++ /** ++ * Set asynch depth of processing with QSV ++ * Format: 0 and more ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int async_depth; ++ ++ /** ++ * Range of numbers that indicate trade-offs between quality and speed. ++ * Format: from 1/MFX_TARGETUSAGE_BEST_QUALITY to 7/MFX_TARGETUSAGE_BEST_SPEED inclusive ++ * ++ * - encoding: Set by user. ++ * - decoding: unused ++ */ ++ int target_usage; ++ ++ /** ++ * Number of reference frames; if NumRefFrame = 0, this parameter is not specified. ++ * Format: 0 and more ++ * ++ * - encoding: Set by user. ++ * - decoding: unused ++ */ ++ int num_ref_frame; ++ ++ /** ++ * Distance between I- or P- key frames; if it is zero, the GOP structure is unspecified. ++ * Note: If GopRefDist = 1, there are no B-frames used. ++ * ++ * - encoding: Set by user. ++ * - decoding: unused ++ */ ++ int gop_ref_dist; ++ ++ /** ++ * Number of pictures within the current GOP (Group of Pictures); if GopPicSize=0, ++ * then the GOP size is unspecified. If GopPicSize=1, only I-frames are used. ++ * ++ * - encoding: Set by user. ++ * - decoding: unused ++ */ ++ int gop_pic_size; ++ ++ /** ++ * Set type of surfaces used with QSV ++ * Format: "IOPattern enum" of Media SDK ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int io_pattern; ++ ++ /** ++ * Set amount of additional surfaces might be needed ++ * Format: ammount of additional buffers(surfaces+syncs) ++ * to allocate in advance ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int additional_buffers; ++ ++ /** ++ * If pipeline should be sync. ++ * Format: wait time in milliseconds, ++ * AV_QSV_SYNC_TIME_DEFAULT/10000 might be a good value ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int sync_need; ++ ++ /** ++ * Type of implementation needed ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int impl_requested; ++ ++ /** ++ * if QSV usage is multithreaded. ++ * Format: Yes/No, 1/0 ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ int usage_threaded; ++ ++ /** ++ * if QSV use an external allocation (valid per session/mfxSession) ++ * Format: pointer to allocators, if default: 0 ++ * ++ * note that: ++ * System Memory: can be used without provided and external allocator, ++ * meaning MediaSDK will use an internal one ++ * Video Memory: in this case - we must provide an external allocator ++ * Also, Media SDK session doesn't require external allocator if the application ++ * uses opaque memory ++ * ++ * Calls SetFrameAllocator/SetBufferAllocator ++ * (MFXVideoCORE_SetFrameAllocator/MFXVideoCORE_SetBufferAllocator) ++ * are to pass allocators to Media SDK ++ * ++ * - encoding: Set by user. ++ * - decoding: Set by user. ++ */ ++ av_qsv_allocators_space *allocators; ++ ++} av_qsv_config; ++ ++#define ANEX_UNKNOWN 0 ++#define ANEX_PREFIX 1 ++#define ANEX_NO_PREFIX 2 ++ ++static const uint8_t ff_prefix_code[] = { 0x00, 0x00, 0x00, 0x01 }; ++ ++int av_qsv_get_free_sync(av_qsv_space *, av_qsv_context *); ++int av_qsv_get_free_surface(av_qsv_space *, av_qsv_context *, mfxFrameInfo *, ++ av_qsv_split); ++int av_qsv_get_free_encode_task(av_qsv_list *); ++ ++int av_is_qsv_available(mfxIMPL, mfxVersion *); ++void av_qsv_wait_on_sync(av_qsv_context *, av_qsv_stage *); ++ ++void av_qsv_add_context_usage(av_qsv_context *, int); ++ ++void av_qsv_pipe_list_create(av_qsv_list **, int); ++void av_qsv_pipe_list_clean(av_qsv_list **); ++ ++void av_qsv_add_stagee(av_qsv_list **, av_qsv_stage *, int); ++av_qsv_stage *av_qsv_get_last_stage(av_qsv_list *); ++av_qsv_list *av_qsv_pipe_by_stage(av_qsv_list *, av_qsv_stage *); ++void av_qsv_flush_stages(av_qsv_list *, av_qsv_list **); ++ ++void av_qsv_dts_ordered_insert(av_qsv_context *, int, int, int64_t, int); ++void av_qsv_dts_pop(av_qsv_context *); ++ ++av_qsv_stage *av_qsv_stage_init(void); ++void av_qsv_stage_clean(av_qsv_stage **); ++int av_qsv_context_clean(av_qsv_context *); ++ ++int ff_qsv_is_sync_in_pipe(mfxSyncPoint *, av_qsv_context *); ++int ff_qsv_is_surface_in_pipe(mfxFrameSurface1 *, av_qsv_context *); ++ ++av_qsv_list *av_qsv_list_init(int); ++int av_qsv_list_lock(av_qsv_list *); ++int av_qsv_list_unlock(av_qsv_list *); ++int av_qsv_list_add(av_qsv_list *, void *); ++void av_qsv_list_rem(av_qsv_list *, void *); ++void av_qsv_list_insert(av_qsv_list *, int, void *); ++void av_qsv_list_close(av_qsv_list **); ++ ++int av_qsv_list_count(av_qsv_list *); ++void *av_qsv_list_item(av_qsv_list *, int); ++ ++/* @} */ ++ ++#endif //AVCODEC_QSV_H +diff -Naur ../../libav-v9.6/libavcodec/qsv_h264.c ./libavcodec/qsv_h264.c +--- ../../libav-v9.6/libavcodec/qsv_h264.c 1970-01-01 01:00:00.000000000 +0100 ++++ ./libavcodec/qsv_h264.c 2013-08-19 21:32:01.705244194 +0200 +@@ -0,0 +1,974 @@ ++/* ********************************************************************* *\ ++ ++Copyright (C) 2013 Intel Corporation. All rights reserved. ++ ++Redistribution and use in source and binary forms, with or without ++modification, are permitted provided that the following conditions are met: ++- Redistributions of source code must retain the above copyright notice, ++this list of conditions and the following disclaimer. ++- Redistributions in binary form must reproduce the above copyright notice, ++this list of conditions and the following disclaimer in the documentation ++and/or other materials provided with the distribution. ++- Neither the name of Intel Corporation nor the names of its contributors ++may be used to endorse or promote products derived from this software ++without specific prior written permission. ++ ++THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION "AS IS" AND ANY EXPRESS OR ++IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES ++OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. ++IN NO EVENT SHALL INTEL CORPORATION BE LIABLE FOR ANY DIRECT, INDIRECT, ++INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT ++NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, ++DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY ++THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF ++THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ ++\* ********************************************************************* */ ++ ++#include "h264.h" ++#include "h264data.h" ++#include "qsv_h264.h" ++ ++static av_qsv_config av_qsv_default_config = { ++ .async_depth = AV_QSV_ASYNC_DEPTH_DEFAULT, ++ .target_usage = MFX_TARGETUSAGE_BALANCED, ++ .num_ref_frame = 0, ++ .gop_ref_dist = 0, ++ .gop_pic_size = 0, ++ .io_pattern = MFX_IOPATTERN_OUT_OPAQUE_MEMORY, ++ .additional_buffers = 0, ++ .sync_need = 0, ++ .impl_requested = MFX_IMPL_HARDWARE, ++ .usage_threaded = 0, ++ .allocators = 0, ++}; ++ ++static av_qsv_allocators_space av_qsv_default_system_allocators = { ++ // fill to access mids ++ .space = 0, ++ ++ .frame_alloc = { ++ .pthis = &av_qsv_default_system_allocators, ++ .Alloc = ff_qsv_mem_frame_alloc, ++ .Lock = ff_qsv_mem_frame_lock, ++ .Unlock = ff_qsv_mem_frame_unlock, ++ .GetHDL = ff_qsv_mem_frame_getHDL, ++ .Free = ff_qsv_mem_frame_free, ++ }, ++ .buffer_alloc = { ++ .pthis = &av_qsv_default_system_allocators, ++ .Alloc = ff_qsv_mem_buffer_alloc, ++ .Lock = ff_qsv_mem_buffer_lock, ++ .Unlock = ff_qsv_mem_buffer_unlock, ++ .Free = ff_qsv_mem_buffer_free, ++ }, ++}; ++ ++static const uint8_t ff_slice_code[] = { 0x00, 0x00, 0x01, 0x65 }; ++ ++int ff_qsv_nal_find_start_code(uint8_t * pb, size_t size) ++{ ++ if ((int) size < 4) ++ return 0; ++ ++ while ((4 <= size) && ((0 != pb[0]) || (0 != pb[1]) || (0 != pb[2]) || (1 != pb[3]))) { ++ pb += 1; ++ size -= 1; ++ } ++ ++ if (4 <= size) ++ return 1; ++ ++ return 0; ++} ++ ++int ff_qsv_dec_init_clean(AVCodecContext *avctx) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_context *qsv = avctx->priv_data; ++ av_qsv_space *qsv_decode = qsv->dec_space; ++ av_qsv_context_clean(qsv); ++ av_freep(&avctx->priv_data); ++} ++int ff_qsv_dec_init(AVCodecContext * avctx) ++{ ++ int ret = 0; ++ mfxStatus sts = MFX_ERR_NONE; ++ size_t current_offset = 6; ++ int header_size = 0; ++ unsigned char *current_position; ++ size_t current_size; ++ ++ av_qsv_context *qsv = avctx->priv_data; ++ av_qsv_space *qsv_decode = qsv->dec_space; ++ av_qsv_config *qsv_config_context = avctx->hwaccel_context; ++ ++ qsv->impl = qsv_config_context->impl_requested; ++ ++ memset(&qsv->mfx_session, 0, sizeof(mfxSession)); ++ qsv->ver.Major = AV_QSV_MSDK_VERSION_MAJOR; ++ qsv->ver.Minor = AV_QSV_MSDK_VERSION_MINOR; ++ ++ sts = MFXInit(qsv->impl, &qsv->ver, &qsv->mfx_session); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ ++ AV_QSV_ZERO_MEMORY(qsv_decode->m_mfxVideoParam); ++ AV_QSV_ZERO_MEMORY(qsv_decode->m_mfxVideoParam.mfx); ++ qsv_decode->m_mfxVideoParam.mfx.CodecId = MFX_CODEC_AVC; ++ qsv_decode->m_mfxVideoParam.IOPattern = ++ qsv_config_context->io_pattern; ++ ++ qsv_decode->m_mfxVideoParam.AsyncDepth = ++ qsv_config_context->async_depth; ++ ++ AV_QSV_ZERO_MEMORY(qsv_decode->bs); ++ { ++ current_position = avctx->extradata; ++ current_size = avctx->extradata_size; ++ ++ if (!ff_qsv_nal_find_start_code(current_position, current_size)) { ++ ++ while (current_offset <= current_size) { ++ int current_nal_size = ++ (unsigned char) current_position[current_offset] << 8 | ++ (unsigned char) current_position[current_offset + 1]; ++ unsigned char nal_type = ++ (unsigned char) current_position[current_offset + 2] & 0x1F; ++ ++ if (nal_type == NAL_SPS || nal_type == NAL_PPS) { ++ memcpy(&qsv_decode->p_buf[header_size], ff_prefix_code, ++ sizeof(ff_prefix_code)); ++ header_size += sizeof(ff_prefix_code); ++ memcpy(&qsv_decode->p_buf[header_size], ++ ¤t_position[current_offset + 2], ++ current_nal_size); ++ ++ // fix for PPS as it comes after SPS, so - last ++ if (nal_type == NAL_PPS) { ++ // fix of MFXVideoDECODE_DecodeHeader: needs one SLICE to find, any SLICE ++ memcpy(&qsv_decode->p_buf ++ [header_size + current_nal_size], ++ ff_slice_code, current_nal_size); ++ header_size += sizeof(ff_slice_code); ++ } ++ } ++ ++ header_size += current_nal_size; ++ current_offset += current_nal_size + 3; ++ } ++ } else { ++ memcpy(&qsv_decode->p_buf[0], avctx->extradata, ++ avctx->extradata_size); ++ header_size = avctx->extradata_size; ++ memcpy(&qsv_decode->p_buf ++ [header_size], ff_slice_code, sizeof(ff_slice_code)); ++ header_size += sizeof(ff_slice_code); ++ } ++ } ++ ++ qsv_decode->bs.Data = qsv_decode->p_buf; ++ qsv_decode->bs.DataLength = header_size; ++ qsv_decode->bs.MaxLength = qsv_decode->p_buf_max_size; ++ ++ if (qsv_decode->bs.DataLength > qsv_decode->bs.MaxLength) { ++ av_log(avctx, AV_LOG_FATAL, "DataLength > MaxLength\n"); ++ return -1; ++ } ++ ++ sts = MFXVideoDECODE_DecodeHeader(qsv->mfx_session, &qsv_decode->bs, ++ &qsv_decode->m_mfxVideoParam); ++ ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ ++ qsv_decode->bs.DataLength -= sizeof(ff_slice_code); ++ ++ memset(&qsv_decode->request, 0, sizeof(mfxFrameAllocRequest) * 2); ++ sts = MFXVideoDECODE_QueryIOSurf(qsv->mfx_session, ++ &qsv_decode->m_mfxVideoParam, ++ &qsv_decode->request); ++ ++ AV_QSV_IGNORE_MFX_STS(sts, MFX_WRN_PARTIAL_ACCELERATION); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ ++ qsv_decode->surface_num = ++ FFMIN(qsv_decode->request[0].NumFrameSuggested + ++ qsv_config_context->async_depth + ++ qsv_config_context->additional_buffers, AV_QSV_SURFACE_NUM); ++ ++ if (qsv_decode->surface_num <= 0) ++ qsv_decode->surface_num = AV_QSV_SURFACE_NUM; ++ ++ if (qsv_decode->m_mfxVideoParam.IOPattern == ++ MFX_IOPATTERN_OUT_SYSTEM_MEMORY) { ++ ++ // as per non-opaque memory: ++ if (!qsv_config_context->allocators) { ++ av_log(avctx, AV_LOG_INFO, ++ "Using default allocators for QSV decode\n"); ++ ((av_qsv_config *) avctx->hwaccel_context)->allocators = ++ &av_qsv_default_system_allocators; ++ } ++ ++ qsv_config_context->allocators->space = qsv_decode; ++ ++ qsv_decode->request[0].NumFrameMin = qsv_decode->surface_num; ++ qsv_decode->request[0].NumFrameSuggested = qsv_decode->surface_num; ++ ++ qsv_decode->request[0].Type = MFX_MEMTYPE_EXTERNAL_FRAME | MFX_MEMTYPE_FROM_DECODE; ++ // qsv_decode->request[0].Type |= m_bd3dAlloc ? MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET : MFX_MEMTYPE_SYSTEM_MEMORY; ++ qsv_decode->request[0].Type |= MFX_MEMTYPE_SYSTEM_MEMORY; ++ ++ qsv_config_context->allocators-> ++ frame_alloc.Alloc(qsv_config_context->allocators, ++ &qsv_decode->request[0], ++ &qsv_decode->response); ++ } ++ ++ for (int i = 0; i < qsv_decode->surface_num; i++) { ++ qsv_decode->p_surfaces[i] = av_mallocz(sizeof(mfxFrameSurface1)); ++ AV_QSV_CHECK_POINTER(qsv_decode->p_surfaces[i], ++ AVERROR(ENOMEM)); ++ memcpy(&(qsv_decode->p_surfaces[i]->Info), ++ &(qsv_decode->request[0].Info), sizeof(mfxFrameInfo)); ++ ++ // for an external(like DX9/11) based allocation: ++ // we bind: ++ // m_pmfxSurfaces[i].Data.MemId = m_mfxResponse.mids[i]; ++ // else, System memory: ++ if (qsv_decode->m_mfxVideoParam.IOPattern == ++ MFX_IOPATTERN_OUT_SYSTEM_MEMORY) { ++ sts = ++ qsv_config_context->allocators-> ++ frame_alloc.Lock(qsv_config_context->allocators, ++ qsv_decode->response.mids[i], ++ &(qsv_decode->p_surfaces[i]->Data)); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ } ++ } ++ ++ qsv_decode->sync_num = FFMIN(qsv_decode->surface_num, AV_QSV_SYNC_NUM); ++ for (int i = 0; i < qsv_decode->sync_num; i++) { ++ qsv_decode->p_syncp[i] = av_mallocz(sizeof(av_qsv_sync)); ++ AV_QSV_CHECK_POINTER(qsv_decode->p_syncp[i], AVERROR(ENOMEM)); ++ qsv_decode->p_syncp[i]->p_sync = av_mallocz(sizeof(mfxSyncPoint)); ++ AV_QSV_CHECK_POINTER(qsv_decode->p_syncp[i]->p_sync, AVERROR(ENOMEM)); ++ } ++ ++ memset(&qsv_decode->ext_opaque_alloc, 0, ++ sizeof(mfxExtOpaqueSurfaceAlloc)); ++ ++ if (qsv_decode->m_mfxVideoParam.IOPattern == ++ MFX_IOPATTERN_OUT_OPAQUE_MEMORY) { ++ qsv_decode->m_mfxVideoParam.NumExtParam = qsv_decode->p_ext_param_num = 1; ++ ++ qsv_decode->p_ext_params = av_mallocz(sizeof(mfxExtBuffer *)*qsv_decode->p_ext_param_num); ++ AV_QSV_CHECK_POINTER(qsv_decode->p_ext_params, AVERROR(ENOMEM)); ++ ++ qsv_decode->m_mfxVideoParam.ExtParam = qsv_decode->p_ext_params; ++ ++ qsv_decode->ext_opaque_alloc.Out.Surfaces = qsv_decode->p_surfaces; ++ qsv_decode->ext_opaque_alloc.Out.NumSurface = qsv_decode->surface_num; ++ qsv_decode->ext_opaque_alloc.Out.Type = qsv_decode->request[0].Type; ++ ++ qsv_decode->ext_opaque_alloc.Header.BufferId = MFX_EXTBUFF_OPAQUE_SURFACE_ALLOCATION; ++ qsv_decode->ext_opaque_alloc.Header.BufferSz = sizeof(mfxExtOpaqueSurfaceAlloc); ++ qsv_decode->p_ext_params[0] = (mfxExtBuffer *) &qsv_decode->ext_opaque_alloc; ++ } ++ ++ sts = ++ MFXVideoDECODE_Init(qsv->mfx_session, ++ &qsv_decode->m_mfxVideoParam); ++ ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ ++ qsv_decode->is_init_done = 1; ++ return ret; ++} ++ ++av_cold int ff_qsv_decode_init(AVCodecContext * avctx) ++{ ++ av_qsv_context *qsv; ++ av_qsv_space *qsv_decode; ++ av_qsv_config **qsv_config_context = ++ (av_qsv_config **) & avctx->hwaccel_context; ++ ++ qsv = avctx->priv_data; ++ ++ if (qsv && qsv->dec_space && qsv->dec_space->is_init_done || !avctx->extradata_size) ++ return 0; ++ ++ if(!qsv) ++ qsv = av_mallocz(sizeof(av_qsv_context)); ++ if (!qsv) ++ return AVERROR(ENOMEM); ++ ++ if(!qsv_decode) ++ qsv_decode = av_mallocz(sizeof(av_qsv_space)); ++ if (!qsv_decode){ ++ free(qsv); ++ return AVERROR(ENOMEM); ++ } ++ avctx->priv_data = qsv; ++ qsv->dec_space = qsv_decode; ++ ++ qsv_decode->p_buf_max_size = AV_QSV_BUF_SIZE_DEFAULT; ++ if(!qsv_decode->p_buf) ++ qsv_decode->p_buf = av_malloc(qsv_decode->p_buf_max_size * sizeof(uint8_t)); ++ if (!qsv_decode->p_buf) ++ return AVERROR(ENOMEM); ++ ++ if (!(*qsv_config_context)) { ++ av_log(avctx, AV_LOG_INFO, ++ "Using default config for QSV decode\n"); ++ avctx->hwaccel_context = &av_qsv_default_config; ++ } else { ++ if ((*qsv_config_context)->io_pattern != ++ MFX_IOPATTERN_OUT_OPAQUE_MEMORY ++ && (*qsv_config_context)->io_pattern != ++ MFX_IOPATTERN_OUT_SYSTEM_MEMORY) { ++ av_log_missing_feature( avctx,"Only MFX_IOPATTERN_OUT_OPAQUE_MEMORY and MFX_IOPATTERN_OUT_SYSTEM_MEMORY are currently supported\n",0); ++ return AVERROR_PATCHWELCOME; ++ } ++ } ++ ++ qsv->qsv_config = avctx->hwaccel_context; ++ ++ av_qsv_add_context_usage(qsv, ++ HAVE_THREADS ++ ? (*qsv_config_context)->usage_threaded : ++ HAVE_THREADS); ++ ++ // allocation of p_syncp and p_surfaces inside of ff_qsv_dec_init ++ return ff_qsv_dec_init(avctx); ++} ++ ++static av_cold int qsv_decode_end(AVCodecContext * avctx) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_context *qsv = avctx->priv_data; ++ av_qsv_config *qsv_config_context = avctx->hwaccel_context; ++ ++ if (qsv) { ++ av_qsv_space *qsv_decode = qsv->dec_space; ++ if (qsv_decode && qsv_decode->is_init_done) { ++ // todo: change to AV_LOG_INFO ++ av_log(avctx, AV_LOG_QUIET, ++ "qsv_decode report done, max_surfaces: %u/%u , max_syncs: %u/%u\n", ++ qsv_decode->surface_num_max_used, ++ qsv_decode->surface_num, qsv_decode->sync_num_max_used, ++ qsv_decode->sync_num); ++ } ++ ++ if (qsv_config_context ++ && qsv_config_context->io_pattern == ++ MFX_IOPATTERN_OUT_SYSTEM_MEMORY) { ++ if (qsv_config_context->allocators) { ++ sts = ++ qsv_config_context->allocators-> ++ frame_alloc.Free(qsv_config_context->allocators, ++ &qsv_decode->response); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ } else { ++ av_log(avctx, AV_LOG_FATAL, ++ "No QSV allocators found for clean up\n"); ++ } ++ } ++ // closing the own resources ++ av_freep(&qsv_decode->p_buf); ++ ++ for (int i = 0; i < qsv_decode->surface_num; i++) { ++ av_freep(&qsv_decode->p_surfaces[i]); ++ } ++ qsv_decode->surface_num = 0; ++ ++ if( qsv_decode->p_ext_param_num || qsv_decode->p_ext_params ) ++ av_freep(&qsv_decode->p_ext_params); ++ qsv_decode->p_ext_param_num = 0; ++ ++ for (int i = 0; i < qsv_decode->sync_num; i++) { ++ av_freep(&qsv_decode->p_syncp[i]->p_sync); ++ av_freep(&qsv_decode->p_syncp[i]); ++ } ++ qsv_decode->sync_num = 0; ++ qsv_decode->is_init_done = 0; ++ ++ av_freep(&qsv->dec_space); ++ ++ // closing commong stuff ++ av_qsv_context_clean(qsv); ++ } ++ ++ return 0; ++} ++ ++static int qsv_decode_frame(AVCodecContext * avctx, void *data, ++ int *data_size, AVPacket * avpkt) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_context *qsv = avctx->priv_data; ++ av_qsv_space *qsv_decode; ++ av_qsv_config *qsv_config_context = avctx->hwaccel_context; ++ int *got_picture_ptr = data_size; ++ int ret_value = 1; ++ uint8_t *current_position = avpkt->data; ++ int current_size = avpkt->size; ++ int frame_processed = 0; ++ size_t frame_length = 0; ++ int surface_idx = 0; ++ int extra_data_workaround = 0; ++ ++ int sync_idx = 0; ++ int current_nal_size; ++ unsigned char nal_type; ++ av_qsv_stage *new_stage = 0; ++ mfxBitstream *input_bs = NULL; ++ size_t current_offset = 2; ++ av_qsv_list *qsv_atom = 0; ++ av_qsv_list *pipe = 0; ++ ++ AVFrame *picture = (AVFrame *) data; ++ ++ *got_picture_ptr = 0; ++ ++ qsv = avctx->priv_data; ++ if(!qsv){ ++ extra_data_workaround = !avctx->extradata_size; ++ if(extra_data_workaround){ ++ avctx->extradata = avpkt->data; ++ avctx->extradata_size = avpkt->size; ++ } ++ sts = ff_qsv_decode_init(avctx); ++ qsv = avctx->priv_data; ++ if(extra_data_workaround){ ++ avctx->extradata = 0; ++ avctx->extradata_size = 0; ++ } ++ if(sts<0){ ++ ff_qsv_dec_init_clean(avctx); ++ *got_picture_ptr = 0; ++ return sts; ++ } ++ } ++ qsv_decode = qsv->dec_space; ++ ++ if (qsv_decode->bs.DataOffset + qsv_decode->bs.DataLength + ++ current_size > qsv_decode->bs.MaxLength) { ++ memmove(&qsv_decode->bs.Data[0], ++ qsv_decode->bs.Data + qsv_decode->bs.DataOffset, ++ qsv_decode->bs.DataLength); ++ qsv_decode->bs.DataOffset = 0; ++ } ++ ++ if (current_size) { ++ if(qsv->is_anex == ANEX_UNKNOWN){ ++ if (ff_qsv_nal_find_start_code(current_position, current_size) && current_position == avpkt->data) ++ qsv->is_anex = ANEX_PREFIX; ++ else ++ qsv->is_anex = ANEX_NO_PREFIX; ++ } ++ if (qsv->is_anex == ANEX_PREFIX){ ++ memcpy(&qsv_decode->bs.Data[0] + ++ qsv_decode->bs.DataLength + ++ qsv_decode->bs.DataOffset, ++ avpkt->data, ++ avpkt->size); ++ qsv_decode->bs.DataLength += avpkt->size; ++ frame_length += avpkt->size; ++ } ++ else ++ while (current_offset <= current_size) { ++ current_nal_size = ++ ((unsigned char) current_position[current_offset - 2] << 24 | ++ (unsigned char) current_position[current_offset - 1] << 16 | ++ (unsigned char) current_position[current_offset] << 8 | ++ (unsigned char) current_position[current_offset + 1]) - 1; ++ nal_type = ++ (unsigned char) current_position[current_offset + 2] & 0x1F; ++ { ++ frame_length += current_nal_size; ++ memcpy(&qsv_decode->bs.Data[0] + ++ qsv_decode->bs.DataLength + ++ qsv_decode->bs.DataOffset, ff_prefix_code, ++ sizeof(ff_prefix_code)); ++ qsv_decode->bs.DataLength += sizeof(ff_prefix_code); ++ memcpy(&qsv_decode->bs.Data[0] + ++ qsv_decode->bs.DataLength + ++ qsv_decode->bs.DataOffset, ++ ¤t_position[current_offset + 2], ++ current_nal_size + 1); ++ qsv_decode->bs.DataLength += current_nal_size + 1; ++ } ++ current_offset += current_nal_size + 5; ++ } ++ ++ if (qsv_decode->bs.DataLength > qsv_decode->bs.MaxLength) { ++ av_log(avctx, AV_LOG_FATAL, "DataLength > MaxLength\n"); ++ return -1; ++ } ++ } ++ ++ if (frame_length || current_size == 0) { ++ ++ qsv_decode->bs.TimeStamp = avpkt->pts; ++ ++ //not a drain ++ if ((current_size || qsv_decode->bs.DataLength)) ++ av_qsv_dts_ordered_insert(qsv, 0, 0, qsv_decode->bs.TimeStamp, 0); ++ ++ sts = MFX_ERR_NONE; ++ // ignore warnings, where warnings >0 , and not error codes <0 ++ while (MFX_ERR_NONE <= sts || MFX_ERR_MORE_SURFACE == sts ++ || MFX_WRN_DEVICE_BUSY == sts) { ++ ++ if (MFX_ERR_MORE_SURFACE == sts || MFX_ERR_NONE == sts) { ++ surface_idx = ++ av_qsv_get_free_surface(qsv_decode, qsv, ++ &qsv_decode->request[0].Info, ++ QSV_PART_ANY); ++ ++ if (surface_idx == -1) { ++ *got_picture_ptr = 0; ++ return 0; ++ } ++ } ++ ++ if (MFX_WRN_DEVICE_BUSY == sts) ++ av_qsv_sleep(10); ++ ++ sync_idx = av_qsv_get_free_sync(qsv_decode, qsv); ++ ++ if (sync_idx == -1) { ++ *got_picture_ptr = 0; ++ return 0; ++ } ++ new_stage = av_qsv_stage_init(); ++ input_bs = NULL; ++ // if to drain last ones ++ if (current_size || qsv_decode->bs.DataLength) ++ input_bs = &qsv_decode->bs; ++ // Decode a frame asynchronously (returns immediately) ++ // very first IDR / SLICE should be with SPS/PPS ++ sts = MFXVideoDECODE_DecodeFrameAsync(qsv->mfx_session, input_bs, ++ qsv_decode->p_surfaces ++ [surface_idx], ++ &new_stage->out.p_surface, ++ qsv_decode->p_syncp[sync_idx]->p_sync); ++ ++ new_stage->out.sync = qsv_decode->p_syncp[sync_idx]; ++ // have some results ++ if (MFX_ERR_NONE <= sts && MFX_WRN_DEVICE_BUSY != sts && ++ MFX_WRN_VIDEO_PARAM_CHANGED != sts) { ++ ++ ff_qsv_atomic_inc(&(new_stage->out.p_surface->Data.Locked)); ++ ++ new_stage->type = AV_QSV_DECODE; ++ new_stage->in.p_bs = input_bs; ++ new_stage->in.p_surface = qsv_decode->p_surfaces[surface_idx]; ++ ++ pipe = av_qsv_list_init(HAVE_THREADS ? qsv_config_context->usage_threaded : HAVE_THREADS); ++ av_qsv_add_stagee(&pipe, new_stage, ++ HAVE_THREADS ? ++ qsv_config_context->usage_threaded : ++ HAVE_THREADS); ++ ++ av_qsv_list_add(qsv->pipes, pipe); ++ qsv_atom = pipe; ++ ++ // usage for forced decode sync and results, can be avoided if sync done by next stage ++ // also note wait time for Sync and possible usage with MFX_WRN_IN_EXECUTION check ++ if (qsv_config_context->sync_need) { ++ sts = ++ MFXVideoCORE_SyncOperation(qsv->mfx_session, ++ qsv_decode->p_syncp[sync_idx]->p_sync, ++ qsv_config_context->sync_need); ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ ++ // no need to wait more -> force off ++ ff_qsv_atomic_dec(&qsv_decode->p_syncp[sync_idx]->in_use); ++ new_stage->out.sync = 0; ++ } ++ ++ sts = MFX_ERR_NONE; ++ break; ++ } ++ av_qsv_stage_clean(&new_stage); ++ ++ /* ++ Can be because of: ++ - runtime situation: ++ - drain procedure: ++ At the end of the bitstream, the application continuously calls the MFXVideoDECODE_DecodeFrameAsync function with a ++ NULL bitstream pointer to drain any remaining frames cached within the Intel ++ Media SDK decoder, until the function returns MFX_ERR_MORE_DATA. ++ */ ++ if (MFX_ERR_MORE_DATA == sts) { ++ // not a drain ++ if (current_size) { ++ *got_picture_ptr = 0; ++ return avpkt->size; ++ } ++ // drain ++ break; ++ } ++ if (MFX_ERR_MORE_SURFACE == sts ){ ++ continue; ++ } ++ ++ AV_QSV_CHECK_RESULT(sts, MFX_ERR_NONE, sts); ++ } ++ ++ frame_processed = 1; ++ } ++ ++ if (frame_processed) { ++ ++ if (current_size) { ++ *got_picture_ptr = 1; ++ ret_value = avpkt->size; ++ } else { ++ if (MFX_ERR_MORE_DATA != sts) { ++ *got_picture_ptr = 1; ++ ret_value = avpkt->size; ++ } else { ++ *got_picture_ptr = 0; ++ return 0; ++ } ++ } ++ ++ picture->pkt_pts = new_stage->out.p_surface->Data.TimeStamp; ++ picture->pts = new_stage->out.p_surface->Data.TimeStamp; ++ ++ picture->repeat_pict = (qsv_decode->m_mfxVideoParam.mfx.FrameInfo.PicStruct & MFX_PICSTRUCT_FIELD_REPEATED); ++ picture->interlaced_frame = !(qsv_decode->m_mfxVideoParam.mfx.FrameInfo.PicStruct & MFX_PICSTRUCT_PROGRESSIVE); ++ picture->top_field_first = (qsv_decode->m_mfxVideoParam.mfx.FrameInfo.PicStruct & MFX_PICSTRUCT_FIELD_TFF); ++ ++ // since we do not know it yet from MSDK, let's do just a simple way for now ++ picture->key_frame = (avctx->frame_number == 0) ? 1 : 0; ++ ++ if (qsv_decode->m_mfxVideoParam.IOPattern == MFX_IOPATTERN_OUT_SYSTEM_MEMORY) { ++ picture->data[0] = new_stage->out.p_surface->Data.Y; ++ picture->data[1] = new_stage->out.p_surface->Data.VU; ++ picture->linesize[0] = new_stage->out.p_surface->Info.Width; ++ picture->linesize[1] = new_stage->out.p_surface->Info.Width; ++ } else { ++ picture->data[0] = 0; ++ picture->data[1] = 0; ++ picture->linesize[0] = 0; ++ picture->linesize[1] = 0; ++ } ++ ++ picture->data[2] = qsv_atom; ++ picture->linesize[2] = 0; ++ } ++ ++ return ret_value; ++} ++ ++// Will be called when seeking ++static void qsv_flush_dpb(AVCodecContext * avctx) ++{ ++ av_qsv_context *qsv = avctx->priv_data; ++ av_qsv_space *qsv_decode = qsv->dec_space; ++ ++ qsv_decode->bs.DataOffset = 0; ++ qsv_decode->bs.DataLength = 0; ++ qsv_decode->bs.MaxLength = qsv_decode->p_buf_max_size; ++} ++ ++ ++mfxStatus ff_qsv_mem_frame_alloc(mfxHDL pthis, ++ mfxFrameAllocRequest * request, ++ mfxFrameAllocResponse * response) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ ++ mfxU32 numAllocated = 0; ++ ++ mfxU32 width = AV_QSV_ALIGN32(request->Info.Width); ++ mfxU32 height = AV_QSV_ALIGN32(request->Info.Height); ++ mfxU32 nbytes; ++ ++ av_qsv_allocators_space *this_alloc = (av_qsv_allocators_space *) pthis; ++ av_qsv_alloc_frame *fs; ++ ++ if (!this_alloc->space) ++ return MFX_ERR_NOT_INITIALIZED; ++ ++ switch (request->Info.FourCC) { ++ case MFX_FOURCC_YV12: ++ case MFX_FOURCC_NV12: ++ nbytes = ++ width * height + (width >> 1) * (height >> 1) + ++ (width >> 1) * (height >> 1); ++ break; ++ case MFX_FOURCC_RGB3: ++ nbytes = width * height + width * height + width * height; ++ break; ++ case MFX_FOURCC_RGB4: ++ nbytes = ++ width * height + width * height + width * height + ++ width * height; ++ break; ++ case MFX_FOURCC_YUY2: ++ nbytes = ++ width * height + (width >> 1) * (height) + ++ (width >> 1) * (height); ++ break; ++ default: ++ return MFX_ERR_UNSUPPORTED; ++ } ++ ++ this_alloc->space->mids = ++ av_malloc(sizeof(mfxMemId) * request->NumFrameSuggested); ++ if (!this_alloc->space->mids) ++ return MFX_ERR_MEMORY_ALLOC; ++ ++ // allocate frames ++ for (numAllocated = 0; numAllocated < request->NumFrameSuggested; ++ numAllocated++) { ++ sts = ++ this_alloc->buffer_alloc.Alloc(this_alloc->buffer_alloc.pthis, ++ nbytes + ++ AV_QSV_ALIGN32(sizeof ++ (av_qsv_alloc_frame)), ++ request->Type, ++ &(this_alloc-> ++ space->mids[numAllocated])); ++ ++ if (MFX_ERR_NONE != sts) ++ break; ++ ++ sts = ++ this_alloc->buffer_alloc.Lock(this_alloc->buffer_alloc.pthis, ++ this_alloc-> ++ space->mids[numAllocated], ++ (mfxU8 **) & fs); ++ ++ if (MFX_ERR_NONE != sts) ++ break; ++ ++ fs->id = AV_QSV_ID_FRAME; ++ fs->info = request->Info; ++ this_alloc->buffer_alloc.Unlock(this_alloc->buffer_alloc.pthis, ++ this_alloc-> ++ space->mids[numAllocated]); ++ } ++ ++ // check the number of allocated frames ++ if (numAllocated < request->NumFrameMin) ++ return MFX_ERR_MEMORY_ALLOC; ++ ++ response->NumFrameActual = (mfxU16) numAllocated; ++ response->mids = this_alloc->space->mids; ++ ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_frame_lock(mfxHDL pthis, mfxMemId mid, ++ mfxFrameData * ptr) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_alloc_frame *fs = 0; ++ mfxU16 width; ++ mfxU16 height; ++ ++ av_qsv_allocators_space *this_alloc = (av_qsv_allocators_space *) pthis; ++ ++ if (!this_alloc->space) ++ return MFX_ERR_NOT_INITIALIZED; ++ if (!ptr) ++ return MFX_ERR_NULL_PTR; ++ ++ ++ sts = ++ this_alloc->buffer_alloc.Lock(this_alloc->buffer_alloc.pthis, mid, ++ (mfxU8 **) & fs); ++ ++ if (MFX_ERR_NONE != sts) ++ return sts; ++ ++ if (AV_QSV_ID_FRAME != fs->id) { ++ this_alloc->buffer_alloc.Unlock(this_alloc->buffer_alloc.pthis, ++ mid); ++ return MFX_ERR_INVALID_HANDLE; ++ } ++ ++ width = (mfxU16) AV_QSV_ALIGN32(fs->info.Width); ++ height = (mfxU16) AV_QSV_ALIGN32(fs->info.Height); ++ ptr->B = ptr->Y = ++ (mfxU8 *) fs + AV_QSV_ALIGN32(sizeof(av_qsv_allocators_space)); ++ ++ switch (fs->info.FourCC) { ++ case MFX_FOURCC_NV12: ++ ptr->U = ptr->Y + width * height; ++ ptr->V = ptr->U + 1; ++ ptr->Pitch = width; ++ break; ++ case MFX_FOURCC_YV12: ++ ptr->V = ptr->Y + width * height; ++ ptr->U = ptr->V + (width >> 1) * (height >> 1); ++ ptr->Pitch = width; ++ break; ++ case MFX_FOURCC_YUY2: ++ ptr->U = ptr->Y + 1; ++ ptr->V = ptr->Y + 3; ++ ptr->Pitch = 2 * width; ++ break; ++ case MFX_FOURCC_RGB3: ++ ptr->G = ptr->B + 1; ++ ptr->R = ptr->B + 2; ++ ptr->Pitch = 3 * width; ++ break; ++ case MFX_FOURCC_RGB4: ++ ptr->G = ptr->B + 1; ++ ptr->R = ptr->B + 2; ++ ptr->A = ptr->B + 3; ++ ptr->Pitch = 4 * width; ++ break; ++ default: ++ return MFX_ERR_UNSUPPORTED; ++ } ++ ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_frame_unlock(mfxHDL pthis, mfxMemId mid, ++ mfxFrameData * ptr) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_allocators_space *this_alloc = (av_qsv_allocators_space *) pthis; ++ ++ sts = ++ this_alloc->buffer_alloc.Unlock(this_alloc->buffer_alloc.pthis, ++ mid); ++ ++ if (MFX_ERR_NONE != sts) ++ return sts; ++ ++ if (NULL != ptr) { ++ ptr->Pitch = 0; ++ ptr->Y = 0; ++ ptr->U = 0; ++ ptr->V = 0; ++ } ++ ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_frame_getHDL(mfxHDL pthis, mfxMemId mid, ++ mfxHDL * handle) ++{ ++ return MFX_ERR_UNSUPPORTED; ++} ++ ++mfxStatus ff_qsv_mem_frame_free(mfxHDL pthis, ++ mfxFrameAllocResponse * response) ++{ ++ mfxStatus sts = MFX_ERR_NONE; ++ av_qsv_allocators_space *this_alloc = (av_qsv_allocators_space *) pthis; ++ mfxU32 i; ++ ++ if (!response) ++ return MFX_ERR_NULL_PTR; ++ ++ if (!this_alloc->space) ++ return MFX_ERR_NOT_INITIALIZED; ++ ++ if (response->mids) ++ for (i = 0; i < response->NumFrameActual; i++) { ++ if (response->mids[i]) { ++ sts = ++ this_alloc->buffer_alloc.Free(this_alloc-> ++ buffer_alloc.pthis, ++ response->mids[i]); ++ if (MFX_ERR_NONE != sts) ++ return sts; ++ } ++ } ++ ++ av_freep(&response->mids); ++ ++ return sts; ++} ++ ++ ++mfxStatus ff_qsv_mem_buffer_alloc(mfxHDL pthis, mfxU32 nbytes, mfxU16 type, ++ mfxMemId * mid) ++{ ++ av_qsv_alloc_buffer *bs; ++ mfxU32 header_size; ++ mfxU8 *buffer_ptr; ++ ++ if (!mid) ++ return MFX_ERR_NULL_PTR; ++ ++ if (0 == (type & MFX_MEMTYPE_SYSTEM_MEMORY)) ++ return MFX_ERR_UNSUPPORTED; ++ ++ header_size = AV_QSV_ALIGN32(sizeof(av_qsv_alloc_buffer)); ++ buffer_ptr = (mfxU8 *) av_malloc(header_size + nbytes); ++ ++ if (!buffer_ptr) ++ return MFX_ERR_MEMORY_ALLOC; ++ ++ bs = (av_qsv_alloc_buffer *) buffer_ptr; ++ bs->id = AV_QSV_ID_BUFFER; ++ bs->type = type; ++ bs->nbytes = nbytes; ++ *mid = (mfxHDL) bs; ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_buffer_lock(mfxHDL pthis, mfxMemId mid, mfxU8 ** ptr) ++{ ++ av_qsv_alloc_buffer *bs; ++ ++ if (!ptr) ++ return MFX_ERR_NULL_PTR; ++ ++ bs = (av_qsv_alloc_buffer *) mid; ++ ++ if (!bs) ++ return MFX_ERR_INVALID_HANDLE; ++ if (AV_QSV_ID_BUFFER != bs->id) ++ return MFX_ERR_INVALID_HANDLE; ++ ++ *ptr = (mfxU8 *) bs + AV_QSV_ALIGN32(sizeof(av_qsv_alloc_buffer)); ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_buffer_unlock(mfxHDL pthis, mfxMemId mid) ++{ ++ av_qsv_alloc_buffer *bs = (av_qsv_alloc_buffer *) mid; ++ ++ if (!bs || AV_QSV_ID_BUFFER != bs->id) ++ return MFX_ERR_INVALID_HANDLE; ++ ++ return MFX_ERR_NONE; ++} ++ ++mfxStatus ff_qsv_mem_buffer_free(mfxHDL pthis, mfxMemId mid) ++{ ++ av_qsv_alloc_buffer *bs = (av_qsv_alloc_buffer *) mid; ++ if (!bs || AV_QSV_ID_BUFFER != bs->id) ++ return MFX_ERR_INVALID_HANDLE; ++ ++ av_freep(&bs); ++ return MFX_ERR_NONE; ++} ++ ++ ++AVCodec ff_h264_qsv_decoder = { ++ .name = "h264_qsv", ++ .type = AVMEDIA_TYPE_VIDEO, ++ .id = AV_CODEC_ID_H264, ++ .init = ff_qsv_decode_init, ++ .close = qsv_decode_end, ++ .decode = qsv_decode_frame, ++ .capabilities = CODEC_CAP_DR1 | CODEC_CAP_DELAY, ++ .flush = qsv_flush_dpb, ++ .long_name = NULL_IF_CONFIG_SMALL("H.264 / AVC / Intel QSV"), ++ .pix_fmts = (const enum PixelFormat[]) {AV_PIX_FMT_QSV_H264, ++ AV_PIX_FMT_NONE}, ++}; +diff -Naur ../../libav-v9.6/libavcodec/qsv_h264.h ./libavcodec/qsv_h264.h +--- ../../libav-v9.6/libavcodec/qsv_h264.h 1970-01-01 01:00:00.000000000 +0100 ++++ ./libavcodec/qsv_h264.h 2013-08-19 21:32:01.710244809 +0200 +@@ -0,0 +1,65 @@ ++/* ********************************************************************* *\ ++ ++Copyright (C) 2013 Intel Corporation. All rights reserved. ++ ++Redistribution and use in source and binary forms, with or without ++modification, are permitted provided that the following conditions are met: ++- Redistributions of source code must retain the above copyright notice, ++this list of conditions and the following disclaimer. ++- Redistributions in binary form must reproduce the above copyright notice, ++this list of conditions and the following disclaimer in the documentation ++and/or other materials provided with the distribution. ++- Neither the name of Intel Corporation nor the names of its contributors ++may be used to endorse or promote products derived from this software ++without specific prior written permission. ++ ++THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION "AS IS" AND ANY EXPRESS OR ++IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES ++OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. ++IN NO EVENT SHALL INTEL CORPORATION BE LIABLE FOR ANY DIRECT, INDIRECT, ++INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT ++NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, ++DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY ++THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF ++THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ ++\* ********************************************************************* */ ++ ++#ifndef AVCODEC_QSV_H264_H ++#define AVCODEC_QSV_H264_H ++ ++#include "qsv.h" ++ ++int ff_qsv_dec_init(AVCodecContext *); ++int ff_qsv_nal_find_start_code(uint8_t * pb, size_t size); ++ ++int ff_qsv_dec_init_clean(AVCodecContext *avctx); ++av_cold int ff_qsv_decode_init(AVCodecContext * avctx); ++static av_cold int qsv_decode_end(AVCodecContext * avctx); ++static int qsv_decode_frame(AVCodecContext * avctx, void *data, ++ int *data_size, AVPacket * avpkt); ++static void qsv_flush_dpb(AVCodecContext * avctx); ++ ++ ++// Default for SYSTEM MEMORY ++// as from MFXFrameAllocator ++mfxStatus ff_qsv_mem_frame_alloc(mfxHDL pthis, ++ mfxFrameAllocRequest * request, ++ mfxFrameAllocResponse * response); ++mfxStatus ff_qsv_mem_frame_lock(mfxHDL pthis, mfxMemId mid, ++ mfxFrameData * ptr); ++mfxStatus ff_qsv_mem_frame_unlock(mfxHDL pthis, mfxMemId mid, ++ mfxFrameData * ptr); ++mfxStatus ff_qsv_mem_frame_getHDL(mfxHDL pthis, mfxMemId mid, ++ mfxHDL * handle); ++mfxStatus ff_qsv_mem_frame_free(mfxHDL pthis, ++ mfxFrameAllocResponse * response); ++// as from mfxBufferAllocator ++mfxStatus ff_qsv_mem_buffer_alloc(mfxHDL pthis, mfxU32 nbytes, mfxU16 type, ++ mfxMemId * mid); ++mfxStatus ff_qsv_mem_buffer_lock(mfxHDL pthis, mfxMemId mid, mfxU8 ** ptr); ++mfxStatus ff_qsv_mem_buffer_unlock(mfxHDL pthis, mfxMemId mid); ++mfxStatus ff_qsv_mem_buffer_free(mfxHDL pthis, mfxMemId mid); ++ ++#endif //AVCODEC_QSV_H264_H +diff -Naur ../../libav-v9.6/libavutil/pixfmt.h ./libavutil/pixfmt.h +--- ../../libav-v9.6/libavutil/pixfmt.h 2013-05-12 08:39:07.000000000 +0200 ++++ ./libavutil/pixfmt.h 2013-08-14 10:48:00.522497405 +0200 +@@ -178,6 +178,7 @@ + AV_PIX_FMT_YUVA422P16LE, ///< planar YUV 4:2:2 48bpp, (1 Cr & Cb sample per 2x1 Y & A samples, little-endian) + AV_PIX_FMT_YUVA444P16BE, ///< planar YUV 4:4:4 64bpp, (1 Cr & Cb sample per 1x1 Y & A samples, big-endian) + AV_PIX_FMT_YUVA444P16LE, ///< planar YUV 4:4:4 64bpp, (1 Cr & Cb sample per 1x1 Y & A samples, little-endian) ++ AV_PIX_FMT_QSV_H264, ///< H.264 HW decoding with QSV, data[2] contains qsv_atom information for MFX_IOPATTERN_OUT_OPAQUE_MEMORY, MFX_IOPATTERN_OUT_VIDEO_MEMORY + AV_PIX_FMT_NB, ///< number of pixel formats, DO NOT USE THIS if you want to link with shared libav* because the number of formats might differ between versions + + #if FF_API_PIX_FMT |