日韩av黄I国产麻豆传媒I国产91av视频在线观看I日韩一区二区三区在线看I美女国产在线I麻豆视频国产在线观看I成人黄色短片

歡迎訪問 生活随笔!

生活随笔

當(dāng)前位置: 首頁 > 运维知识 > Android >内容正文

Android

Android 使用FFmpeg3.3.9基于命令实现视频压缩

發(fā)布時(shí)間:2023/12/3 Android 62 豆豆
生活随笔 收集整理的這篇文章主要介紹了 Android 使用FFmpeg3.3.9基于命令实现视频压缩 小編覺得挺不錯(cuò)的,現(xiàn)在分享給大家,幫大家做個(gè)參考.

前言

首先利用linux平臺(tái)編譯ffmpeg的so庫,具體詳情請(qǐng)查看文章:Android NDK(ndk-r16b)交叉編譯FFmpeg(3.3.9)_jszlittlecat_720的博客-CSDN博客

1.創(chuàng)建VideoCompress類

?

package com.suoer.ndk.ffmpegtestapplication;public class VideoCompress {/*** compressVideo native 方法* @param compressCommand 壓縮命令* @param callback 壓縮回調(diào)*/public native void compressVideo(String[] compressCommand,CompressCallback callback);//CompressCallback 壓縮回調(diào)public interface CompressCallback{/*** onCompress* @param current 壓縮的當(dāng)前進(jìn)度* @param total 總進(jìn)度*/public void onCompress(int current,int total);} } compressVideo報(bào)紅,鼠標(biāo)停在上面,左邊會(huì)出現(xiàn)紅色小燈泡,點(diǎn)擊紅色小燈泡

?點(diǎn)擊Create JNI function for compressVideo

自動(dòng)打開native-lib.cpp并創(chuàng)建完成Java_com_suoer_ndk_ffmpegtestapplication_VideoCompress_compressVideo 方法

?在此方法下實(shí)現(xiàn)壓縮視頻

2.MainActivity實(shí)現(xiàn)點(diǎn)擊TextView實(shí)現(xiàn)壓縮視頻

2.1權(quán)限問題處理

壓縮視頻需要讀寫文件的權(quán)限

打開AndroidManifest.xml

添加讀寫文件權(quán)限

<uses-permission android:name="android.permission.WRITE_EXTERNAL_STORAGE"></uses-permission><uses-permission android:name="android.permission.READ_EXTERNAL_STORAGE"></uses-permission>

?權(quán)限處理使用rxpermissions

rxpermissions github地址:GitHub - tbruyelle/RxPermissions: Android runtime permissions powered by RxJava2

rxpermissions使用方式

打開項(xiàng)目下的build.gradle添加如下代碼:

maven { url 'https://jitpack.io' }

?打開app下的build.gradle添加如下代碼:

implementation 'com.github.tbruyelle:rxpermissions:0.12'

?點(diǎn)擊同步按鈕,同步文件

?2.2子線程處理耗時(shí)操作

壓縮視頻是耗時(shí)操作,子線程處理耗時(shí)操作。

使用rxandroid實(shí)現(xiàn)子線程耗時(shí)操作并實(shí)現(xiàn)子線程和主線程切換

rxandroid github地址:GitHub - ReactiveX/RxAndroid: RxJava bindings for Android

?rxandroid使用方式:

打開項(xiàng)目下的build.gradle添加如下代碼:

maven { url "https://oss.jfrog.org/libs-snapshot" }

?打開app下的build.gradle添加如下代碼:

implementation 'io.reactivex.rxjava3:rxandroid:3.0.0'

?點(diǎn)擊同步按鈕同步文件

MainActivity.java內(nèi)容如下:

package com.suoer.ndk.ffmpegtestapplication;import android.Manifest; import android.os.Bundle; import android.os.Environment; import android.util.Log; import android.view.View; import android.widget.TextView;import com.tbruyelle.rxpermissions3.RxPermissions;import java.io.File;import androidx.appcompat.app.AppCompatActivity; import io.reactivex.rxjava3.android.schedulers.AndroidSchedulers; import io.reactivex.rxjava3.core.Observable; import io.reactivex.rxjava3.functions.Consumer; import io.reactivex.rxjava3.functions.Function; import io.reactivex.rxjava3.schedulers.Schedulers;public class MainActivity extends AppCompatActivity {private File mInFile=new File(Environment.getExternalStorageDirectory(),"test.mp4");//mInFile 需要壓縮的文件路徑private File mOutFile=new File(Environment.getExternalStorageDirectory(),"out.mp4");//mOutFile壓縮后的文件路徑// Used to load the 'native-lib' library on application startup.static {System.loadLibrary("native-lib");}@Overrideprotected void onCreate(Bundle savedInstanceState) {super.onCreate(savedInstanceState);setContentView(R.layout.activity_main);// Example of a call to a native methodTextView tv = findViewById(R.id.sample_text);//tv.setText("ffmpeg版本:"+stringFromJNI());tv.setText("壓縮");//tv的點(diǎn)擊事件 點(diǎn)擊按鈕實(shí)現(xiàn)視頻壓縮tv.setOnClickListener(new View.OnClickListener() {@Overridepublic void onClick(View v) {// 壓縮文件 需要讀寫文件權(quán)限 申請(qǐng)權(quán)限RxPermissions rxPermissions=new RxPermissions(MainActivity.this);rxPermissions.request(Manifest.permission.READ_EXTERNAL_STORAGE, Manifest.permission.WRITE_EXTERNAL_STORAGE).subscribe(new Consumer<Boolean>() {@Overridepublic void accept(Boolean aBoolean) throws Throwable {if(aBoolean){//權(quán)限已經(jīng)獲取 壓縮視頻compressVideo();}}});}});}/*** 開啟子線程處理耗時(shí)壓縮問題*/private void compressVideo() {//ffmpeg的壓縮命令:ffmpeg -i test.mp4 -b:v 1024k out.mp4//ffmpeg -i test.mp4 -b:v 1024k out.mp4//-b:v 1024k 1024k為碼率 碼率越高視頻越清晰,而且視頻越大//test.mp4需要壓縮的文件//out.mp4 壓縮之后的文件String[] compressCommand={"ffmpeg","-i",mInFile.getAbsolutePath(),"-b:v","1024k",mOutFile.getAbsolutePath()};//壓縮是耗時(shí)的,需要子線程處理Observable.just(compressCommand).map(new Function<String[], File>() {@Overridepublic File apply(String[] compressCommand) throws Throwable {VideoCompress videoCompress=new VideoCompress();videoCompress.compressVideo(compressCommand, new VideoCompress.CompressCallback() {@Overridepublic void onCompress(int current, int total) {Log.e("TAG", "onCompress: 壓縮進(jìn)度:"+current+"/"+total);}});return mOutFile;}}).subscribeOn(Schedulers.io()).observeOn(AndroidSchedulers.mainThread()).subscribe(new Consumer<File>() {@Overridepublic void accept(File file) throws Throwable {//壓縮完成Log.e("TAG", "accept: 壓縮完成!" );}});}/*** A native method that is implemented by the 'native-lib' native library,* which is packaged with this application.*/public native String stringFromJNI(); }

3.拷貝視頻壓縮使用命令實(shí)現(xiàn)需要的其他文件

include目錄下創(chuàng)建compat文件夾把所需要的頭文件?os2threads.h?va_copy.h?w32pthreads.h拷貝此文件夾中

這三個(gè)頭文件存在于下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9的compat文件夾中

?在jniLibs文件夾中創(chuàng)建other文件夾

?拷貝linux平臺(tái)編譯后的文件cmdutils.h cmdutils_common_opts.h config.h ffmpeg.h 拷貝至other文件夾中

?

?將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中的cmdutils.c?

ffmpeg.c?ffmpeg_filter.c?ffmpeg_opt.c 四個(gè)文件拷貝至cpp文件夾中

?

?配置CMakeLists.txt文件修改如下內(nèi)容

?CMakeLists.txt內(nèi)容如下:

# For more information about using CMake with Android Studio, read the # documentation: https://d.android.com/studio/projects/add-native-code.html# Sets the minimum version of CMake required to build the native library.cmake_minimum_required(VERSION 3.10.2)# Declares and names the project.project("ffmpegtestapplication")#判斷編譯器類型,如果是gcc編譯器,則在編譯選項(xiàng)中加入c++11支持 if(CMAKE_COMPILER_IS_GNUCXX)set(CMAKE_CXX_FLAGS "-std=c++11 ${CMAKE_CXX_FLAGS}")message(STATUS "optional:-std=c++11") endif(CMAKE_COMPILER_IS_GNUCXX)#需要引入我們頭文件,以這個(gè)配置的目錄為基準(zhǔn)include_directories(${CMAKE_SOURCE_DIR}/../jniLibs/include) include_directories(${CMAKE_SOURCE_DIR}/../jniLibs/other)# Creates and names a library, sets it as either STATIC # or SHARED, and provides the relative paths to its source code. # You can define multiple libraries, and CMake builds them for you. # Gradle automatically packages shared libraries with your APK.add_library( # Sets the name of the library.native-lib# Sets the library as a shared library.SHARED# Provides a relative path to your source file(s).native-lib.cpp#添加額外的c文件cmdutils.cffmpeg.cffmpeg_filter.cffmpeg_opt.c) # 編解碼(最重要的庫) add_library(avcodecSHAREDIMPORTED) set_target_properties(avcodecPROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavcodec-57.so)# 設(shè)備信息 add_library(avdeviceSHAREDIMPORTED) set_target_properties(avdevicePROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavdevice-57.so)# 濾鏡特效處理庫 add_library(avfilterSHAREDIMPORTED) set_target_properties(avfilterPROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavfilter-6.so)# 封裝格式處理庫 add_library(avformatSHAREDIMPORTED) set_target_properties(avformatPROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavformat-57.so) # 重采樣處理庫 add_library(avresampleSHAREDIMPORTED) set_target_properties(avresamplePROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavresample-3.so) # 工具庫(大部分庫都需要這個(gè)庫的支持) add_library(avutilSHAREDIMPORTED) set_target_properties(avutilPROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libavutil-55.so)# 后期處理 add_library(postprocSHAREDIMPORTED) set_target_properties(postprocPROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libpostproc-54.so)# 音頻采樣數(shù)據(jù)格式轉(zhuǎn)換庫 add_library(swresampleSHAREDIMPORTED) set_target_properties(swresamplePROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libswresample-2.so)# 視頻像素?cái)?shù)據(jù)格式轉(zhuǎn)換 add_library(swscaleSHAREDIMPORTED) set_target_properties(swscalePROPERTIES IMPORTED_LOCATION${CMAKE_SOURCE_DIR}/../jniLibs/armeabi-v7a/libswscale-4.so)# Searches for a specified prebuilt library and stores the path as a # variable. Because CMake includes system libraries in the search path by # default, you only need to specify the name of the public NDK library # you want to add. CMake verifies that the library exists before # completing its build.find_library( # Sets the name of the path variable.log-lib# Specifies the name of the NDK library that# you want CMake to locate.log )# Specifies libraries CMake should link to your target library. You # can link multiple libraries, such as libraries you define in this # build script, prebuilt third-party libraries, or system libraries.target_link_libraries( # Specifies the target library.native-lib avcodec avdevice avfilter avformat avresample avutil postproc swresample swscale# Links the target library to the log library# included in the NDK.${log-lib} )

運(yùn)行Run app至手機(jī)設(shè)備,出現(xiàn)如下如所示錯(cuò)誤

?D:\suowei\androidproject\FFmpegTestApplication\app\src\main\cpp\cmdutils.c:47:10: fatal error: 'libavutil/libm.h' file not found

解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavutil中的libm.h拷貝至android項(xiàng)目jniLibs下的libavutil中

?

?D:\suowei\androidproject\FFmpegTestApplication\app\src\main\cpp\ffmpeg.c:52:10: fatal error: 'libavutil/internal.h' file not found

?解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavutil中的internal.h拷貝至android項(xiàng)目jniLibs下的libavutil中

?

?重新Run app至手機(jī)設(shè)備,出現(xiàn)如下圖所示錯(cuò)誤

?D:\suowei\androidproject\FFmpegTestApplication\app\src\main\cpp\cmdutils.c:58:10: fatal error: 'libavformat/network.h' file not found

解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavformat中的network.h拷貝至android項(xiàng)目jniLibs下的libavformat中

?

?D:\suowei\androidproject\FFmpegTestApplication\app\src\main\jniLibs\include\libavutil\internal.h:42:10: fatal error: 'timer.h' file not found

?解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavutil中的timer.h拷貝至android項(xiàng)目jniLibs下的libavutil中

?

?

?重新Run app至手機(jī)設(shè)備,出現(xiàn)如下圖所示錯(cuò)誤

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/../jniLibs/include\libavformat/network.h:29:10: fatal error: 'os_support.h' file not found
#include "os_support.h"

??解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavformat中的os_support.h拷貝至android項(xiàng)目jniLibs下的libavformat中

?

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/../jniLibs/include\libavutil/timer.h:44:13: fatal error: 'arm/timer.h' file not found

解決方式:android項(xiàng)目jniLibs目錄下的libavutil文件夾下創(chuàng)建arm文件夾

?將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavutil中的arm文件夾下的timer.h拷貝至android項(xiàng)目jniLibs下的libavutil中的arm文件夾下

?

?重新Run app至手機(jī)設(shè)備,出現(xiàn)如下圖所示錯(cuò)誤

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/ffmpeg.c:65:10: fatal error: 'libavcodec/mathops.h' file not found

解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavcodec中的mathops.h拷貝至android項(xiàng)目jniLibs下的libavcodec中

?

?重新Run app至手機(jī)設(shè)備,出現(xiàn)如下圖所示錯(cuò)誤

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/../jniLibs/include\libavformat/network.h:31:10: fatal error: 'url.h' file not found

解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavformat中的url.h拷貝至android項(xiàng)目jniLibs下的libavformat中

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/../jniLibs/include\libavcodec/mathops.h:28:10: fatal error: 'libavutil/reverse.h' file not found

解決方式:將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavutil中的reverse.h拷貝至android項(xiàng)目jniLibs下的libavutil中

?

?

?重新Run app至手機(jī)設(shè)備,出現(xiàn)如下圖所示錯(cuò)誤

?D:/suowei/androidproject/FFmpegTestApplication/app/src/main/cpp/../jniLibs/include\libavcodec/mathops.h:40:13: fatal error: 'arm/mathops.h' file not found

解決方式:android項(xiàng)目jniLibs目錄下的libavcodec文件夾下創(chuàng)建arm文件夾

?

?將下載之后的ffmpeg-3.3.9.tar.gz解壓之后的文件夾ffmpeg-3.3.9中l(wèi)ibavcodec中的arm文件夾下的mathops.h拷貝至android項(xiàng)目jniLibs下的libavcodec中的arm文件夾下

?

?重新Run app至手機(jī)設(shè)備,android項(xiàng)目運(yùn)行成功。

4.編寫native-cpp實(shí)現(xiàn)視頻壓縮

?

修改ffmpeg.c,找到main方法修改方法名稱為:run_ffmpeg_command 并注釋掉此行代碼

exit_program(received_nb_signals ? 255 : main_return_code); //命令函數(shù)的入口 int run_ffmpeg_command(int argc, char **argv) {int i, ret;int64_t ti;init_dynload();register_exit(ffmpeg_cleanup);setvbuf(stderr,NULL,_IONBF,0); /* win32 runtime needs this */av_log_set_flags(AV_LOG_SKIP_REPEATED);parse_loglevel(argc, argv, options);if(argc>1 && !strcmp(argv[1], "-d")){run_as_daemon=1;av_log_set_callback(log_callback_null);argc--;argv++;}avcodec_register_all(); #if CONFIG_AVDEVICEavdevice_register_all(); #endifavfilter_register_all();av_register_all();avformat_network_init();show_banner(argc, argv, options);/* parse options and open all input/output files */ret = ffmpeg_parse_options(argc, argv);if (ret < 0)exit_program(1);if (nb_output_files <= 0 && nb_input_files == 0) {show_usage();av_log(NULL, AV_LOG_WARNING, "Use -h to get full help or, even better, run 'man %s'\n", program_name);exit_program(1);}/* file converter / grab */if (nb_output_files <= 0) {av_log(NULL, AV_LOG_FATAL, "At least one output file must be specified\n");exit_program(1);}// if (nb_input_files == 0) { // av_log(NULL, AV_LOG_FATAL, "At least one input file must be specified\n"); // exit_program(1); // }for (i = 0; i < nb_output_files; i++) {if (strcmp(output_files[i]->ctx->oformat->name, "rtp"))want_sdp = 0;}current_time = ti = getutime();if (transcode() < 0)exit_program(1);ti = getutime() - ti;if (do_benchmark) {av_log(NULL, AV_LOG_INFO, "bench: utime=%0.3fs\n", ti / 1000000.0);}av_log(NULL, AV_LOG_DEBUG, "%"PRIu64" frames successfully decoded, %"PRIu64" decoding errors\n",decode_error_stat[0], decode_error_stat[1]);if ((decode_error_stat[0] + decode_error_stat[1]) * max_error_rate < decode_error_stat[1])exit_program(69);//注釋掉此行代碼 如果不注釋掉該行代碼,命令執(zhí)行完會(huì)導(dǎo)致app退出//exit_program(received_nb_signals ? 255 : main_return_code);return main_return_code; }

native-lib.cpp代碼如下:

#include <jni.h> #include <string> #include <android/log.h> #define TAG "JNI_TAG" #define LOGE(...) __android_log_print(ANDROID_LOG_ERROR,TAG,__VA_ARGS__); extern "C" { #include "libavutil/avutil.h"//聲明方法 argc命令的個(gè)數(shù) argv 二維數(shù)組 int run_ffmpeg_command(int argc, char **argv); } extern "C" JNIEXPORT jstring JNICALL Java_com_suoer_ndk_ffmpegtestapplication_MainActivity_stringFromJNI(JNIEnv* env,jobject /* this */) {//std::string hello = "Hello from C++";return env->NewStringUTF(av_version_info()); }extern "C" JNIEXPORT void JNICALL Java_com_suoer_ndk_ffmpegtestapplication_VideoCompress_compressVideo(JNIEnv *env, jobject thiz,jobjectArray compress_command,jobject callback) {//ffmpeg 處理視頻壓縮//arm這個(gè)里面的so都是用來處理音視頻的,include都是頭文件//還有幾個(gè)沒有被打包編譯成so,因?yàn)檫@些不算是音視頻的處理代碼,只是我們現(xiàn)在支持命令(封裝)//1.獲取命令個(gè)數(shù)int argc=env->GetArrayLength(compress_command);//2.給char **argv填充數(shù)據(jù)char **argv=(char **)malloc(sizeof(char*)*argc);for (int i = 0; i <argc ; ++i) {jstring j_param=(jstring)env->GetObjectArrayElement(compress_command,i);argv[i]= (char *)(env->GetStringUTFChars(j_param, NULL));LOGE("參數(shù):%s",argv[i]);}//3.調(diào)用命令函數(shù)去壓縮run_ffmpeg_command(argc,argv);//4.釋放內(nèi)存for (int i = 0; i <argc ; ++i) {free(argv[i]);}free(argv); }

視頻壓縮回調(diào)處理,視頻壓縮進(jìn)度回調(diào)

修改ffmpeg.c內(nèi)容如下:

print_report方法

?

?transcode方法

?

?

?run_ffmpeg_command方法

?

?

?ffmpeg.c內(nèi)容如下:

/** Copyright (c) 2000-2003 Fabrice Bellard** This file is part of FFmpeg.** FFmpeg is free software; you can redistribute it and/or* modify it under the terms of the GNU Lesser General Public* License as published by the Free Software Foundation; either* version 2.1 of the License, or (at your option) any later version.** FFmpeg is distributed in the hope that it will be useful,* but WITHOUT ANY WARRANTY; without even the implied warranty of* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU* Lesser General Public License for more details.** You should have received a copy of the GNU Lesser General Public* License along with FFmpeg; if not, write to the Free Software* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA*//*** @file* multimedia converter based on the FFmpeg libraries*/#include "config.h" #include <ctype.h> #include <string.h> #include <math.h> #include <stdlib.h> #include <errno.h> #include <limits.h> #include <stdatomic.h> #include <stdint.h>#if HAVE_IO_H #include <io.h> #endif #if HAVE_UNISTD_H #include <unistd.h> #endif#include "libavformat/avformat.h" #include "libavdevice/avdevice.h" #include "libswresample/swresample.h" #include "libavutil/opt.h" #include "libavutil/channel_layout.h" #include "libavutil/parseutils.h" #include "libavutil/samplefmt.h" #include "libavutil/fifo.h" #include "libavutil/hwcontext.h" #include "libavutil/internal.h" #include "libavutil/intreadwrite.h" #include "libavutil/dict.h" #include "libavutil/display.h" #include "libavutil/mathematics.h" #include "libavutil/pixdesc.h" #include "libavutil/avstring.h" #include "libavutil/libm.h" #include "libavutil/imgutils.h" #include "libavutil/timestamp.h" #include "libavutil/bprint.h" #include "libavutil/time.h" #include "libavutil/threadmessage.h" #include "libavcodec/mathops.h" #include "libavformat/os_support.h"# include "libavfilter/avfilter.h" # include "libavfilter/buffersrc.h" # include "libavfilter/buffersink.h"#if HAVE_SYS_RESOURCE_H #include <sys/time.h> #include <sys/types.h> #include <sys/resource.h> #elif HAVE_GETPROCESSTIMES #include <windows.h> #endif #if HAVE_GETPROCESSMEMORYINFO #include <windows.h> #include <psapi.h> #endif #if HAVE_SETCONSOLECTRLHANDLER #include <windows.h> #endif#if HAVE_SYS_SELECT_H #include <sys/select.h> #endif#if HAVE_TERMIOS_H #include <fcntl.h> #include <sys/ioctl.h> #include <sys/time.h> #include <termios.h> #elif HAVE_KBHIT #include <conio.h> #endif#if HAVE_PTHREADS #include <pthread.h> #endif#include <time.h> #include <android/log.h>#include "ffmpeg.h" #include "cmdutils.h"#include "libavutil/avassert.h"const char program_name[] = "ffmpeg"; const int program_birth_year = 2000;static FILE *vstats_file;const char *const forced_keyframes_const_names[] = {"n","n_forced","prev_forced_n","prev_forced_t","t",NULL };static void do_video_stats(OutputStream *ost, int frame_size); static int64_t getutime(void); static int64_t getmaxrss(void); static int ifilter_has_all_input_formats(FilterGraph *fg);static int run_as_daemon = 0; static int nb_frames_dup = 0; static unsigned dup_warning = 1000; static int nb_frames_drop = 0; static int64_t decode_error_stat[2];static int want_sdp = 1;static int current_time; AVIOContext *progress_avio = NULL;static uint8_t *subtitle_out;InputStream **input_streams = NULL; int nb_input_streams = 0; InputFile **input_files = NULL; int nb_input_files = 0;OutputStream **output_streams = NULL; int nb_output_streams = 0; OutputFile **output_files = NULL; int nb_output_files = 0;FilterGraph **filtergraphs; int nb_filtergraphs;#if HAVE_TERMIOS_H/* init terminal so that we can grab keys */ static struct termios oldtty; static int restore_tty; #endif#if HAVE_PTHREADS static void free_input_threads(void); #endif/* sub2video hack:Convert subtitles to video with alpha to insert them in filter graphs.This is a temporary solution until libavfilter gets real subtitles support.*/static int sub2video_get_blank_frame(InputStream *ist) {int ret;AVFrame *frame = ist->sub2video.frame;av_frame_unref(frame);ist->sub2video.frame->width = ist->dec_ctx->width ? ist->dec_ctx->width : ist->sub2video.w;ist->sub2video.frame->height = ist->dec_ctx->height ? ist->dec_ctx->height : ist->sub2video.h;ist->sub2video.frame->format = AV_PIX_FMT_RGB32;if ((ret = av_frame_get_buffer(frame, 32)) < 0)return ret;memset(frame->data[0], 0, frame->height * frame->linesize[0]);return 0; }static void sub2video_copy_rect(uint8_t *dst, int dst_linesize, int w, int h,AVSubtitleRect *r) {uint32_t *pal, *dst2;uint8_t *src, *src2;int x, y;if (r->type != SUBTITLE_BITMAP) {av_log(NULL, AV_LOG_WARNING, "sub2video: non-bitmap subtitle\n");return;}if (r->x < 0 || r->x + r->w > w || r->y < 0 || r->y + r->h > h) {av_log(NULL, AV_LOG_WARNING, "sub2video: rectangle (%d %d %d %d) overflowing %d %d\n",r->x, r->y, r->w, r->h, w, h);return;}dst += r->y * dst_linesize + r->x * 4;src = r->data[0];pal = (uint32_t *)r->data[1];for (y = 0; y < r->h; y++) {dst2 = (uint32_t *)dst;src2 = src;for (x = 0; x < r->w; x++)*(dst2++) = pal[*(src2++)];dst += dst_linesize;src += r->linesize[0];} }static void sub2video_push_ref(InputStream *ist, int64_t pts) {AVFrame *frame = ist->sub2video.frame;int i;av_assert1(frame->data[0]);ist->sub2video.last_pts = frame->pts = pts;for (i = 0; i < ist->nb_filters; i++)av_buffersrc_add_frame_flags(ist->filters[i]->filter, frame,AV_BUFFERSRC_FLAG_KEEP_REF |AV_BUFFERSRC_FLAG_PUSH); }void sub2video_update(InputStream *ist, AVSubtitle *sub) {AVFrame *frame = ist->sub2video.frame;int8_t *dst;int dst_linesize;int num_rects, i;int64_t pts, end_pts;if (!frame)return;if (sub) {pts = av_rescale_q(sub->pts + sub->start_display_time * 1000LL,AV_TIME_BASE_Q, ist->st->time_base);end_pts = av_rescale_q(sub->pts + sub->end_display_time * 1000LL,AV_TIME_BASE_Q, ist->st->time_base);num_rects = sub->num_rects;} else {pts = ist->sub2video.end_pts;end_pts = INT64_MAX;num_rects = 0;}if (sub2video_get_blank_frame(ist) < 0) {av_log(ist->dec_ctx, AV_LOG_ERROR,"Impossible to get a blank canvas.\n");return;}dst = frame->data [0];dst_linesize = frame->linesize[0];for (i = 0; i < num_rects; i++)sub2video_copy_rect(dst, dst_linesize, frame->width, frame->height, sub->rects[i]);sub2video_push_ref(ist, pts);ist->sub2video.end_pts = end_pts; }static void sub2video_heartbeat(InputStream *ist, int64_t pts) {InputFile *infile = input_files[ist->file_index];int i, j, nb_reqs;int64_t pts2;/* When a frame is read from a file, examine all sub2video streams inthe same file and send the sub2video frame again. Otherwise, decodedvideo frames could be accumulating in the filter graph while a filter(possibly overlay) is desperately waiting for a subtitle frame. */for (i = 0; i < infile->nb_streams; i++) {InputStream *ist2 = input_streams[infile->ist_index + i];if (!ist2->sub2video.frame)continue;/* subtitles seem to be usually muxed ahead of other streams;if not, subtracting a larger time here is necessary */pts2 = av_rescale_q(pts, ist->st->time_base, ist2->st->time_base) - 1;/* do not send the heartbeat frame if the subtitle is already ahead */if (pts2 <= ist2->sub2video.last_pts)continue;if (pts2 >= ist2->sub2video.end_pts || !ist2->sub2video.frame->data[0])sub2video_update(ist2, NULL);for (j = 0, nb_reqs = 0; j < ist2->nb_filters; j++)nb_reqs += av_buffersrc_get_nb_failed_requests(ist2->filters[j]->filter);if (nb_reqs)sub2video_push_ref(ist2, pts2);} }static void sub2video_flush(InputStream *ist) {int i;if (ist->sub2video.end_pts < INT64_MAX)sub2video_update(ist, NULL);for (i = 0; i < ist->nb_filters; i++)av_buffersrc_add_frame(ist->filters[i]->filter, NULL); }/* end of sub2video hack */static void term_exit_sigsafe(void) { #if HAVE_TERMIOS_Hif(restore_tty)tcsetattr (0, TCSANOW, &oldtty); #endif }void term_exit(void) {av_log(NULL, AV_LOG_QUIET, "%s", "");term_exit_sigsafe(); }static volatile int received_sigterm = 0; static volatile int received_nb_signals = 0; static atomic_int transcode_init_done = ATOMIC_VAR_INIT(0); static volatile int ffmpeg_exited = 0; static int main_return_code = 0;static void sigterm_handler(int sig) {received_sigterm = sig;received_nb_signals++;term_exit_sigsafe();if(received_nb_signals > 3) {write(2/*STDERR_FILENO*/, "Received > 3 system signals, hard exiting\n",strlen("Received > 3 system signals, hard exiting\n"));exit(123);} }#if HAVE_SETCONSOLECTRLHANDLER static BOOL WINAPI CtrlHandler(DWORD fdwCtrlType) {av_log(NULL, AV_LOG_DEBUG, "\nReceived windows signal %ld\n", fdwCtrlType);switch (fdwCtrlType){case CTRL_C_EVENT:case CTRL_BREAK_EVENT:sigterm_handler(SIGINT);return TRUE;case CTRL_CLOSE_EVENT:case CTRL_LOGOFF_EVENT:case CTRL_SHUTDOWN_EVENT:sigterm_handler(SIGTERM);/* Basically, with these 3 events, when we return from this method theprocess is hard terminated, so stall as long as we need toto try and let the main thread(s) clean up and gracefully terminate(we have at most 5 seconds, but should be done far before that). */while (!ffmpeg_exited) {Sleep(0);}return TRUE;default:av_log(NULL, AV_LOG_ERROR, "Received unknown windows signal %ld\n", fdwCtrlType);return FALSE;} } #endifvoid term_init(void) { #if HAVE_TERMIOS_Hif (!run_as_daemon && stdin_interaction) {struct termios tty;if (tcgetattr (0, &tty) == 0) {oldtty = tty;restore_tty = 1;tty.c_iflag &= ~(IGNBRK|BRKINT|PARMRK|ISTRIP|INLCR|IGNCR|ICRNL|IXON);tty.c_oflag |= OPOST;tty.c_lflag &= ~(ECHO|ECHONL|ICANON|IEXTEN);tty.c_cflag &= ~(CSIZE|PARENB);tty.c_cflag |= CS8;tty.c_cc[VMIN] = 1;tty.c_cc[VTIME] = 0;tcsetattr (0, TCSANOW, &tty);}signal(SIGQUIT, sigterm_handler); /* Quit (POSIX). */} #endifsignal(SIGINT , sigterm_handler); /* Interrupt (ANSI). */signal(SIGTERM, sigterm_handler); /* Termination (ANSI). */ #ifdef SIGXCPUsignal(SIGXCPU, sigterm_handler); #endif #if HAVE_SETCONSOLECTRLHANDLERSetConsoleCtrlHandler((PHANDLER_ROUTINE) CtrlHandler, TRUE); #endif }/* read a key without blocking */ static int read_key(void) {unsigned char ch; #if HAVE_TERMIOS_Hint n = 1;struct timeval tv;fd_set rfds;FD_ZERO(&rfds);FD_SET(0, &rfds);tv.tv_sec = 0;tv.tv_usec = 0;n = select(1, &rfds, NULL, NULL, &tv);if (n > 0) {n = read(0, &ch, 1);if (n == 1)return ch;return n;} #elif HAVE_KBHIT # if HAVE_PEEKNAMEDPIPEstatic int is_pipe;static HANDLE input_handle;DWORD dw, nchars;if(!input_handle){input_handle = GetStdHandle(STD_INPUT_HANDLE);is_pipe = !GetConsoleMode(input_handle, &dw);}if (is_pipe) {/* When running under a GUI, you will end here. */if (!PeekNamedPipe(input_handle, NULL, 0, NULL, &nchars, NULL)) {// input pipe may have been closed by the program that ran ffmpegreturn -1;}//Read itif(nchars != 0) {read(0, &ch, 1);return ch;}else{return -1;}} # endifif(kbhit())return(getch()); #endifreturn -1; }static int decode_interrupt_cb(void *ctx) {return received_nb_signals > atomic_load(&transcode_init_done); }const AVIOInterruptCB int_cb = { decode_interrupt_cb, NULL };static void ffmpeg_cleanup(int ret) {int i, j;if (do_benchmark) {int maxrss = getmaxrss() / 1024;av_log(NULL, AV_LOG_INFO, "bench: maxrss=%ikB\n", maxrss);}for (i = 0; i < nb_filtergraphs; i++) {FilterGraph *fg = filtergraphs[i];avfilter_graph_free(&fg->graph);for (j = 0; j < fg->nb_inputs; j++) {while (av_fifo_size(fg->inputs[j]->frame_queue)) {AVFrame *frame;av_fifo_generic_read(fg->inputs[j]->frame_queue, &frame,sizeof(frame), NULL);av_frame_free(&frame);}av_fifo_free(fg->inputs[j]->frame_queue);if (fg->inputs[j]->ist->sub2video.sub_queue) {while (av_fifo_size(fg->inputs[j]->ist->sub2video.sub_queue)) {AVSubtitle sub;av_fifo_generic_read(fg->inputs[j]->ist->sub2video.sub_queue,&sub, sizeof(sub), NULL);avsubtitle_free(&sub);}av_fifo_free(fg->inputs[j]->ist->sub2video.sub_queue);}av_buffer_unref(&fg->inputs[j]->hw_frames_ctx);av_freep(&fg->inputs[j]->name);av_freep(&fg->inputs[j]);}av_freep(&fg->inputs);for (j = 0; j < fg->nb_outputs; j++) {av_freep(&fg->outputs[j]->name);av_freep(&fg->outputs[j]->formats);av_freep(&fg->outputs[j]->channel_layouts);av_freep(&fg->outputs[j]->sample_rates);av_freep(&fg->outputs[j]);}av_freep(&fg->outputs);av_freep(&fg->graph_desc);av_freep(&filtergraphs[i]);}av_freep(&filtergraphs);av_freep(&subtitle_out);/* close files */for (i = 0; i < nb_output_files; i++) {OutputFile *of = output_files[i];AVFormatContext *s;if (!of)continue;s = of->ctx;if (s && s->oformat && !(s->oformat->flags & AVFMT_NOFILE))avio_closep(&s->pb);avformat_free_context(s);av_dict_free(&of->opts);av_freep(&output_files[i]);}for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];if (!ost)continue;for (j = 0; j < ost->nb_bitstream_filters; j++)av_bsf_free(&ost->bsf_ctx[j]);av_freep(&ost->bsf_ctx);av_freep(&ost->bsf_extradata_updated);av_frame_free(&ost->filtered_frame);av_frame_free(&ost->last_frame);av_dict_free(&ost->encoder_opts);av_parser_close(ost->parser);avcodec_free_context(&ost->parser_avctx);av_freep(&ost->forced_keyframes);av_expr_free(ost->forced_keyframes_pexpr);av_freep(&ost->avfilter);av_freep(&ost->logfile_prefix);av_freep(&ost->audio_channels_map);ost->audio_channels_mapped = 0;av_dict_free(&ost->sws_dict);avcodec_free_context(&ost->enc_ctx);avcodec_parameters_free(&ost->ref_par);if (ost->muxing_queue) {while (av_fifo_size(ost->muxing_queue)) {AVPacket pkt;av_fifo_generic_read(ost->muxing_queue, &pkt, sizeof(pkt), NULL);av_packet_unref(&pkt);}av_fifo_freep(&ost->muxing_queue);}av_freep(&output_streams[i]);} #if HAVE_PTHREADSfree_input_threads(); #endiffor (i = 0; i < nb_input_files; i++) {avformat_close_input(&input_files[i]->ctx);av_freep(&input_files[i]);}for (i = 0; i < nb_input_streams; i++) {InputStream *ist = input_streams[i];av_frame_free(&ist->decoded_frame);av_frame_free(&ist->filter_frame);av_dict_free(&ist->decoder_opts);avsubtitle_free(&ist->prev_sub.subtitle);av_frame_free(&ist->sub2video.frame);av_freep(&ist->filters);av_freep(&ist->hwaccel_device);av_freep(&ist->dts_buffer);avcodec_free_context(&ist->dec_ctx);av_freep(&input_streams[i]);}if (vstats_file) {if (fclose(vstats_file))av_log(NULL, AV_LOG_ERROR,"Error closing vstats file, loss of information possible: %s\n",av_err2str(AVERROR(errno)));}av_freep(&vstats_filename);av_freep(&input_streams);av_freep(&input_files);av_freep(&output_streams);av_freep(&output_files);uninit_opts();avformat_network_deinit();if (received_sigterm) {av_log(NULL, AV_LOG_INFO, "Exiting normally, received signal %d.\n",(int) received_sigterm);} else if (ret && atomic_load(&transcode_init_done)) {av_log(NULL, AV_LOG_INFO, "Conversion failed!\n");}term_exit();ffmpeg_exited = 1; }void remove_avoptions(AVDictionary **a, AVDictionary *b) {AVDictionaryEntry *t = NULL;while ((t = av_dict_get(b, "", t, AV_DICT_IGNORE_SUFFIX))) {av_dict_set(a, t->key, NULL, AV_DICT_MATCH_CASE);} }void assert_avoptions(AVDictionary *m) {AVDictionaryEntry *t;if ((t = av_dict_get(m, "", NULL, AV_DICT_IGNORE_SUFFIX))) {av_log(NULL, AV_LOG_FATAL, "Option %s not found.\n", t->key);exit_program(1);} }static void abort_codec_experimental(AVCodec *c, int encoder) {exit_program(1); }static void update_benchmark(const char *fmt, ...) {if (do_benchmark_all) {int64_t t = getutime();va_list va;char buf[1024];if (fmt) {va_start(va, fmt);vsnprintf(buf, sizeof(buf), fmt, va);va_end(va);av_log(NULL, AV_LOG_INFO, "bench: %8"PRIu64" %s \n", t - current_time, buf);}current_time = t;} }static void close_all_output_streams(OutputStream *ost, OSTFinished this_stream, OSTFinished others) {int i;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost2 = output_streams[i];ost2->finished |= ost == ost2 ? this_stream : others;} }static void write_packet(OutputFile *of, AVPacket *pkt, OutputStream *ost, int unqueue) {AVFormatContext *s = of->ctx;AVStream *st = ost->st;int ret;/** Audio encoders may split the packets -- #frames in != #packets out.* But there is no reordering, so we can limit the number of output packets* by simply dropping them here.* Counting encoded video frames needs to be done separately because of* reordering, see do_video_out().* Do not count the packet when unqueued because it has been counted when queued.*/if (!(st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && ost->encoding_needed) && !unqueue) {if (ost->frame_number >= ost->max_frames) {av_packet_unref(pkt);return;}ost->frame_number++;}if (!of->header_written) {AVPacket tmp_pkt = {0};/* the muxer is not initialized yet, buffer the packet */if (!av_fifo_space(ost->muxing_queue)) {int new_size = FFMIN(2 * av_fifo_size(ost->muxing_queue),ost->max_muxing_queue_size);if (new_size <= av_fifo_size(ost->muxing_queue)) {av_log(NULL, AV_LOG_ERROR,"Too many packets buffered for output stream %d:%d.\n",ost->file_index, ost->st->index);exit_program(1);}ret = av_fifo_realloc2(ost->muxing_queue, new_size);if (ret < 0)exit_program(1);}ret = av_packet_ref(&tmp_pkt, pkt);if (ret < 0)exit_program(1);av_fifo_generic_write(ost->muxing_queue, &tmp_pkt, sizeof(tmp_pkt), NULL);av_packet_unref(pkt);return;}if ((st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && video_sync_method == VSYNC_DROP) ||(st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO && audio_sync_method < 0))pkt->pts = pkt->dts = AV_NOPTS_VALUE;if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {int i;uint8_t *sd = av_packet_get_side_data(pkt, AV_PKT_DATA_QUALITY_STATS,NULL);ost->quality = sd ? AV_RL32(sd) : -1;ost->pict_type = sd ? sd[4] : AV_PICTURE_TYPE_NONE;for (i = 0; i<FF_ARRAY_ELEMS(ost->error); i++) {if (sd && i < sd[5])ost->error[i] = AV_RL64(sd + 8 + 8*i);elseost->error[i] = -1;}if (ost->frame_rate.num && ost->is_cfr) {if (pkt->duration > 0)av_log(NULL, AV_LOG_WARNING, "Overriding packet duration by frame rate, this should not happen\n");pkt->duration = av_rescale_q(1, av_inv_q(ost->frame_rate),ost->mux_timebase);}}av_packet_rescale_ts(pkt, ost->mux_timebase, ost->st->time_base);if (!(s->oformat->flags & AVFMT_NOTIMESTAMPS)) {if (pkt->dts != AV_NOPTS_VALUE &&pkt->pts != AV_NOPTS_VALUE &&pkt->dts > pkt->pts) {av_log(s, AV_LOG_WARNING, "Invalid DTS: %"PRId64" PTS: %"PRId64" in output stream %d:%d, replacing by guess\n",pkt->dts, pkt->pts,ost->file_index, ost->st->index);pkt->pts =pkt->dts = pkt->pts + pkt->dts + ost->last_mux_dts + 1- FFMIN3(pkt->pts, pkt->dts, ost->last_mux_dts + 1)- FFMAX3(pkt->pts, pkt->dts, ost->last_mux_dts + 1);}if ((st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO || st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) &&pkt->dts != AV_NOPTS_VALUE &&!(st->codecpar->codec_id == AV_CODEC_ID_VP9 && ost->stream_copy) &&ost->last_mux_dts != AV_NOPTS_VALUE) {int64_t max = ost->last_mux_dts + !(s->oformat->flags & AVFMT_TS_NONSTRICT);if (pkt->dts < max) {int loglevel = max - pkt->dts > 2 || st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO ? AV_LOG_WARNING : AV_LOG_DEBUG;av_log(s, loglevel, "Non-monotonous DTS in output stream ""%d:%d; previous: %"PRId64", current: %"PRId64"; ",ost->file_index, ost->st->index, ost->last_mux_dts, pkt->dts);if (exit_on_error) {av_log(NULL, AV_LOG_FATAL, "aborting.\n");exit_program(1);}av_log(s, loglevel, "changing to %"PRId64". This may result ""in incorrect timestamps in the output file.\n",max);if (pkt->pts >= pkt->dts)pkt->pts = FFMAX(pkt->pts, max);pkt->dts = max;}}}ost->last_mux_dts = pkt->dts;ost->data_size += pkt->size;ost->packets_written++;pkt->stream_index = ost->index;if (debug_ts) {av_log(NULL, AV_LOG_INFO, "muxer <- type:%s ""pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s size:%d\n",av_get_media_type_string(ost->enc_ctx->codec_type),av_ts2str(pkt->pts), av_ts2timestr(pkt->pts, &ost->st->time_base),av_ts2str(pkt->dts), av_ts2timestr(pkt->dts, &ost->st->time_base),pkt->size);}ret = av_interleaved_write_frame(s, pkt);if (ret < 0) {print_error("av_interleaved_write_frame()", ret);main_return_code = 1;close_all_output_streams(ost, MUXER_FINISHED | ENCODER_FINISHED, ENCODER_FINISHED);}av_packet_unref(pkt); }static void close_output_stream(OutputStream *ost) {OutputFile *of = output_files[ost->file_index];ost->finished |= ENCODER_FINISHED;if (of->shortest) {int64_t end = av_rescale_q(ost->sync_opts - ost->first_pts, ost->enc_ctx->time_base, AV_TIME_BASE_Q);of->recording_time = FFMIN(of->recording_time, end);} }static void output_packet(OutputFile *of, AVPacket *pkt, OutputStream *ost) {int ret = 0;/* apply the output bitstream filters, if any */if (ost->nb_bitstream_filters) {int idx;ret = av_bsf_send_packet(ost->bsf_ctx[0], pkt);if (ret < 0)goto finish;idx = 1;while (idx) {/* get a packet from the previous filter up the chain */ret = av_bsf_receive_packet(ost->bsf_ctx[idx - 1], pkt);if (ret == AVERROR(EAGAIN)) {ret = 0;idx--;continue;} else if (ret < 0)goto finish;/* HACK! - aac_adtstoasc updates extradata after filtering the first frame when* the api states this shouldn't happen after init(). Propagate it here to the* muxer and to the next filters in the chain to workaround this.* TODO/FIXME - Make aac_adtstoasc use new packet side data instead of changing* par_out->extradata and adapt muxers accordingly to get rid of this. */if (!(ost->bsf_extradata_updated[idx - 1] & 1)) {ret = avcodec_parameters_copy(ost->st->codecpar, ost->bsf_ctx[idx - 1]->par_out);if (ret < 0)goto finish;ost->bsf_extradata_updated[idx - 1] |= 1;}/* send it to the next filter down the chain or to the muxer */if (idx < ost->nb_bitstream_filters) {/* HACK/FIXME! - See above */if (!(ost->bsf_extradata_updated[idx] & 2)) {ret = avcodec_parameters_copy(ost->bsf_ctx[idx]->par_out, ost->bsf_ctx[idx - 1]->par_out);if (ret < 0)goto finish;ost->bsf_extradata_updated[idx] |= 2;}ret = av_bsf_send_packet(ost->bsf_ctx[idx], pkt);if (ret < 0)goto finish;idx++;} elsewrite_packet(of, pkt, ost, 0);}} elsewrite_packet(of, pkt, ost, 0);finish:if (ret < 0 && ret != AVERROR_EOF) {av_log(NULL, AV_LOG_ERROR, "Error applying bitstream filters to an output ""packet for stream #%d:%d.\n", ost->file_index, ost->index);if(exit_on_error)exit_program(1);} }static int check_recording_time(OutputStream *ost) {OutputFile *of = output_files[ost->file_index];if (of->recording_time != INT64_MAX &&av_compare_ts(ost->sync_opts - ost->first_pts, ost->enc_ctx->time_base, of->recording_time,AV_TIME_BASE_Q) >= 0) {close_output_stream(ost);return 0;}return 1; }static void do_audio_out(OutputFile *of, OutputStream *ost,AVFrame *frame) {AVCodecContext *enc = ost->enc_ctx;AVPacket pkt;int ret;av_init_packet(&pkt);pkt.data = NULL;pkt.size = 0;if (!check_recording_time(ost))return;if (frame->pts == AV_NOPTS_VALUE || audio_sync_method < 0)frame->pts = ost->sync_opts;ost->sync_opts = frame->pts + frame->nb_samples;ost->samples_encoded += frame->nb_samples;ost->frames_encoded++;av_assert0(pkt.size || !pkt.data);update_benchmark(NULL);if (debug_ts) {av_log(NULL, AV_LOG_INFO, "encoder <- type:audio ""frame_pts:%s frame_pts_time:%s time_base:%d/%d\n",av_ts2str(frame->pts), av_ts2timestr(frame->pts, &enc->time_base),enc->time_base.num, enc->time_base.den);}ret = avcodec_send_frame(enc, frame);if (ret < 0)goto error;while (1) {ret = avcodec_receive_packet(enc, &pkt);if (ret == AVERROR(EAGAIN))break;if (ret < 0)goto error;update_benchmark("encode_audio %d.%d", ost->file_index, ost->index);av_packet_rescale_ts(&pkt, enc->time_base, ost->mux_timebase);if (debug_ts) {av_log(NULL, AV_LOG_INFO, "encoder -> type:audio ""pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s\n",av_ts2str(pkt.pts), av_ts2timestr(pkt.pts, &enc->time_base),av_ts2str(pkt.dts), av_ts2timestr(pkt.dts, &enc->time_base));}output_packet(of, &pkt, ost);}return; error:av_log(NULL, AV_LOG_FATAL, "Audio encoding failed\n");exit_program(1); }static void do_subtitle_out(OutputFile *of,OutputStream *ost,AVSubtitle *sub) {int subtitle_out_max_size = 1024 * 1024;int subtitle_out_size, nb, i;AVCodecContext *enc;AVPacket pkt;int64_t pts;if (sub->pts == AV_NOPTS_VALUE) {av_log(NULL, AV_LOG_ERROR, "Subtitle packets must have a pts\n");if (exit_on_error)exit_program(1);return;}enc = ost->enc_ctx;if (!subtitle_out) {subtitle_out = av_malloc(subtitle_out_max_size);if (!subtitle_out) {av_log(NULL, AV_LOG_FATAL, "Failed to allocate subtitle_out\n");exit_program(1);}}/* Note: DVB subtitle need one packet to draw them and one otherpacket to clear them *//* XXX: signal it in the codec context ? */if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE)nb = 2;elsenb = 1;/* shift timestamp to honor -ss and make check_recording_time() work with -t */pts = sub->pts;if (output_files[ost->file_index]->start_time != AV_NOPTS_VALUE)pts -= output_files[ost->file_index]->start_time;for (i = 0; i < nb; i++) {unsigned save_num_rects = sub->num_rects;ost->sync_opts = av_rescale_q(pts, AV_TIME_BASE_Q, enc->time_base);if (!check_recording_time(ost))return;sub->pts = pts;// start_display_time is required to be 0sub->pts += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, AV_TIME_BASE_Q);sub->end_display_time -= sub->start_display_time;sub->start_display_time = 0;if (i == 1)sub->num_rects = 0;ost->frames_encoded++;subtitle_out_size = avcodec_encode_subtitle(enc, subtitle_out,subtitle_out_max_size, sub);if (i == 1)sub->num_rects = save_num_rects;if (subtitle_out_size < 0) {av_log(NULL, AV_LOG_FATAL, "Subtitle encoding failed\n");exit_program(1);}av_init_packet(&pkt);pkt.data = subtitle_out;pkt.size = subtitle_out_size;pkt.pts = av_rescale_q(sub->pts, AV_TIME_BASE_Q, ost->mux_timebase);pkt.duration = av_rescale_q(sub->end_display_time, (AVRational){ 1, 1000 }, ost->mux_timebase);if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE) {/* XXX: the pts correction is handled here. Maybe handlingit in the codec would be better */if (i == 0)pkt.pts += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, ost->mux_timebase);elsepkt.pts += av_rescale_q(sub->end_display_time, (AVRational){ 1, 1000 }, ost->mux_timebase);}pkt.dts = pkt.pts;output_packet(of, &pkt, ost);} }static void do_video_out(OutputFile *of,OutputStream *ost,AVFrame *next_picture,double sync_ipts) {int ret, format_video_sync;AVPacket pkt;AVCodecContext *enc = ost->enc_ctx;AVCodecParameters *mux_par = ost->st->codecpar;AVRational frame_rate;int nb_frames, nb0_frames, i;double delta, delta0;double duration = 0;int frame_size = 0;InputStream *ist = NULL;AVFilterContext *filter = ost->filter->filter;if (ost->source_index >= 0)ist = input_streams[ost->source_index];frame_rate = av_buffersink_get_frame_rate(filter);if (frame_rate.num > 0 && frame_rate.den > 0)duration = 1/(av_q2d(frame_rate) * av_q2d(enc->time_base));if(ist && ist->st->start_time != AV_NOPTS_VALUE && ist->st->first_dts != AV_NOPTS_VALUE && ost->frame_rate.num)duration = FFMIN(duration, 1/(av_q2d(ost->frame_rate) * av_q2d(enc->time_base)));if (!ost->filters_script &&!ost->filters &&next_picture &&ist &&lrintf(av_frame_get_pkt_duration(next_picture) * av_q2d(ist->st->time_base) / av_q2d(enc->time_base)) > 0) {duration = lrintf(av_frame_get_pkt_duration(next_picture) * av_q2d(ist->st->time_base) / av_q2d(enc->time_base));}if (!next_picture) {//end, flushingnb0_frames = nb_frames = mid_pred(ost->last_nb0_frames[0],ost->last_nb0_frames[1],ost->last_nb0_frames[2]);} else {delta0 = sync_ipts - ost->sync_opts; // delta0 is the "drift" between the input frame (next_picture) and where it would fall in the output.delta = delta0 + duration;/* by default, we output a single frame */nb0_frames = 0; // tracks the number of times the PREVIOUS frame should be duplicated, mostly for variable framerate (VFR)nb_frames = 1;format_video_sync = video_sync_method;if (format_video_sync == VSYNC_AUTO) {if(!strcmp(of->ctx->oformat->name, "avi")) {format_video_sync = VSYNC_VFR;} elseformat_video_sync = (of->ctx->oformat->flags & AVFMT_VARIABLE_FPS) ? ((of->ctx->oformat->flags & AVFMT_NOTIMESTAMPS) ? VSYNC_PASSTHROUGH : VSYNC_VFR) : VSYNC_CFR;if ( ist&& format_video_sync == VSYNC_CFR&& input_files[ist->file_index]->ctx->nb_streams == 1&& input_files[ist->file_index]->input_ts_offset == 0) {format_video_sync = VSYNC_VSCFR;}if (format_video_sync == VSYNC_CFR && copy_ts) {format_video_sync = VSYNC_VSCFR;}}ost->is_cfr = (format_video_sync == VSYNC_CFR || format_video_sync == VSYNC_VSCFR);if (delta0 < 0 &&delta > 0 &&format_video_sync != VSYNC_PASSTHROUGH &&format_video_sync != VSYNC_DROP) {if (delta0 < -0.6) {av_log(NULL, AV_LOG_WARNING, "Past duration %f too large\n", -delta0);} elseav_log(NULL, AV_LOG_DEBUG, "Clipping frame in rate conversion by %f\n", -delta0);sync_ipts = ost->sync_opts;duration += delta0;delta0 = 0;}switch (format_video_sync) {case VSYNC_VSCFR:if (ost->frame_number == 0 && delta0 >= 0.5) {av_log(NULL, AV_LOG_DEBUG, "Not duplicating %d initial frames\n", (int)lrintf(delta0));delta = duration;delta0 = 0;ost->sync_opts = lrint(sync_ipts);}case VSYNC_CFR:// FIXME set to 0.5 after we fix some dts/pts bugs like in avidec.cif (frame_drop_threshold && delta < frame_drop_threshold && ost->frame_number) {nb_frames = 0;} else if (delta < -1.1)nb_frames = 0;else if (delta > 1.1) {nb_frames = lrintf(delta);if (delta0 > 1.1)nb0_frames = lrintf(delta0 - 0.6);}break;case VSYNC_VFR:if (delta <= -0.6)nb_frames = 0;else if (delta > 0.6)ost->sync_opts = lrint(sync_ipts);break;case VSYNC_DROP:case VSYNC_PASSTHROUGH:ost->sync_opts = lrint(sync_ipts);break;default:av_assert0(0);}}nb_frames = FFMIN(nb_frames, ost->max_frames - ost->frame_number);nb0_frames = FFMIN(nb0_frames, nb_frames);memmove(ost->last_nb0_frames + 1,ost->last_nb0_frames,sizeof(ost->last_nb0_frames[0]) * (FF_ARRAY_ELEMS(ost->last_nb0_frames) - 1));ost->last_nb0_frames[0] = nb0_frames;if (nb0_frames == 0 && ost->last_dropped) {nb_frames_drop++;av_log(NULL, AV_LOG_VERBOSE,"*** dropping frame %d from stream %d at ts %"PRId64"\n",ost->frame_number, ost->st->index, ost->last_frame->pts);}if (nb_frames > (nb0_frames && ost->last_dropped) + (nb_frames > nb0_frames)) {if (nb_frames > dts_error_threshold * 30) {av_log(NULL, AV_LOG_ERROR, "%d frame duplication too large, skipping\n", nb_frames - 1);nb_frames_drop++;return;}nb_frames_dup += nb_frames - (nb0_frames && ost->last_dropped) - (nb_frames > nb0_frames);av_log(NULL, AV_LOG_VERBOSE, "*** %d dup!\n", nb_frames - 1);if (nb_frames_dup > dup_warning) {av_log(NULL, AV_LOG_WARNING, "More than %d frames duplicated\n", dup_warning);dup_warning *= 10;}}ost->last_dropped = nb_frames == nb0_frames && next_picture;/* duplicates frame if needed */for (i = 0; i < nb_frames; i++) {AVFrame *in_picture;av_init_packet(&pkt);pkt.data = NULL;pkt.size = 0;if (i < nb0_frames && ost->last_frame) {in_picture = ost->last_frame;} elsein_picture = next_picture;if (!in_picture)return;in_picture->pts = ost->sync_opts;#if 1if (!check_recording_time(ost)) #elseif (ost->frame_number >= ost->max_frames) #endifreturn;#if FF_API_LAVF_FMT_RAWPICTUREif (of->ctx->oformat->flags & AVFMT_RAWPICTURE &&enc->codec->id == AV_CODEC_ID_RAWVIDEO) {/* raw pictures are written as AVPicture structure toavoid any copies. We support temporarily the oldermethod. */if (in_picture->interlaced_frame)mux_par->field_order = in_picture->top_field_first ? AV_FIELD_TB:AV_FIELD_BT;elsemux_par->field_order = AV_FIELD_PROGRESSIVE;pkt.data = (uint8_t *)in_picture;pkt.size = sizeof(AVPicture);pkt.pts = av_rescale_q(in_picture->pts, enc->time_base, ost->mux_timebase);pkt.flags |= AV_PKT_FLAG_KEY;output_packet(of, &pkt, ost);} else #endif{int forced_keyframe = 0;double pts_time;if (enc->flags & (AV_CODEC_FLAG_INTERLACED_DCT | AV_CODEC_FLAG_INTERLACED_ME) &&ost->top_field_first >= 0)in_picture->top_field_first = !!ost->top_field_first;if (in_picture->interlaced_frame) {if (enc->codec->id == AV_CODEC_ID_MJPEG)mux_par->field_order = in_picture->top_field_first ? AV_FIELD_TT:AV_FIELD_BB;elsemux_par->field_order = in_picture->top_field_first ? AV_FIELD_TB:AV_FIELD_BT;} elsemux_par->field_order = AV_FIELD_PROGRESSIVE;in_picture->quality = enc->global_quality;in_picture->pict_type = 0;pts_time = in_picture->pts != AV_NOPTS_VALUE ?in_picture->pts * av_q2d(enc->time_base) : NAN;if (ost->forced_kf_index < ost->forced_kf_count &&in_picture->pts >= ost->forced_kf_pts[ost->forced_kf_index]) {ost->forced_kf_index++;forced_keyframe = 1;} else if (ost->forced_keyframes_pexpr) {double res;ost->forced_keyframes_expr_const_values[FKF_T] = pts_time;res = av_expr_eval(ost->forced_keyframes_pexpr,ost->forced_keyframes_expr_const_values, NULL);ff_dlog(NULL, "force_key_frame: n:%f n_forced:%f prev_forced_n:%f t:%f prev_forced_t:%f -> res:%f\n",ost->forced_keyframes_expr_const_values[FKF_N],ost->forced_keyframes_expr_const_values[FKF_N_FORCED],ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_N],ost->forced_keyframes_expr_const_values[FKF_T],ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_T],res);if (res) {forced_keyframe = 1;ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_N] =ost->forced_keyframes_expr_const_values[FKF_N];ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_T] =ost->forced_keyframes_expr_const_values[FKF_T];ost->forced_keyframes_expr_const_values[FKF_N_FORCED] += 1;}ost->forced_keyframes_expr_const_values[FKF_N] += 1;} else if ( ost->forced_keyframes&& !strncmp(ost->forced_keyframes, "source", 6)&& in_picture->key_frame==1) {forced_keyframe = 1;}if (forced_keyframe) {in_picture->pict_type = AV_PICTURE_TYPE_I;av_log(NULL, AV_LOG_DEBUG, "Forced keyframe at time %f\n", pts_time);}update_benchmark(NULL);if (debug_ts) {av_log(NULL, AV_LOG_INFO, "encoder <- type:video ""frame_pts:%s frame_pts_time:%s time_base:%d/%d\n",av_ts2str(in_picture->pts), av_ts2timestr(in_picture->pts, &enc->time_base),enc->time_base.num, enc->time_base.den);}ost->frames_encoded++;ret = avcodec_send_frame(enc, in_picture);if (ret < 0)goto error;while (1) {ret = avcodec_receive_packet(enc, &pkt);update_benchmark("encode_video %d.%d", ost->file_index, ost->index);if (ret == AVERROR(EAGAIN))break;if (ret < 0)goto error;if (debug_ts) {av_log(NULL, AV_LOG_INFO, "encoder -> type:video ""pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s\n",av_ts2str(pkt.pts), av_ts2timestr(pkt.pts, &enc->time_base),av_ts2str(pkt.dts), av_ts2timestr(pkt.dts, &enc->time_base));}if (pkt.pts == AV_NOPTS_VALUE && !(enc->codec->capabilities & AV_CODEC_CAP_DELAY))pkt.pts = ost->sync_opts;av_packet_rescale_ts(&pkt, enc->time_base, ost->mux_timebase);if (debug_ts) {av_log(NULL, AV_LOG_INFO, "encoder -> type:video ""pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s\n",av_ts2str(pkt.pts), av_ts2timestr(pkt.pts, &ost->mux_timebase),av_ts2str(pkt.dts), av_ts2timestr(pkt.dts, &ost->mux_timebase));}frame_size = pkt.size;output_packet(of, &pkt, ost);/* if two pass, output log */if (ost->logfile && enc->stats_out) {fprintf(ost->logfile, "%s", enc->stats_out);}}}ost->sync_opts++;/** For video, number of frames in == number of packets out.* But there may be reordering, so we can't throw away frames on encoder* flush, we need to limit them here, before they go into encoder.*/ost->frame_number++;if (vstats_filename && frame_size)do_video_stats(ost, frame_size);}if (!ost->last_frame)ost->last_frame = av_frame_alloc();av_frame_unref(ost->last_frame);if (next_picture && ost->last_frame)av_frame_ref(ost->last_frame, next_picture);elseav_frame_free(&ost->last_frame);return; error:av_log(NULL, AV_LOG_FATAL, "Video encoding failed\n");exit_program(1); }static double psnr(double d) {return -10.0 * log10(d); }static void do_video_stats(OutputStream *ost, int frame_size) {AVCodecContext *enc;int frame_number;double ti1, bitrate, avg_bitrate;/* this is executed just the first time do_video_stats is called */if (!vstats_file) {vstats_file = fopen(vstats_filename, "w");if (!vstats_file) {perror("fopen");exit_program(1);}}enc = ost->enc_ctx;if (enc->codec_type == AVMEDIA_TYPE_VIDEO) {frame_number = ost->st->nb_frames;if (vstats_version <= 1) {fprintf(vstats_file, "frame= %5d q= %2.1f ", frame_number,ost->quality / (float)FF_QP2LAMBDA);} else {fprintf(vstats_file, "out= %2d st= %2d frame= %5d q= %2.1f ", ost->file_index, ost->index, frame_number,ost->quality / (float)FF_QP2LAMBDA);}if (ost->error[0]>=0 && (enc->flags & AV_CODEC_FLAG_PSNR))fprintf(vstats_file, "PSNR= %6.2f ", psnr(ost->error[0] / (enc->width * enc->height * 255.0 * 255.0)));fprintf(vstats_file,"f_size= %6d ", frame_size);/* compute pts value */ti1 = av_stream_get_end_pts(ost->st) * av_q2d(ost->st->time_base);if (ti1 < 0.01)ti1 = 0.01;bitrate = (frame_size * 8) / av_q2d(enc->time_base) / 1000.0;avg_bitrate = (double)(ost->data_size * 8) / ti1 / 1000.0;fprintf(vstats_file, "s_size= %8.0fkB time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ",(double)ost->data_size / 1024, ti1, bitrate, avg_bitrate);fprintf(vstats_file, "type= %c\n", av_get_picture_type_char(ost->pict_type));} }static int init_output_stream(OutputStream *ost, char *error, int error_len);static void finish_output_stream(OutputStream *ost) {OutputFile *of = output_files[ost->file_index];int i;ost->finished = ENCODER_FINISHED | MUXER_FINISHED;if (of->shortest) {for (i = 0; i < of->ctx->nb_streams; i++)output_streams[of->ost_index + i]->finished = ENCODER_FINISHED | MUXER_FINISHED;} }/*** Get and encode new output from any of the filtergraphs, without causing* activity.** @return 0 for success, <0 for severe errors*/ static int reap_filters(int flush) {AVFrame *filtered_frame = NULL;int i;/* Reap all buffers present in the buffer sinks */for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];OutputFile *of = output_files[ost->file_index];AVFilterContext *filter;AVCodecContext *enc = ost->enc_ctx;int ret = 0;if (!ost->filter || !ost->filter->graph->graph)continue;filter = ost->filter->filter;if (!ost->initialized) {char error[1024] = "";ret = init_output_stream(ost, error, sizeof(error));if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error initializing output stream %d:%d -- %s\n",ost->file_index, ost->index, error);exit_program(1);}}if (!ost->filtered_frame && !(ost->filtered_frame = av_frame_alloc())) {return AVERROR(ENOMEM);}filtered_frame = ost->filtered_frame;while (1) {double float_pts = AV_NOPTS_VALUE; // this is identical to filtered_frame.pts but with higher precisionret = av_buffersink_get_frame_flags(filter, filtered_frame,AV_BUFFERSINK_FLAG_NO_REQUEST);if (ret < 0) {if (ret != AVERROR(EAGAIN) && ret != AVERROR_EOF) {av_log(NULL, AV_LOG_WARNING,"Error in av_buffersink_get_frame_flags(): %s\n", av_err2str(ret));} else if (flush && ret == AVERROR_EOF) {if (av_buffersink_get_type(filter) == AVMEDIA_TYPE_VIDEO)do_video_out(of, ost, NULL, AV_NOPTS_VALUE);}break;}if (ost->finished) {av_frame_unref(filtered_frame);continue;}if (filtered_frame->pts != AV_NOPTS_VALUE) {int64_t start_time = (of->start_time == AV_NOPTS_VALUE) ? 0 : of->start_time;AVRational filter_tb = av_buffersink_get_time_base(filter);AVRational tb = enc->time_base;int extra_bits = av_clip(29 - av_log2(tb.den), 0, 16);tb.den <<= extra_bits;float_pts =av_rescale_q(filtered_frame->pts, filter_tb, tb) -av_rescale_q(start_time, AV_TIME_BASE_Q, tb);float_pts /= 1 << extra_bits;// avoid exact midoints to reduce the chance of rounding differences, this can be removed in case the fps code is changed to work with integersfloat_pts += FFSIGN(float_pts) * 1.0 / (1<<17);filtered_frame->pts =av_rescale_q(filtered_frame->pts, filter_tb, enc->time_base) -av_rescale_q(start_time, AV_TIME_BASE_Q, enc->time_base);}//if (ost->source_index >= 0)// *filtered_frame= *input_streams[ost->source_index]->decoded_frame; //for me_thresholdswitch (av_buffersink_get_type(filter)) {case AVMEDIA_TYPE_VIDEO:if (!ost->frame_aspect_ratio.num)enc->sample_aspect_ratio = filtered_frame->sample_aspect_ratio;if (debug_ts) {av_log(NULL, AV_LOG_INFO, "filter -> pts:%s pts_time:%s exact:%f time_base:%d/%d\n",av_ts2str(filtered_frame->pts), av_ts2timestr(filtered_frame->pts, &enc->time_base),float_pts,enc->time_base.num, enc->time_base.den);}do_video_out(of, ost, filtered_frame, float_pts);break;case AVMEDIA_TYPE_AUDIO:if (!(enc->codec->capabilities & AV_CODEC_CAP_PARAM_CHANGE) &&enc->channels != av_frame_get_channels(filtered_frame)) {av_log(NULL, AV_LOG_ERROR,"Audio filter graph output is not normalized and encoder does not support parameter changes\n");break;}do_audio_out(of, ost, filtered_frame);break;default:// TODO support subtitle filtersav_assert0(0);}av_frame_unref(filtered_frame);}}return 0; }static void print_final_stats(int64_t total_size) {uint64_t video_size = 0, audio_size = 0, extra_size = 0, other_size = 0;uint64_t subtitle_size = 0;uint64_t data_size = 0;float percent = -1.0;int i, j;int pass1_used = 1;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];switch (ost->enc_ctx->codec_type) {case AVMEDIA_TYPE_VIDEO: video_size += ost->data_size; break;case AVMEDIA_TYPE_AUDIO: audio_size += ost->data_size; break;case AVMEDIA_TYPE_SUBTITLE: subtitle_size += ost->data_size; break;default: other_size += ost->data_size; break;}extra_size += ost->enc_ctx->extradata_size;data_size += ost->data_size;if ( (ost->enc_ctx->flags & (AV_CODEC_FLAG_PASS1 | AV_CODEC_FLAG_PASS2))!= AV_CODEC_FLAG_PASS1)pass1_used = 0;}if (data_size && total_size>0 && total_size >= data_size)percent = 100.0 * (total_size - data_size) / data_size;av_log(NULL, AV_LOG_INFO, "video:%1.0fkB audio:%1.0fkB subtitle:%1.0fkB other streams:%1.0fkB global headers:%1.0fkB muxing overhead: ",video_size / 1024.0,audio_size / 1024.0,subtitle_size / 1024.0,other_size / 1024.0,extra_size / 1024.0);if (percent >= 0.0)av_log(NULL, AV_LOG_INFO, "%f%%", percent);elseav_log(NULL, AV_LOG_INFO, "unknown");av_log(NULL, AV_LOG_INFO, "\n");/* print verbose per-stream stats */for (i = 0; i < nb_input_files; i++) {InputFile *f = input_files[i];uint64_t total_packets = 0, total_size = 0;av_log(NULL, AV_LOG_VERBOSE, "Input file #%d (%s):\n",i, f->ctx->filename);for (j = 0; j < f->nb_streams; j++) {InputStream *ist = input_streams[f->ist_index + j];enum AVMediaType type = ist->dec_ctx->codec_type;total_size += ist->data_size;total_packets += ist->nb_packets;av_log(NULL, AV_LOG_VERBOSE, " Input stream #%d:%d (%s): ",i, j, media_type_string(type));av_log(NULL, AV_LOG_VERBOSE, "%"PRIu64" packets read (%"PRIu64" bytes); ",ist->nb_packets, ist->data_size);if (ist->decoding_needed) {av_log(NULL, AV_LOG_VERBOSE, "%"PRIu64" frames decoded",ist->frames_decoded);if (type == AVMEDIA_TYPE_AUDIO)av_log(NULL, AV_LOG_VERBOSE, " (%"PRIu64" samples)", ist->samples_decoded);av_log(NULL, AV_LOG_VERBOSE, "; ");}av_log(NULL, AV_LOG_VERBOSE, "\n");}av_log(NULL, AV_LOG_VERBOSE, " Total: %"PRIu64" packets (%"PRIu64" bytes) demuxed\n",total_packets, total_size);}for (i = 0; i < nb_output_files; i++) {OutputFile *of = output_files[i];uint64_t total_packets = 0, total_size = 0;av_log(NULL, AV_LOG_VERBOSE, "Output file #%d (%s):\n",i, of->ctx->filename);for (j = 0; j < of->ctx->nb_streams; j++) {OutputStream *ost = output_streams[of->ost_index + j];enum AVMediaType type = ost->enc_ctx->codec_type;total_size += ost->data_size;total_packets += ost->packets_written;av_log(NULL, AV_LOG_VERBOSE, " Output stream #%d:%d (%s): ",i, j, media_type_string(type));if (ost->encoding_needed) {av_log(NULL, AV_LOG_VERBOSE, "%"PRIu64" frames encoded",ost->frames_encoded);if (type == AVMEDIA_TYPE_AUDIO)av_log(NULL, AV_LOG_VERBOSE, " (%"PRIu64" samples)", ost->samples_encoded);av_log(NULL, AV_LOG_VERBOSE, "; ");}av_log(NULL, AV_LOG_VERBOSE, "%"PRIu64" packets muxed (%"PRIu64" bytes); ",ost->packets_written, ost->data_size);av_log(NULL, AV_LOG_VERBOSE, "\n");}av_log(NULL, AV_LOG_VERBOSE, " Total: %"PRIu64" packets (%"PRIu64" bytes) muxed\n",total_packets, total_size);}if(video_size + data_size + audio_size + subtitle_size + extra_size == 0){av_log(NULL, AV_LOG_WARNING, "Output file is empty, nothing was encoded ");if (pass1_used) {av_log(NULL, AV_LOG_WARNING, "\n");} else {av_log(NULL, AV_LOG_WARNING, "(check -ss / -t / -frames parameters if used)\n");}} } //打印輸出信息的方法 static void print_report(int is_last_report, int64_t timer_start, int64_t cur_time,void(call_back)(int current,int total)) {char buf[1024];AVBPrint buf_script;OutputStream *ost;AVFormatContext *oc;int64_t total_size;AVCodecContext *enc;int frame_number, vid, i;double bitrate;double speed;int64_t pts = INT64_MIN + 1;static int64_t last_time = -1;static int qp_histogram[52];int hours, mins, secs, us;int ret;float t;if (!print_stats && !is_last_report && !progress_avio)return;if (!is_last_report) {if (last_time == -1) {last_time = cur_time;return;}if ((cur_time - last_time) < 500000)return;last_time = cur_time;}t = (cur_time-timer_start) / 1000000.0;oc = output_files[0]->ctx;total_size = avio_size(oc->pb);if (total_size <= 0) // FIXME improve avio_size() so it works with non seekable output toototal_size = avio_tell(oc->pb);buf[0] = '\0';vid = 0;av_bprint_init(&buf_script, 0, 1);for (i = 0; i < nb_output_streams; i++) {float q = -1;ost = output_streams[i];enc = ost->enc_ctx;if (!ost->stream_copy)q = ost->quality / (float) FF_QP2LAMBDA;if (vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "q=%2.1f ", q);av_bprintf(&buf_script, "stream_%d_%d_q=%.1f\n",ost->file_index, ost->index, q);}if (!vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {float fps;frame_number = ost->frame_number;fps = t > 1 ? frame_number / t : 0;snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "frame=%5d fps=%3.*f q=%3.1f ",frame_number, fps < 9.95, fps, q);av_bprintf(&buf_script, "frame=%d\n", frame_number);av_bprintf(&buf_script, "fps=%.1f\n", fps);av_bprintf(&buf_script, "stream_%d_%d_q=%.1f\n",ost->file_index, ost->index, q);//__android_log_print(ANDROID_LOG_ERROR,"JNI_TAG","當(dāng)前壓縮幀:%d",frame_number);int total_frame_number=input_files[0]->ctx->streams[0]->nb_frames;call_back(frame_number,total_frame_number);if (is_last_report)snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "L");if (qp_hist) {int j;int qp = lrintf(q);if (qp >= 0 && qp < FF_ARRAY_ELEMS(qp_histogram))qp_histogram[qp]++;for (j = 0; j < 32; j++)snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%X", av_log2(qp_histogram[j] + 1));}if ((enc->flags & AV_CODEC_FLAG_PSNR) && (ost->pict_type != AV_PICTURE_TYPE_NONE || is_last_report)) {int j;double error, error_sum = 0;double scale, scale_sum = 0;double p;char type[3] = { 'Y','U','V' };snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "PSNR=");for (j = 0; j < 3; j++) {if (is_last_report) {error = enc->error[j];scale = enc->width * enc->height * 255.0 * 255.0 * frame_number;} else {error = ost->error[j];scale = enc->width * enc->height * 255.0 * 255.0;}if (j)scale /= 4;error_sum += error;scale_sum += scale;p = psnr(error / scale);snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%c:%2.2f ", type[j], p);av_bprintf(&buf_script, "stream_%d_%d_psnr_%c=%2.2f\n",ost->file_index, ost->index, type[j] | 32, p);}p = psnr(error_sum / scale_sum);snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "*:%2.2f ", psnr(error_sum / scale_sum));av_bprintf(&buf_script, "stream_%d_%d_psnr_all=%2.2f\n",ost->file_index, ost->index, p);}vid = 1;}/* compute min output value */if (av_stream_get_end_pts(ost->st) != AV_NOPTS_VALUE)pts = FFMAX(pts, av_rescale_q(av_stream_get_end_pts(ost->st),ost->st->time_base, AV_TIME_BASE_Q));if (is_last_report)nb_frames_drop += ost->last_dropped;}secs = FFABS(pts) / AV_TIME_BASE;us = FFABS(pts) % AV_TIME_BASE;mins = secs / 60;secs %= 60;hours = mins / 60;mins %= 60;bitrate = pts && total_size >= 0 ? total_size * 8 / (pts / 1000.0) : -1;speed = t != 0.0 ? (double)pts / AV_TIME_BASE / t : -1;if (total_size < 0) snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),"size=N/A time=");else snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),"size=%8.0fkB time=", total_size / 1024.0);if (pts < 0)snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "-");snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),"%02d:%02d:%02d.%02d ", hours, mins, secs,(100 * us) / AV_TIME_BASE);if (bitrate < 0) {snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),"bitrate=N/A");av_bprintf(&buf_script, "bitrate=N/A\n");}else{snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),"bitrate=%6.1fkbits/s", bitrate);av_bprintf(&buf_script, "bitrate=%6.1fkbits/s\n", bitrate);}if (total_size < 0) av_bprintf(&buf_script, "total_size=N/A\n");else av_bprintf(&buf_script, "total_size=%"PRId64"\n", total_size);av_bprintf(&buf_script, "out_time_ms=%"PRId64"\n", pts);av_bprintf(&buf_script, "out_time=%02d:%02d:%02d.%06d\n",hours, mins, secs, us);if (nb_frames_dup || nb_frames_drop)snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), " dup=%d drop=%d",nb_frames_dup, nb_frames_drop);av_bprintf(&buf_script, "dup_frames=%d\n", nb_frames_dup);av_bprintf(&buf_script, "drop_frames=%d\n", nb_frames_drop);if (speed < 0) {snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf)," speed=N/A");av_bprintf(&buf_script, "speed=N/A\n");} else {snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf)," speed=%4.3gx", speed);av_bprintf(&buf_script, "speed=%4.3gx\n", speed);}if (print_stats || is_last_report) {const char end = is_last_report ? '\n' : '\r';if (print_stats==1 && AV_LOG_INFO > av_log_get_level()) {fprintf(stderr, "%s %c", buf, end);} elseav_log(NULL, AV_LOG_INFO, "%s %c", buf, end);fflush(stderr);}if (progress_avio) {av_bprintf(&buf_script, "progress=%s\n",is_last_report ? "end" : "continue");avio_write(progress_avio, buf_script.str,FFMIN(buf_script.len, buf_script.size - 1));avio_flush(progress_avio);av_bprint_finalize(&buf_script, NULL);if (is_last_report) {if ((ret = avio_closep(&progress_avio)) < 0)av_log(NULL, AV_LOG_ERROR,"Error closing progress log, loss of information possible: %s\n", av_err2str(ret));}}if (is_last_report)print_final_stats(total_size); }static void flush_encoders(void) {int i, ret;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];AVCodecContext *enc = ost->enc_ctx;OutputFile *of = output_files[ost->file_index];if (!ost->encoding_needed)continue;// Try to enable encoding with no input frames.// Maybe we should just let encoding fail instead.if (!ost->initialized) {FilterGraph *fg = ost->filter->graph;char error[1024] = "";av_log(NULL, AV_LOG_WARNING,"Finishing stream %d:%d without any data written to it.\n",ost->file_index, ost->st->index);if (ost->filter && !fg->graph) {int x;for (x = 0; x < fg->nb_inputs; x++) {InputFilter *ifilter = fg->inputs[x];if (ifilter->format < 0) {AVCodecParameters *par = ifilter->ist->st->codecpar;// We never got any input. Set a fake format, which will// come from libavformat.ifilter->format = par->format;ifilter->sample_rate = par->sample_rate;ifilter->channels = par->channels;ifilter->channel_layout = par->channel_layout;ifilter->width = par->width;ifilter->height = par->height;ifilter->sample_aspect_ratio = par->sample_aspect_ratio;}}if (!ifilter_has_all_input_formats(fg))continue;ret = configure_filtergraph(fg);if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error configuring filter graph\n");exit_program(1);}finish_output_stream(ost);}ret = init_output_stream(ost, error, sizeof(error));if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error initializing output stream %d:%d -- %s\n",ost->file_index, ost->index, error);exit_program(1);}}if (enc->codec_type == AVMEDIA_TYPE_AUDIO && enc->frame_size <= 1)continue; #if FF_API_LAVF_FMT_RAWPICTUREif (enc->codec_type == AVMEDIA_TYPE_VIDEO && (of->ctx->oformat->flags & AVFMT_RAWPICTURE) && enc->codec->id == AV_CODEC_ID_RAWVIDEO)continue; #endifif (enc->codec_type != AVMEDIA_TYPE_VIDEO && enc->codec_type != AVMEDIA_TYPE_AUDIO)continue;for (;;) {const char *desc = NULL;AVPacket pkt;int pkt_size;switch (enc->codec_type) {case AVMEDIA_TYPE_AUDIO:desc = "audio";break;case AVMEDIA_TYPE_VIDEO:desc = "video";break;default:av_assert0(0);}av_init_packet(&pkt);pkt.data = NULL;pkt.size = 0;update_benchmark(NULL);while ((ret = avcodec_receive_packet(enc, &pkt)) == AVERROR(EAGAIN)) {ret = avcodec_send_frame(enc, NULL);if (ret < 0) {av_log(NULL, AV_LOG_FATAL, "%s encoding failed: %s\n",desc,av_err2str(ret));exit_program(1);}}update_benchmark("flush_%s %d.%d", desc, ost->file_index, ost->index);if (ret < 0 && ret != AVERROR_EOF) {av_log(NULL, AV_LOG_FATAL, "%s encoding failed: %s\n",desc,av_err2str(ret));exit_program(1);}if (ost->logfile && enc->stats_out) {fprintf(ost->logfile, "%s", enc->stats_out);}if (ret == AVERROR_EOF) {break;}if (ost->finished & MUXER_FINISHED) {av_packet_unref(&pkt);continue;}av_packet_rescale_ts(&pkt, enc->time_base, ost->mux_timebase);pkt_size = pkt.size;output_packet(of, &pkt, ost);if (ost->enc_ctx->codec_type == AVMEDIA_TYPE_VIDEO && vstats_filename) {do_video_stats(ost, pkt_size);}}} }/** Check whether a packet from ist should be written into ost at this time*/ static int check_output_constraints(InputStream *ist, OutputStream *ost) {OutputFile *of = output_files[ost->file_index];int ist_index = input_files[ist->file_index]->ist_index + ist->st->index;if (ost->source_index != ist_index)return 0;if (ost->finished)return 0;if (of->start_time != AV_NOPTS_VALUE && ist->pts < of->start_time)return 0;return 1; }static void do_streamcopy(InputStream *ist, OutputStream *ost, const AVPacket *pkt) {OutputFile *of = output_files[ost->file_index];InputFile *f = input_files [ist->file_index];int64_t start_time = (of->start_time == AV_NOPTS_VALUE) ? 0 : of->start_time;int64_t ost_tb_start_time = av_rescale_q(start_time, AV_TIME_BASE_Q, ost->mux_timebase);AVPicture pict;AVPacket opkt;av_init_packet(&opkt);if ((!ost->frame_number && !(pkt->flags & AV_PKT_FLAG_KEY)) &&!ost->copy_initial_nonkeyframes)return;if (!ost->frame_number && !ost->copy_prior_start) {int64_t comp_start = start_time;if (copy_ts && f->start_time != AV_NOPTS_VALUE)comp_start = FFMAX(start_time, f->start_time + f->ts_offset);if (pkt->pts == AV_NOPTS_VALUE ?ist->pts < comp_start :pkt->pts < av_rescale_q(comp_start, AV_TIME_BASE_Q, ist->st->time_base))return;}if (of->recording_time != INT64_MAX &&ist->pts >= of->recording_time + start_time) {close_output_stream(ost);return;}if (f->recording_time != INT64_MAX) {start_time = f->ctx->start_time;if (f->start_time != AV_NOPTS_VALUE && copy_ts)start_time += f->start_time;if (ist->pts >= f->recording_time + start_time) {close_output_stream(ost);return;}}/* force the input stream PTS */if (ost->enc_ctx->codec_type == AVMEDIA_TYPE_VIDEO)ost->sync_opts++;if (pkt->pts != AV_NOPTS_VALUE)opkt.pts = av_rescale_q(pkt->pts, ist->st->time_base, ost->mux_timebase) - ost_tb_start_time;elseopkt.pts = AV_NOPTS_VALUE;if (pkt->dts == AV_NOPTS_VALUE)opkt.dts = av_rescale_q(ist->dts, AV_TIME_BASE_Q, ost->mux_timebase);elseopkt.dts = av_rescale_q(pkt->dts, ist->st->time_base, ost->mux_timebase);opkt.dts -= ost_tb_start_time;if (ost->st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO && pkt->dts != AV_NOPTS_VALUE) {int duration = av_get_audio_frame_duration(ist->dec_ctx, pkt->size);if(!duration)duration = ist->dec_ctx->frame_size;opkt.dts = opkt.pts = av_rescale_delta(ist->st->time_base, pkt->dts,(AVRational){1, ist->dec_ctx->sample_rate}, duration, &ist->filter_in_rescale_delta_last,ost->mux_timebase) - ost_tb_start_time;}opkt.duration = av_rescale_q(pkt->duration, ist->st->time_base, ost->mux_timebase);opkt.flags = pkt->flags;// FIXME remove the following 2 lines they shall be replaced by the bitstream filtersif ( ost->st->codecpar->codec_id != AV_CODEC_ID_H264&& ost->st->codecpar->codec_id != AV_CODEC_ID_MPEG1VIDEO&& ost->st->codecpar->codec_id != AV_CODEC_ID_MPEG2VIDEO&& ost->st->codecpar->codec_id != AV_CODEC_ID_VC1) {int ret = av_parser_change(ost->parser, ost->parser_avctx,&opkt.data, &opkt.size,pkt->data, pkt->size,pkt->flags & AV_PKT_FLAG_KEY);if (ret < 0) {av_log(NULL, AV_LOG_FATAL, "av_parser_change failed: %s\n",av_err2str(ret));exit_program(1);}if (ret) {opkt.buf = av_buffer_create(opkt.data, opkt.size, av_buffer_default_free, NULL, 0);if (!opkt.buf)exit_program(1);}} else {opkt.data = pkt->data;opkt.size = pkt->size;}av_copy_packet_side_data(&opkt, pkt);#if FF_API_LAVF_FMT_RAWPICTUREif (ost->st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO &&ost->st->codecpar->codec_id == AV_CODEC_ID_RAWVIDEO &&(of->ctx->oformat->flags & AVFMT_RAWPICTURE)) {/* store AVPicture in AVPacket, as expected by the output format */int ret = avpicture_fill(&pict, opkt.data, ost->st->codecpar->format, ost->st->codecpar->width, ost->st->codecpar->height);if (ret < 0) {av_log(NULL, AV_LOG_FATAL, "avpicture_fill failed: %s\n",av_err2str(ret));exit_program(1);}opkt.data = (uint8_t *)&pict;opkt.size = sizeof(AVPicture);opkt.flags |= AV_PKT_FLAG_KEY;} #endifoutput_packet(of, &opkt, ost); }int guess_input_channel_layout(InputStream *ist) {AVCodecContext *dec = ist->dec_ctx;if (!dec->channel_layout) {char layout_name[256];if (dec->channels > ist->guess_layout_max)return 0;dec->channel_layout = av_get_default_channel_layout(dec->channels);if (!dec->channel_layout)return 0;av_get_channel_layout_string(layout_name, sizeof(layout_name),dec->channels, dec->channel_layout);av_log(NULL, AV_LOG_WARNING, "Guessed Channel Layout for Input Stream ""#%d.%d : %s\n", ist->file_index, ist->st->index, layout_name);}return 1; }static void check_decode_result(InputStream *ist, int *got_output, int ret) {if (*got_output || ret<0)decode_error_stat[ret<0] ++;if (ret < 0 && exit_on_error)exit_program(1);if (exit_on_error && *got_output && ist) {if (av_frame_get_decode_error_flags(ist->decoded_frame) || (ist->decoded_frame->flags & AV_FRAME_FLAG_CORRUPT)) {av_log(NULL, AV_LOG_FATAL, "%s: corrupt decoded frame in stream %d\n", input_files[ist->file_index]->ctx->filename, ist->st->index);exit_program(1);}} }// Filters can be configured only if the formats of all inputs are known. static int ifilter_has_all_input_formats(FilterGraph *fg) {int i;for (i = 0; i < fg->nb_inputs; i++) {if (fg->inputs[i]->format < 0 && (fg->inputs[i]->type == AVMEDIA_TYPE_AUDIO ||fg->inputs[i]->type == AVMEDIA_TYPE_VIDEO))return 0;}return 1; }static int ifilter_send_frame(InputFilter *ifilter, AVFrame *frame) {FilterGraph *fg = ifilter->graph;int need_reinit, ret, i;/* determine if the parameters for this input changed */need_reinit = ifilter->format != frame->format;switch (ifilter->ist->st->codecpar->codec_type) {case AVMEDIA_TYPE_AUDIO:need_reinit |= ifilter->sample_rate != frame->sample_rate ||ifilter->channels != frame->channels ||ifilter->channel_layout != frame->channel_layout;break;case AVMEDIA_TYPE_VIDEO:need_reinit |= ifilter->width != frame->width ||ifilter->height != frame->height;break;}if (!ifilter->ist->reinit_filters && fg->graph)need_reinit = 0;if (!!ifilter->hw_frames_ctx != !!frame->hw_frames_ctx ||(ifilter->hw_frames_ctx && ifilter->hw_frames_ctx->data != frame->hw_frames_ctx->data))need_reinit = 1;if (need_reinit) {ret = ifilter_parameters_from_frame(ifilter, frame);if (ret < 0)return ret;}/* (re)init the graph if possible, otherwise buffer the frame and return */if (need_reinit || !fg->graph) {for (i = 0; i < fg->nb_inputs; i++) {if (!ifilter_has_all_input_formats(fg)) {AVFrame *tmp = av_frame_clone(frame);if (!tmp)return AVERROR(ENOMEM);av_frame_unref(frame);if (!av_fifo_space(ifilter->frame_queue)) {ret = av_fifo_realloc2(ifilter->frame_queue, 2 * av_fifo_size(ifilter->frame_queue));if (ret < 0) {av_frame_free(&tmp);return ret;}}av_fifo_generic_write(ifilter->frame_queue, &tmp, sizeof(tmp), NULL);return 0;}}ret = reap_filters(1);if (ret < 0 && ret != AVERROR_EOF) {char errbuf[128];av_strerror(ret, errbuf, sizeof(errbuf));av_log(NULL, AV_LOG_ERROR, "Error while filtering: %s\n", errbuf);return ret;}ret = configure_filtergraph(fg);if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error reinitializing filters!\n");return ret;}}ret = av_buffersrc_add_frame_flags(ifilter->filter, frame, AV_BUFFERSRC_FLAG_PUSH);if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error while filtering\n");return ret;}return 0; }static int ifilter_send_eof(InputFilter *ifilter) {int i, j, ret;ifilter->eof = 1;if (ifilter->filter) {ret = av_buffersrc_add_frame_flags(ifilter->filter, NULL, AV_BUFFERSRC_FLAG_PUSH);if (ret < 0)return ret;} else {// the filtergraph was never configuredFilterGraph *fg = ifilter->graph;for (i = 0; i < fg->nb_inputs; i++)if (!fg->inputs[i]->eof)break;if (i == fg->nb_inputs) {// All the input streams have finished without the filtergraph// ever being configured.// Mark the output streams as finished.for (j = 0; j < fg->nb_outputs; j++)finish_output_stream(fg->outputs[j]->ost);}}return 0; }// This does not quite work like avcodec_decode_audio4/avcodec_decode_video2. // There is the following difference: if you got a frame, you must call // it again with pkt=NULL. pkt==NULL is treated differently from pkt.size==0 // (pkt==NULL means get more output, pkt.size==0 is a flush/drain packet) static int decode(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt) {int ret;*got_frame = 0;if (pkt) {ret = avcodec_send_packet(avctx, pkt);// In particular, we don't expect AVERROR(EAGAIN), because we read all// decoded frames with avcodec_receive_frame() until done.if (ret < 0 && ret != AVERROR_EOF)return ret;}ret = avcodec_receive_frame(avctx, frame);if (ret < 0 && ret != AVERROR(EAGAIN))return ret;if (ret >= 0)*got_frame = 1;return 0; }static int send_frame_to_filters(InputStream *ist, AVFrame *decoded_frame) {int i, ret;AVFrame *f;av_assert1(ist->nb_filters > 0); /* ensure ret is initialized */for (i = 0; i < ist->nb_filters; i++) {if (i < ist->nb_filters - 1) {f = ist->filter_frame;ret = av_frame_ref(f, decoded_frame);if (ret < 0)break;} elsef = decoded_frame;ret = ifilter_send_frame(ist->filters[i], f);if (ret == AVERROR_EOF)ret = 0; /* ignore */if (ret < 0) {av_log(NULL, AV_LOG_ERROR,"Failed to inject frame into filter network: %s\n", av_err2str(ret));break;}}return ret; }static int decode_audio(InputStream *ist, AVPacket *pkt, int *got_output,int *decode_failed) {AVFrame *decoded_frame;AVCodecContext *avctx = ist->dec_ctx;int ret, err = 0;AVRational decoded_frame_tb;if (!ist->decoded_frame && !(ist->decoded_frame = av_frame_alloc()))return AVERROR(ENOMEM);if (!ist->filter_frame && !(ist->filter_frame = av_frame_alloc()))return AVERROR(ENOMEM);decoded_frame = ist->decoded_frame;update_benchmark(NULL);ret = decode(avctx, decoded_frame, got_output, pkt);update_benchmark("decode_audio %d.%d", ist->file_index, ist->st->index);if (ret < 0)*decode_failed = 1;if (ret >= 0 && avctx->sample_rate <= 0) {av_log(avctx, AV_LOG_ERROR, "Sample rate %d invalid\n", avctx->sample_rate);ret = AVERROR_INVALIDDATA;}if (ret != AVERROR_EOF)check_decode_result(ist, got_output, ret);if (!*got_output || ret < 0)return ret;ist->samples_decoded += decoded_frame->nb_samples;ist->frames_decoded++;#if 1/* increment next_dts to use for the case where the input stream does nothave timestamps or there are multiple frames in the packet */ist->next_pts += ((int64_t)AV_TIME_BASE * decoded_frame->nb_samples) /avctx->sample_rate;ist->next_dts += ((int64_t)AV_TIME_BASE * decoded_frame->nb_samples) /avctx->sample_rate; #endifif (decoded_frame->pts != AV_NOPTS_VALUE) {decoded_frame_tb = ist->st->time_base;} else if (pkt && pkt->pts != AV_NOPTS_VALUE) {decoded_frame->pts = pkt->pts;decoded_frame_tb = ist->st->time_base;}else {decoded_frame->pts = ist->dts;decoded_frame_tb = AV_TIME_BASE_Q;}if (decoded_frame->pts != AV_NOPTS_VALUE)decoded_frame->pts = av_rescale_delta(decoded_frame_tb, decoded_frame->pts,(AVRational){1, avctx->sample_rate}, decoded_frame->nb_samples, &ist->filter_in_rescale_delta_last,(AVRational){1, avctx->sample_rate});ist->nb_samples = decoded_frame->nb_samples;err = send_frame_to_filters(ist, decoded_frame);av_frame_unref(ist->filter_frame);av_frame_unref(decoded_frame);return err < 0 ? err : ret; }static int decode_video(InputStream *ist, AVPacket *pkt, int *got_output, int eof,int *decode_failed) {AVFrame *decoded_frame;int i, ret = 0, err = 0;int64_t best_effort_timestamp;int64_t dts = AV_NOPTS_VALUE;AVPacket avpkt;// With fate-indeo3-2, we're getting 0-sized packets before EOF for some// reason. This seems like a semi-critical bug. Don't trigger EOF, and// skip the packet.if (!eof && pkt && pkt->size == 0)return 0;if (!ist->decoded_frame && !(ist->decoded_frame = av_frame_alloc()))return AVERROR(ENOMEM);if (!ist->filter_frame && !(ist->filter_frame = av_frame_alloc()))return AVERROR(ENOMEM);decoded_frame = ist->decoded_frame;if (ist->dts != AV_NOPTS_VALUE)dts = av_rescale_q(ist->dts, AV_TIME_BASE_Q, ist->st->time_base);if (pkt) {avpkt = *pkt;avpkt.dts = dts; // ffmpeg.c probably shouldn't do this}// The old code used to set dts on the drain packet, which does not work// with the new API anymore.if (eof) {void *new = av_realloc_array(ist->dts_buffer, ist->nb_dts_buffer + 1, sizeof(ist->dts_buffer[0]));if (!new)return AVERROR(ENOMEM);ist->dts_buffer = new;ist->dts_buffer[ist->nb_dts_buffer++] = dts;}update_benchmark(NULL);ret = decode(ist->dec_ctx, decoded_frame, got_output, pkt ? &avpkt : NULL);update_benchmark("decode_video %d.%d", ist->file_index, ist->st->index);if (ret < 0)*decode_failed = 1;// The following line may be required in some cases where there is no parser// or the parser does not has_b_frames correctlyif (ist->st->codecpar->video_delay < ist->dec_ctx->has_b_frames) {if (ist->dec_ctx->codec_id == AV_CODEC_ID_H264) {ist->st->codecpar->video_delay = ist->dec_ctx->has_b_frames;} elseav_log(ist->dec_ctx, AV_LOG_WARNING,"video_delay is larger in decoder than demuxer %d > %d.\n""If you want to help, upload a sample ""of this file to ftp://upload.ffmpeg.org/incoming/ ""and contact the ffmpeg-devel mailing list. (ffmpeg-devel@ffmpeg.org)\n",ist->dec_ctx->has_b_frames,ist->st->codecpar->video_delay);}if (ret != AVERROR_EOF)check_decode_result(ist, got_output, ret);if (*got_output && ret >= 0) {if (ist->dec_ctx->width != decoded_frame->width ||ist->dec_ctx->height != decoded_frame->height ||ist->dec_ctx->pix_fmt != decoded_frame->format) {av_log(NULL, AV_LOG_DEBUG, "Frame parameters mismatch context %d,%d,%d != %d,%d,%d\n",decoded_frame->width,decoded_frame->height,decoded_frame->format,ist->dec_ctx->width,ist->dec_ctx->height,ist->dec_ctx->pix_fmt);}}if (!*got_output || ret < 0)return ret;if(ist->top_field_first>=0)decoded_frame->top_field_first = ist->top_field_first;ist->frames_decoded++;if (ist->hwaccel_retrieve_data && decoded_frame->format == ist->hwaccel_pix_fmt) {err = ist->hwaccel_retrieve_data(ist->dec_ctx, decoded_frame);if (err < 0)goto fail;}ist->hwaccel_retrieved_pix_fmt = decoded_frame->format;best_effort_timestamp= av_frame_get_best_effort_timestamp(decoded_frame);if (ist->framerate.num)best_effort_timestamp = ist->cfr_next_pts++;if (eof && best_effort_timestamp == AV_NOPTS_VALUE && ist->nb_dts_buffer > 0) {best_effort_timestamp = ist->dts_buffer[0];for (i = 0; i < ist->nb_dts_buffer - 1; i++)ist->dts_buffer[i] = ist->dts_buffer[i + 1];ist->nb_dts_buffer--;}if(best_effort_timestamp != AV_NOPTS_VALUE) {int64_t ts = av_rescale_q(decoded_frame->pts = best_effort_timestamp, ist->st->time_base, AV_TIME_BASE_Q);if (ts != AV_NOPTS_VALUE)ist->next_pts = ist->pts = ts;}if (debug_ts) {av_log(NULL, AV_LOG_INFO, "decoder -> ist_index:%d type:video ""frame_pts:%s frame_pts_time:%s best_effort_ts:%"PRId64" best_effort_ts_time:%s keyframe:%d frame_type:%d time_base:%d/%d\n",ist->st->index, av_ts2str(decoded_frame->pts),av_ts2timestr(decoded_frame->pts, &ist->st->time_base),best_effort_timestamp,av_ts2timestr(best_effort_timestamp, &ist->st->time_base),decoded_frame->key_frame, decoded_frame->pict_type,ist->st->time_base.num, ist->st->time_base.den);}if (ist->st->sample_aspect_ratio.num)decoded_frame->sample_aspect_ratio = ist->st->sample_aspect_ratio;err = send_frame_to_filters(ist, decoded_frame);fail:av_frame_unref(ist->filter_frame);av_frame_unref(decoded_frame);return err < 0 ? err : ret; }static int transcode_subtitles(InputStream *ist, AVPacket *pkt, int *got_output,int *decode_failed) {AVSubtitle subtitle;int free_sub = 1;int i, ret = avcodec_decode_subtitle2(ist->dec_ctx,&subtitle, got_output, pkt);check_decode_result(NULL, got_output, ret);if (ret < 0 || !*got_output) {*decode_failed = 1;if (!pkt->size)sub2video_flush(ist);return ret;}if (ist->fix_sub_duration) {int end = 1;if (ist->prev_sub.got_output) {end = av_rescale(subtitle.pts - ist->prev_sub.subtitle.pts,1000, AV_TIME_BASE);if (end < ist->prev_sub.subtitle.end_display_time) {av_log(ist->dec_ctx, AV_LOG_DEBUG,"Subtitle duration reduced from %"PRId32" to %d%s\n",ist->prev_sub.subtitle.end_display_time, end,end <= 0 ? ", dropping it" : "");ist->prev_sub.subtitle.end_display_time = end;}}FFSWAP(int, *got_output, ist->prev_sub.got_output);FFSWAP(int, ret, ist->prev_sub.ret);FFSWAP(AVSubtitle, subtitle, ist->prev_sub.subtitle);if (end <= 0)goto out;}if (!*got_output)return ret;if (ist->sub2video.frame) {sub2video_update(ist, &subtitle);} else if (ist->nb_filters) {if (!ist->sub2video.sub_queue)ist->sub2video.sub_queue = av_fifo_alloc(8 * sizeof(AVSubtitle));if (!ist->sub2video.sub_queue)exit_program(1);if (!av_fifo_space(ist->sub2video.sub_queue)) {ret = av_fifo_realloc2(ist->sub2video.sub_queue, 2 * av_fifo_size(ist->sub2video.sub_queue));if (ret < 0)exit_program(1);}av_fifo_generic_write(ist->sub2video.sub_queue, &subtitle, sizeof(subtitle), NULL);free_sub = 0;}if (!subtitle.num_rects)goto out;ist->frames_decoded++;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];if (!check_output_constraints(ist, ost) || !ost->encoding_needed|| ost->enc->type != AVMEDIA_TYPE_SUBTITLE)continue;do_subtitle_out(output_files[ost->file_index], ost, &subtitle);}out:if (free_sub)avsubtitle_free(&subtitle);return ret; }static int send_filter_eof(InputStream *ist) {int i, ret;for (i = 0; i < ist->nb_filters; i++) {ret = ifilter_send_eof(ist->filters[i]);if (ret < 0)return ret;}return 0; }/* pkt = NULL means EOF (needed to flush decoder buffers) */ static int process_input_packet(InputStream *ist, const AVPacket *pkt, int no_eof) {int ret = 0, i;int repeating = 0;int eof_reached = 0;AVPacket avpkt;if (!ist->saw_first_ts) {ist->dts = ist->st->avg_frame_rate.num ? - ist->dec_ctx->has_b_frames * AV_TIME_BASE / av_q2d(ist->st->avg_frame_rate) : 0;ist->pts = 0;if (pkt && pkt->pts != AV_NOPTS_VALUE && !ist->decoding_needed) {ist->dts += av_rescale_q(pkt->pts, ist->st->time_base, AV_TIME_BASE_Q);ist->pts = ist->dts; //unused but better to set it to a value thats not totally wrong}ist->saw_first_ts = 1;}if (ist->next_dts == AV_NOPTS_VALUE)ist->next_dts = ist->dts;if (ist->next_pts == AV_NOPTS_VALUE)ist->next_pts = ist->pts;if (!pkt) {/* EOF handling */av_init_packet(&avpkt);avpkt.data = NULL;avpkt.size = 0;} else {avpkt = *pkt;}if (pkt && pkt->dts != AV_NOPTS_VALUE) {ist->next_dts = ist->dts = av_rescale_q(pkt->dts, ist->st->time_base, AV_TIME_BASE_Q);if (ist->dec_ctx->codec_type != AVMEDIA_TYPE_VIDEO || !ist->decoding_needed)ist->next_pts = ist->pts = ist->dts;}// while we have more to decode or while the decoder did output something on EOFwhile (ist->decoding_needed) {int duration = 0;int got_output = 0;int decode_failed = 0;ist->pts = ist->next_pts;ist->dts = ist->next_dts;switch (ist->dec_ctx->codec_type) {case AVMEDIA_TYPE_AUDIO:ret = decode_audio (ist, repeating ? NULL : &avpkt, &got_output,&decode_failed);break;case AVMEDIA_TYPE_VIDEO:ret = decode_video (ist, repeating ? NULL : &avpkt, &got_output, !pkt,&decode_failed);if (!repeating || !pkt || got_output) {if (pkt && pkt->duration) {duration = av_rescale_q(pkt->duration, ist->st->time_base, AV_TIME_BASE_Q);} else if(ist->dec_ctx->framerate.num != 0 && ist->dec_ctx->framerate.den != 0) {int ticks= av_stream_get_parser(ist->st) ? av_stream_get_parser(ist->st)->repeat_pict+1 : ist->dec_ctx->ticks_per_frame;duration = ((int64_t)AV_TIME_BASE *ist->dec_ctx->framerate.den * ticks) /ist->dec_ctx->framerate.num / ist->dec_ctx->ticks_per_frame;}if(ist->dts != AV_NOPTS_VALUE && duration) {ist->next_dts += duration;}elseist->next_dts = AV_NOPTS_VALUE;}if (got_output)ist->next_pts += duration; //FIXME the duration is not correct in some casesbreak;case AVMEDIA_TYPE_SUBTITLE:if (repeating)break;ret = transcode_subtitles(ist, &avpkt, &got_output, &decode_failed);if (!pkt && ret >= 0)ret = AVERROR_EOF;break;default:return -1;}if (ret == AVERROR_EOF) {eof_reached = 1;break;}if (ret < 0) {if (decode_failed) {av_log(NULL, AV_LOG_ERROR, "Error while decoding stream #%d:%d: %s\n",ist->file_index, ist->st->index, av_err2str(ret));} else {av_log(NULL, AV_LOG_FATAL, "Error while processing the decoded ""data for stream #%d:%d\n", ist->file_index, ist->st->index);}if (!decode_failed || exit_on_error)exit_program(1);break;}if (got_output)ist->got_output = 1;if (!got_output)break;// During draining, we might get multiple output frames in this loop.// ffmpeg.c does not drain the filter chain on configuration changes,// which means if we send multiple frames at once to the filters, and// one of those frames changes configuration, the buffered frames will// be lost. This can upset certain FATE tests.// Decode only 1 frame per call on EOF to appease these FATE tests.// The ideal solution would be to rewrite decoding to use the new// decoding API in a better way.if (!pkt)break;repeating = 1;}/* after flushing, send an EOF on all the filter inputs attached to the stream *//* except when looping we need to flush but not to send an EOF */if (!pkt && ist->decoding_needed && eof_reached && !no_eof) {int ret = send_filter_eof(ist);if (ret < 0) {av_log(NULL, AV_LOG_FATAL, "Error marking filters as finished\n");exit_program(1);}}/* handle stream copy */if (!ist->decoding_needed) {ist->dts = ist->next_dts;switch (ist->dec_ctx->codec_type) {case AVMEDIA_TYPE_AUDIO:if (ist->dec_ctx->sample_rate) {ist->next_dts += ((int64_t)AV_TIME_BASE * ist->dec_ctx->frame_size) /ist->dec_ctx->sample_rate;} else {ist->next_dts += av_rescale_q(pkt->duration, ist->st->time_base, AV_TIME_BASE_Q);}break;case AVMEDIA_TYPE_VIDEO:if (ist->framerate.num) {// TODO: Remove work-around for c99-to-c89 issue 7AVRational time_base_q = AV_TIME_BASE_Q;int64_t next_dts = av_rescale_q(ist->next_dts, time_base_q, av_inv_q(ist->framerate));ist->next_dts = av_rescale_q(next_dts + 1, av_inv_q(ist->framerate), time_base_q);} else if (pkt->duration) {ist->next_dts += av_rescale_q(pkt->duration, ist->st->time_base, AV_TIME_BASE_Q);} else if(ist->dec_ctx->framerate.num != 0) {int ticks= av_stream_get_parser(ist->st) ? av_stream_get_parser(ist->st)->repeat_pict + 1 : ist->dec_ctx->ticks_per_frame;ist->next_dts += ((int64_t)AV_TIME_BASE *ist->dec_ctx->framerate.den * ticks) /ist->dec_ctx->framerate.num / ist->dec_ctx->ticks_per_frame;}break;}ist->pts = ist->dts;ist->next_pts = ist->next_dts;}for (i = 0; pkt && i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];if (!check_output_constraints(ist, ost) || ost->encoding_needed)continue;do_streamcopy(ist, ost, pkt);}return !eof_reached; }static void print_sdp(void) {char sdp[16384];int i;int j;AVIOContext *sdp_pb;AVFormatContext **avc;for (i = 0; i < nb_output_files; i++) {if (!output_files[i]->header_written)return;}avc = av_malloc_array(nb_output_files, sizeof(*avc));if (!avc)exit_program(1);for (i = 0, j = 0; i < nb_output_files; i++) {if (!strcmp(output_files[i]->ctx->oformat->name, "rtp")) {avc[j] = output_files[i]->ctx;j++;}}if (!j)goto fail;av_sdp_create(avc, j, sdp, sizeof(sdp));if (!sdp_filename) {printf("SDP:\n%s\n", sdp);fflush(stdout);} else {if (avio_open2(&sdp_pb, sdp_filename, AVIO_FLAG_WRITE, &int_cb, NULL) < 0) {av_log(NULL, AV_LOG_ERROR, "Failed to open sdp file '%s'\n", sdp_filename);} else {avio_printf(sdp_pb, "SDP:\n%s", sdp);avio_closep(&sdp_pb);av_freep(&sdp_filename);}}fail:av_freep(&avc); }static const HWAccel *get_hwaccel(enum AVPixelFormat pix_fmt) {int i;for (i = 0; hwaccels[i].name; i++)if (hwaccels[i].pix_fmt == pix_fmt)return &hwaccels[i];return NULL; }static enum AVPixelFormat get_format(AVCodecContext *s, const enum AVPixelFormat *pix_fmts) {InputStream *ist = s->opaque;const enum AVPixelFormat *p;int ret;for (p = pix_fmts; *p != -1; p++) {const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(*p);const HWAccel *hwaccel;if (!(desc->flags & AV_PIX_FMT_FLAG_HWACCEL))break;hwaccel = get_hwaccel(*p);if (!hwaccel ||(ist->active_hwaccel_id && ist->active_hwaccel_id != hwaccel->id) ||(ist->hwaccel_id != HWACCEL_AUTO && ist->hwaccel_id != hwaccel->id))continue;ret = hwaccel->init(s);if (ret < 0) {if (ist->hwaccel_id == hwaccel->id) {av_log(NULL, AV_LOG_FATAL,"%s hwaccel requested for input stream #%d:%d, ""but cannot be initialized.\n", hwaccel->name,ist->file_index, ist->st->index);return AV_PIX_FMT_NONE;}continue;}if (ist->hw_frames_ctx) {s->hw_frames_ctx = av_buffer_ref(ist->hw_frames_ctx);if (!s->hw_frames_ctx)return AV_PIX_FMT_NONE;}ist->active_hwaccel_id = hwaccel->id;ist->hwaccel_pix_fmt = *p;break;}return *p; }static int get_buffer(AVCodecContext *s, AVFrame *frame, int flags) {InputStream *ist = s->opaque;if (ist->hwaccel_get_buffer && frame->format == ist->hwaccel_pix_fmt)return ist->hwaccel_get_buffer(s, frame, flags);return avcodec_default_get_buffer2(s, frame, flags); }static int init_input_stream(int ist_index, char *error, int error_len) {int ret;InputStream *ist = input_streams[ist_index];if (ist->decoding_needed) {AVCodec *codec = ist->dec;if (!codec) {snprintf(error, error_len, "Decoder (codec %s) not found for input stream #%d:%d",avcodec_get_name(ist->dec_ctx->codec_id), ist->file_index, ist->st->index);return AVERROR(EINVAL);}ist->dec_ctx->opaque = ist;ist->dec_ctx->get_format = get_format;ist->dec_ctx->get_buffer2 = get_buffer;ist->dec_ctx->thread_safe_callbacks = 1;av_opt_set_int(ist->dec_ctx, "refcounted_frames", 1, 0);if (ist->dec_ctx->codec_id == AV_CODEC_ID_DVB_SUBTITLE &&(ist->decoding_needed & DECODING_FOR_OST)) {av_dict_set(&ist->decoder_opts, "compute_edt", "1", AV_DICT_DONT_OVERWRITE);if (ist->decoding_needed & DECODING_FOR_FILTER)av_log(NULL, AV_LOG_WARNING, "Warning using DVB subtitles for filtering and output at the same time is not fully supported, also see -compute_edt [0|1]\n");}av_dict_set(&ist->decoder_opts, "sub_text_format", "ass", AV_DICT_DONT_OVERWRITE);/* Useful for subtitles retiming by lavf (FIXME), skipping samples in* audio, and video decoders such as cuvid or mediacodec */av_codec_set_pkt_timebase(ist->dec_ctx, ist->st->time_base);if (!av_dict_get(ist->decoder_opts, "threads", NULL, 0))av_dict_set(&ist->decoder_opts, "threads", "auto", 0);if ((ret = avcodec_open2(ist->dec_ctx, codec, &ist->decoder_opts)) < 0) {if (ret == AVERROR_EXPERIMENTAL)abort_codec_experimental(codec, 0);snprintf(error, error_len,"Error while opening decoder for input stream ""#%d:%d : %s",ist->file_index, ist->st->index, av_err2str(ret));return ret;}assert_avoptions(ist->decoder_opts);}ist->next_pts = AV_NOPTS_VALUE;ist->next_dts = AV_NOPTS_VALUE;return 0; }static InputStream *get_input_stream(OutputStream *ost) {if (ost->source_index >= 0)return input_streams[ost->source_index];return NULL; }static int compare_int64(const void *a, const void *b) {return FFDIFFSIGN(*(const int64_t *)a, *(const int64_t *)b); }/* open the muxer when all the streams are initialized */ static int check_init_output_file(OutputFile *of, int file_index) {int ret, i;for (i = 0; i < of->ctx->nb_streams; i++) {OutputStream *ost = output_streams[of->ost_index + i];if (!ost->initialized)return 0;}of->ctx->interrupt_callback = int_cb;ret = avformat_write_header(of->ctx, &of->opts);if (ret < 0) {av_log(NULL, AV_LOG_ERROR,"Could not write header for output file #%d ""(incorrect codec parameters ?): %s\n",file_index, av_err2str(ret));return ret;}//assert_avoptions(of->opts);of->header_written = 1;av_dump_format(of->ctx, file_index, of->ctx->filename, 1);if (sdp_filename || want_sdp)print_sdp();/* flush the muxing queues */for (i = 0; i < of->ctx->nb_streams; i++) {OutputStream *ost = output_streams[of->ost_index + i];/* try to improve muxing time_base (only possible if nothing has been written yet) */if (!av_fifo_size(ost->muxing_queue))ost->mux_timebase = ost->st->time_base;while (av_fifo_size(ost->muxing_queue)) {AVPacket pkt;av_fifo_generic_read(ost->muxing_queue, &pkt, sizeof(pkt), NULL);write_packet(of, &pkt, ost, 1);}}return 0; }static int init_output_bsfs(OutputStream *ost) {AVBSFContext *ctx;int i, ret;if (!ost->nb_bitstream_filters)return 0;for (i = 0; i < ost->nb_bitstream_filters; i++) {ctx = ost->bsf_ctx[i];ret = avcodec_parameters_copy(ctx->par_in,i ? ost->bsf_ctx[i - 1]->par_out : ost->st->codecpar);if (ret < 0)return ret;ctx->time_base_in = i ? ost->bsf_ctx[i - 1]->time_base_out : ost->st->time_base;ret = av_bsf_init(ctx);if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error initializing bitstream filter: %s\n",ost->bsf_ctx[i]->filter->name);return ret;}}ctx = ost->bsf_ctx[ost->nb_bitstream_filters - 1];ret = avcodec_parameters_copy(ost->st->codecpar, ctx->par_out);if (ret < 0)return ret;ost->st->time_base = ctx->time_base_out;return 0; }static int init_output_stream_streamcopy(OutputStream *ost) {OutputFile *of = output_files[ost->file_index];InputStream *ist = get_input_stream(ost);AVCodecParameters *par_dst = ost->st->codecpar;AVCodecParameters *par_src = ost->ref_par;AVRational sar;int i, ret;uint32_t codec_tag = par_dst->codec_tag;av_assert0(ist && !ost->filter);ret = avcodec_parameters_to_context(ost->enc_ctx, ist->st->codecpar);if (ret >= 0)ret = av_opt_set_dict(ost->enc_ctx, &ost->encoder_opts);if (ret < 0) {av_log(NULL, AV_LOG_FATAL,"Error setting up codec context options.\n");return ret;}avcodec_parameters_from_context(par_src, ost->enc_ctx);if (!codec_tag) {unsigned int codec_tag_tmp;if (!of->ctx->oformat->codec_tag ||av_codec_get_id (of->ctx->oformat->codec_tag, par_src->codec_tag) == par_src->codec_id ||!av_codec_get_tag2(of->ctx->oformat->codec_tag, par_src->codec_id, &codec_tag_tmp))codec_tag = par_src->codec_tag;}ret = avcodec_parameters_copy(par_dst, par_src);if (ret < 0)return ret;par_dst->codec_tag = codec_tag;if (!ost->frame_rate.num)ost->frame_rate = ist->framerate;ost->st->avg_frame_rate = ost->frame_rate;ret = avformat_transfer_internal_stream_timing_info(of->ctx->oformat, ost->st, ist->st, copy_tb);if (ret < 0)return ret;// copy timebase while removing common factorsif (ost->st->time_base.num <= 0 || ost->st->time_base.den <= 0)ost->st->time_base = av_add_q(av_stream_get_codec_timebase(ost->st), (AVRational){0, 1});// copy estimated duration as a hint to the muxerif (ost->st->duration <= 0 && ist->st->duration > 0)ost->st->duration = av_rescale_q(ist->st->duration, ist->st->time_base, ost->st->time_base);// copy dispositionost->st->disposition = ist->st->disposition;if (ist->st->nb_side_data) {ost->st->side_data = av_realloc_array(NULL, ist->st->nb_side_data,sizeof(*ist->st->side_data));if (!ost->st->side_data)return AVERROR(ENOMEM);ost->st->nb_side_data = 0;for (i = 0; i < ist->st->nb_side_data; i++) {const AVPacketSideData *sd_src = &ist->st->side_data[i];AVPacketSideData *sd_dst = &ost->st->side_data[ost->st->nb_side_data];sd_dst->data = av_malloc(sd_src->size);if (!sd_dst->data)return AVERROR(ENOMEM);memcpy(sd_dst->data, sd_src->data, sd_src->size);sd_dst->size = sd_src->size;sd_dst->type = sd_src->type;ost->st->nb_side_data++;}}if (ost->rotate_overridden) {uint8_t *sd = av_stream_new_side_data(ost->st, AV_PKT_DATA_DISPLAYMATRIX,sizeof(int32_t) * 9);if (sd)av_display_rotation_set((int32_t *)sd, -ost->rotate_override_value);}ost->parser = av_parser_init(par_dst->codec_id);ost->parser_avctx = avcodec_alloc_context3(NULL);if (!ost->parser_avctx)return AVERROR(ENOMEM);switch (par_dst->codec_type) {case AVMEDIA_TYPE_AUDIO:if (audio_volume != 256) {av_log(NULL, AV_LOG_FATAL, "-acodec copy and -vol are incompatible (frames are not decoded)\n");exit_program(1);}if((par_dst->block_align == 1 || par_dst->block_align == 1152 || par_dst->block_align == 576) && par_dst->codec_id == AV_CODEC_ID_MP3)par_dst->block_align= 0;if(par_dst->codec_id == AV_CODEC_ID_AC3)par_dst->block_align= 0;break;case AVMEDIA_TYPE_VIDEO:if (ost->frame_aspect_ratio.num) { // overridden by the -aspect cli optionsar =av_mul_q(ost->frame_aspect_ratio,(AVRational){ par_dst->height, par_dst->width });av_log(NULL, AV_LOG_WARNING, "Overriding aspect ratio ""with stream copy may produce invalid files\n");}else if (ist->st->sample_aspect_ratio.num)sar = ist->st->sample_aspect_ratio;elsesar = par_src->sample_aspect_ratio;ost->st->sample_aspect_ratio = par_dst->sample_aspect_ratio = sar;ost->st->avg_frame_rate = ist->st->avg_frame_rate;ost->st->r_frame_rate = ist->st->r_frame_rate;break;}ost->mux_timebase = ist->st->time_base;return 0; }static void set_encoder_id(OutputFile *of, OutputStream *ost) {AVDictionaryEntry *e;uint8_t *encoder_string;int encoder_string_len;int format_flags = 0;int codec_flags = 0;if (av_dict_get(ost->st->metadata, "encoder", NULL, 0))return;e = av_dict_get(of->opts, "fflags", NULL, 0);if (e) {const AVOption *o = av_opt_find(of->ctx, "fflags", NULL, 0, 0);if (!o)return;av_opt_eval_flags(of->ctx, o, e->value, &format_flags);}e = av_dict_get(ost->encoder_opts, "flags", NULL, 0);if (e) {const AVOption *o = av_opt_find(ost->enc_ctx, "flags", NULL, 0, 0);if (!o)return;av_opt_eval_flags(ost->enc_ctx, o, e->value, &codec_flags);}encoder_string_len = sizeof(LIBAVCODEC_IDENT) + strlen(ost->enc->name) + 2;encoder_string = av_mallocz(encoder_string_len);if (!encoder_string)exit_program(1);if (!(format_flags & AVFMT_FLAG_BITEXACT) && !(codec_flags & AV_CODEC_FLAG_BITEXACT))av_strlcpy(encoder_string, LIBAVCODEC_IDENT " ", encoder_string_len);elseav_strlcpy(encoder_string, "Lavc ", encoder_string_len);av_strlcat(encoder_string, ost->enc->name, encoder_string_len);av_dict_set(&ost->st->metadata, "encoder", encoder_string,AV_DICT_DONT_STRDUP_VAL | AV_DICT_DONT_OVERWRITE); }static void parse_forced_key_frames(char *kf, OutputStream *ost,AVCodecContext *avctx) {char *p;int n = 1, i, size, index = 0;int64_t t, *pts;for (p = kf; *p; p++)if (*p == ',')n++;size = n;pts = av_malloc_array(size, sizeof(*pts));if (!pts) {av_log(NULL, AV_LOG_FATAL, "Could not allocate forced key frames array.\n");exit_program(1);}p = kf;for (i = 0; i < n; i++) {char *next = strchr(p, ',');if (next)*next++ = 0;if (!memcmp(p, "chapters", 8)) {AVFormatContext *avf = output_files[ost->file_index]->ctx;int j;if (avf->nb_chapters > INT_MAX - size ||!(pts = av_realloc_f(pts, size += avf->nb_chapters - 1,sizeof(*pts)))) {av_log(NULL, AV_LOG_FATAL,"Could not allocate forced key frames array.\n");exit_program(1);}t = p[8] ? parse_time_or_die("force_key_frames", p + 8, 1) : 0;t = av_rescale_q(t, AV_TIME_BASE_Q, avctx->time_base);for (j = 0; j < avf->nb_chapters; j++) {AVChapter *c = avf->chapters[j];av_assert1(index < size);pts[index++] = av_rescale_q(c->start, c->time_base,avctx->time_base) + t;}} else {t = parse_time_or_die("force_key_frames", p, 1);av_assert1(index < size);pts[index++] = av_rescale_q(t, AV_TIME_BASE_Q, avctx->time_base);}p = next;}av_assert0(index == size);qsort(pts, size, sizeof(*pts), compare_int64);ost->forced_kf_count = size;ost->forced_kf_pts = pts; }static int init_output_stream_encode(OutputStream *ost) {InputStream *ist = get_input_stream(ost);AVCodecContext *enc_ctx = ost->enc_ctx;AVCodecContext *dec_ctx = NULL;AVFormatContext *oc = output_files[ost->file_index]->ctx;int j, ret;set_encoder_id(output_files[ost->file_index], ost);// Muxers use AV_PKT_DATA_DISPLAYMATRIX to signal rotation. On the other// hand, the legacy API makes demuxers set "rotate" metadata entries,// which have to be filtered out to prevent leaking them to output files.av_dict_set(&ost->st->metadata, "rotate", NULL, 0);if (ist) {ost->st->disposition = ist->st->disposition;dec_ctx = ist->dec_ctx;enc_ctx->chroma_sample_location = dec_ctx->chroma_sample_location;} else {for (j = 0; j < oc->nb_streams; j++) {AVStream *st = oc->streams[j];if (st != ost->st && st->codecpar->codec_type == ost->st->codecpar->codec_type)break;}if (j == oc->nb_streams)if (ost->st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO ||ost->st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)ost->st->disposition = AV_DISPOSITION_DEFAULT;}if (enc_ctx->codec_type == AVMEDIA_TYPE_VIDEO) {if (!ost->frame_rate.num)ost->frame_rate = av_buffersink_get_frame_rate(ost->filter->filter);if (ist && !ost->frame_rate.num)ost->frame_rate = ist->framerate;if (ist && !ost->frame_rate.num)ost->frame_rate = ist->st->r_frame_rate;if (ist && !ost->frame_rate.num) {ost->frame_rate = (AVRational){25, 1};av_log(NULL, AV_LOG_WARNING,"No information ""about the input framerate is available. Falling ""back to a default value of 25fps for output stream #%d:%d. Use the -r option ""if you want a different framerate.\n",ost->file_index, ost->index);} // ost->frame_rate = ist->st->avg_frame_rate.num ? ist->st->avg_frame_rate : (AVRational){25, 1};if (ost->enc->supported_framerates && !ost->force_fps) {int idx = av_find_nearest_q_idx(ost->frame_rate, ost->enc->supported_framerates);ost->frame_rate = ost->enc->supported_framerates[idx];}// reduce frame rate for mpeg4 to be within the spec limitsif (enc_ctx->codec_id == AV_CODEC_ID_MPEG4) {av_reduce(&ost->frame_rate.num, &ost->frame_rate.den,ost->frame_rate.num, ost->frame_rate.den, 65535);}}switch (enc_ctx->codec_type) {case AVMEDIA_TYPE_AUDIO:enc_ctx->sample_fmt = av_buffersink_get_format(ost->filter->filter);if (dec_ctx)enc_ctx->bits_per_raw_sample = FFMIN(dec_ctx->bits_per_raw_sample,av_get_bytes_per_sample(enc_ctx->sample_fmt) << 3);enc_ctx->sample_rate = av_buffersink_get_sample_rate(ost->filter->filter);enc_ctx->channel_layout = av_buffersink_get_channel_layout(ost->filter->filter);enc_ctx->channels = av_buffersink_get_channels(ost->filter->filter);enc_ctx->time_base = (AVRational){ 1, enc_ctx->sample_rate };break;case AVMEDIA_TYPE_VIDEO:enc_ctx->time_base = av_inv_q(ost->frame_rate);if (!(enc_ctx->time_base.num && enc_ctx->time_base.den))enc_ctx->time_base = av_buffersink_get_time_base(ost->filter->filter);if ( av_q2d(enc_ctx->time_base) < 0.001 && video_sync_method != VSYNC_PASSTHROUGH&& (video_sync_method == VSYNC_CFR || video_sync_method == VSYNC_VSCFR || (video_sync_method == VSYNC_AUTO && !(oc->oformat->flags & AVFMT_VARIABLE_FPS)))){av_log(oc, AV_LOG_WARNING, "Frame rate very high for a muxer not efficiently supporting it.\n""Please consider specifying a lower framerate, a different muxer or -vsync 2\n");}for (j = 0; j < ost->forced_kf_count; j++)ost->forced_kf_pts[j] = av_rescale_q(ost->forced_kf_pts[j],AV_TIME_BASE_Q,enc_ctx->time_base);enc_ctx->width = av_buffersink_get_w(ost->filter->filter);enc_ctx->height = av_buffersink_get_h(ost->filter->filter);enc_ctx->sample_aspect_ratio = ost->st->sample_aspect_ratio =ost->frame_aspect_ratio.num ? // overridden by the -aspect cli optionav_mul_q(ost->frame_aspect_ratio, (AVRational){ enc_ctx->height, enc_ctx->width }) :av_buffersink_get_sample_aspect_ratio(ost->filter->filter);if (!strncmp(ost->enc->name, "libx264", 7) &&enc_ctx->pix_fmt == AV_PIX_FMT_NONE &&av_buffersink_get_format(ost->filter->filter) != AV_PIX_FMT_YUV420P)av_log(NULL, AV_LOG_WARNING,"No pixel format specified, %s for H.264 encoding chosen.\n""Use -pix_fmt yuv420p for compatibility with outdated media players.\n",av_get_pix_fmt_name(av_buffersink_get_format(ost->filter->filter)));if (!strncmp(ost->enc->name, "mpeg2video", 10) &&enc_ctx->pix_fmt == AV_PIX_FMT_NONE &&av_buffersink_get_format(ost->filter->filter) != AV_PIX_FMT_YUV420P)av_log(NULL, AV_LOG_WARNING,"No pixel format specified, %s for MPEG-2 encoding chosen.\n""Use -pix_fmt yuv420p for compatibility with outdated media players.\n",av_get_pix_fmt_name(av_buffersink_get_format(ost->filter->filter)));enc_ctx->pix_fmt = av_buffersink_get_format(ost->filter->filter);if (dec_ctx)enc_ctx->bits_per_raw_sample = FFMIN(dec_ctx->bits_per_raw_sample,av_pix_fmt_desc_get(enc_ctx->pix_fmt)->comp[0].depth);enc_ctx->framerate = ost->frame_rate;ost->st->avg_frame_rate = ost->frame_rate;if (!dec_ctx ||enc_ctx->width != dec_ctx->width ||enc_ctx->height != dec_ctx->height ||enc_ctx->pix_fmt != dec_ctx->pix_fmt) {enc_ctx->bits_per_raw_sample = frame_bits_per_raw_sample;}if (ost->forced_keyframes) {if (!strncmp(ost->forced_keyframes, "expr:", 5)) {ret = av_expr_parse(&ost->forced_keyframes_pexpr, ost->forced_keyframes+5,forced_keyframes_const_names, NULL, NULL, NULL, NULL, 0, NULL);if (ret < 0) {av_log(NULL, AV_LOG_ERROR,"Invalid force_key_frames expression '%s'\n", ost->forced_keyframes+5);return ret;}ost->forced_keyframes_expr_const_values[FKF_N] = 0;ost->forced_keyframes_expr_const_values[FKF_N_FORCED] = 0;ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_N] = NAN;ost->forced_keyframes_expr_const_values[FKF_PREV_FORCED_T] = NAN;// Don't parse the 'forced_keyframes' in case of 'keep-source-keyframes',// parse it only for static kf timings} else if(strncmp(ost->forced_keyframes, "source", 6)) {parse_forced_key_frames(ost->forced_keyframes, ost, ost->enc_ctx);}}break;case AVMEDIA_TYPE_SUBTITLE:enc_ctx->time_base = AV_TIME_BASE_Q;if (!enc_ctx->width) {enc_ctx->width = input_streams[ost->source_index]->st->codecpar->width;enc_ctx->height = input_streams[ost->source_index]->st->codecpar->height;}break;case AVMEDIA_TYPE_DATA:break;default:abort();break;}ost->mux_timebase = enc_ctx->time_base;return 0; }static int init_output_stream(OutputStream *ost, char *error, int error_len) {int ret = 0;if (ost->encoding_needed) {AVCodec *codec = ost->enc;AVCodecContext *dec = NULL;InputStream *ist;ret = init_output_stream_encode(ost);if (ret < 0)return ret;if ((ist = get_input_stream(ost)))dec = ist->dec_ctx;if (dec && dec->subtitle_header) {/* ASS code assumes this buffer is null terminated so add extra byte. */ost->enc_ctx->subtitle_header = av_mallocz(dec->subtitle_header_size + 1);if (!ost->enc_ctx->subtitle_header)return AVERROR(ENOMEM);memcpy(ost->enc_ctx->subtitle_header, dec->subtitle_header, dec->subtitle_header_size);ost->enc_ctx->subtitle_header_size = dec->subtitle_header_size;}if (!av_dict_get(ost->encoder_opts, "threads", NULL, 0))av_dict_set(&ost->encoder_opts, "threads", "auto", 0);if (ost->enc->type == AVMEDIA_TYPE_AUDIO &&!codec->defaults &&!av_dict_get(ost->encoder_opts, "b", NULL, 0) &&!av_dict_get(ost->encoder_opts, "ab", NULL, 0))av_dict_set(&ost->encoder_opts, "b", "128000", 0);if (ost->filter && av_buffersink_get_hw_frames_ctx(ost->filter->filter) &&((AVHWFramesContext*)av_buffersink_get_hw_frames_ctx(ost->filter->filter)->data)->format ==av_buffersink_get_format(ost->filter->filter)) {ost->enc_ctx->hw_frames_ctx = av_buffer_ref(av_buffersink_get_hw_frames_ctx(ost->filter->filter));if (!ost->enc_ctx->hw_frames_ctx)return AVERROR(ENOMEM);}if ((ret = avcodec_open2(ost->enc_ctx, codec, &ost->encoder_opts)) < 0) {if (ret == AVERROR_EXPERIMENTAL)abort_codec_experimental(codec, 1);snprintf(error, error_len,"Error while opening encoder for output stream #%d:%d - ""maybe incorrect parameters such as bit_rate, rate, width or height",ost->file_index, ost->index);return ret;}if (ost->enc->type == AVMEDIA_TYPE_AUDIO &&!(ost->enc->capabilities & AV_CODEC_CAP_VARIABLE_FRAME_SIZE))av_buffersink_set_frame_size(ost->filter->filter,ost->enc_ctx->frame_size);assert_avoptions(ost->encoder_opts);if (ost->enc_ctx->bit_rate && ost->enc_ctx->bit_rate < 1000)av_log(NULL, AV_LOG_WARNING, "The bitrate parameter is set too low."" It takes bits/s as argument, not kbits/s\n");ret = avcodec_parameters_from_context(ost->st->codecpar, ost->enc_ctx);if (ret < 0) {av_log(NULL, AV_LOG_FATAL,"Error initializing the output stream codec context.\n");exit_program(1);}/** FIXME: ost->st->codec should't be needed here anymore.*/ret = avcodec_copy_context(ost->st->codec, ost->enc_ctx);if (ret < 0)return ret;if (ost->enc_ctx->nb_coded_side_data) {int i;ost->st->side_data = av_realloc_array(NULL, ost->enc_ctx->nb_coded_side_data,sizeof(*ost->st->side_data));if (!ost->st->side_data)return AVERROR(ENOMEM);for (i = 0; i < ost->enc_ctx->nb_coded_side_data; i++) {const AVPacketSideData *sd_src = &ost->enc_ctx->coded_side_data[i];AVPacketSideData *sd_dst = &ost->st->side_data[i];sd_dst->data = av_malloc(sd_src->size);if (!sd_dst->data)return AVERROR(ENOMEM);memcpy(sd_dst->data, sd_src->data, sd_src->size);sd_dst->size = sd_src->size;sd_dst->type = sd_src->type;ost->st->nb_side_data++;}}/** Add global input side data. For now this is naive, and copies it* from the input stream's global side data. All side data should* really be funneled over AVFrame and libavfilter, then added back to* packet side data, and then potentially using the first packet for* global side data.*/if (ist) {int i;for (i = 0; i < ist->st->nb_side_data; i++) {AVPacketSideData *sd = &ist->st->side_data[i];uint8_t *dst = av_stream_new_side_data(ost->st, sd->type, sd->size);if (!dst)return AVERROR(ENOMEM);memcpy(dst, sd->data, sd->size);if (ist->autorotate && sd->type == AV_PKT_DATA_DISPLAYMATRIX)av_display_rotation_set((uint32_t *)dst, 0);}}// copy timebase while removing common factorsif (ost->st->time_base.num <= 0 || ost->st->time_base.den <= 0)ost->st->time_base = av_add_q(ost->enc_ctx->time_base, (AVRational){0, 1});// copy estimated duration as a hint to the muxerif (ost->st->duration <= 0 && ist && ist->st->duration > 0)ost->st->duration = av_rescale_q(ist->st->duration, ist->st->time_base, ost->st->time_base);ost->st->codec->codec= ost->enc_ctx->codec;} else if (ost->stream_copy) {ret = init_output_stream_streamcopy(ost);if (ret < 0)return ret;/** FIXME: will the codec context used by the parser during streamcopy* This should go away with the new parser API.*/ret = avcodec_parameters_to_context(ost->parser_avctx, ost->st->codecpar);if (ret < 0)return ret;}// parse user provided disposition, and update stream valuesif (ost->disposition) {static const AVOption opts[] = {{ "disposition" , NULL, 0, AV_OPT_TYPE_FLAGS, { .i64 = 0 }, INT64_MIN, INT64_MAX, .unit = "flags" },{ "default" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_DEFAULT }, .unit = "flags" },{ "dub" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_DUB }, .unit = "flags" },{ "original" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_ORIGINAL }, .unit = "flags" },{ "comment" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_COMMENT }, .unit = "flags" },{ "lyrics" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_LYRICS }, .unit = "flags" },{ "karaoke" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_KARAOKE }, .unit = "flags" },{ "forced" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_FORCED }, .unit = "flags" },{ "hearing_impaired" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_HEARING_IMPAIRED }, .unit = "flags" },{ "visual_impaired" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_VISUAL_IMPAIRED }, .unit = "flags" },{ "clean_effects" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_CLEAN_EFFECTS }, .unit = "flags" },{ "captions" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_CAPTIONS }, .unit = "flags" },{ "descriptions" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_DESCRIPTIONS }, .unit = "flags" },{ "metadata" , NULL, 0, AV_OPT_TYPE_CONST, { .i64 = AV_DISPOSITION_METADATA }, .unit = "flags" },{ NULL },};static const AVClass class = {.class_name = "",.item_name = av_default_item_name,.option = opts,.version = LIBAVUTIL_VERSION_INT,};const AVClass *pclass = &class;ret = av_opt_eval_flags(&pclass, &opts[0], ost->disposition, &ost->st->disposition);if (ret < 0)return ret;}/* initialize bitstream filters for the output stream* needs to be done here, because the codec id for streamcopy is not* known until now */ret = init_output_bsfs(ost);if (ret < 0)return ret;ost->initialized = 1;ret = check_init_output_file(output_files[ost->file_index], ost->file_index);if (ret < 0)return ret;return ret; }static void report_new_stream(int input_index, AVPacket *pkt) {InputFile *file = input_files[input_index];AVStream *st = file->ctx->streams[pkt->stream_index];if (pkt->stream_index < file->nb_streams_warn)return;av_log(file->ctx, AV_LOG_WARNING,"New %s stream %d:%d at pos:%"PRId64" and DTS:%ss\n",av_get_media_type_string(st->codecpar->codec_type),input_index, pkt->stream_index,pkt->pos, av_ts2timestr(pkt->dts, &st->time_base));file->nb_streams_warn = pkt->stream_index + 1; }static int transcode_init(void) {int ret = 0, i, j, k;AVFormatContext *oc;OutputStream *ost;InputStream *ist;char error[1024] = {0};for (i = 0; i < nb_filtergraphs; i++) {FilterGraph *fg = filtergraphs[i];for (j = 0; j < fg->nb_outputs; j++) {OutputFilter *ofilter = fg->outputs[j];if (!ofilter->ost || ofilter->ost->source_index >= 0)continue;if (fg->nb_inputs != 1)continue;for (k = nb_input_streams-1; k >= 0 ; k--)if (fg->inputs[0]->ist == input_streams[k])break;ofilter->ost->source_index = k;}}/* init framerate emulation */for (i = 0; i < nb_input_files; i++) {InputFile *ifile = input_files[i];if (ifile->rate_emu)for (j = 0; j < ifile->nb_streams; j++)input_streams[j + ifile->ist_index]->start = av_gettime_relative();}/* init input streams */for (i = 0; i < nb_input_streams; i++)if ((ret = init_input_stream(i, error, sizeof(error))) < 0) {for (i = 0; i < nb_output_streams; i++) {ost = output_streams[i];avcodec_close(ost->enc_ctx);}goto dump_format;}/* open each encoder */for (i = 0; i < nb_output_streams; i++) {// skip streams fed from filtergraphs until we have a frame for themif (output_streams[i]->filter)continue;ret = init_output_stream(output_streams[i], error, sizeof(error));if (ret < 0)goto dump_format;}/* discard unused programs */for (i = 0; i < nb_input_files; i++) {InputFile *ifile = input_files[i];for (j = 0; j < ifile->ctx->nb_programs; j++) {AVProgram *p = ifile->ctx->programs[j];int discard = AVDISCARD_ALL;for (k = 0; k < p->nb_stream_indexes; k++)if (!input_streams[ifile->ist_index + p->stream_index[k]]->discard) {discard = AVDISCARD_DEFAULT;break;}p->discard = discard;}}/* write headers for files with no streams */for (i = 0; i < nb_output_files; i++) {oc = output_files[i]->ctx;if (oc->oformat->flags & AVFMT_NOSTREAMS && oc->nb_streams == 0) {ret = check_init_output_file(output_files[i], i);if (ret < 0)goto dump_format;}}dump_format:/* dump the stream mapping */av_log(NULL, AV_LOG_INFO, "Stream mapping:\n");for (i = 0; i < nb_input_streams; i++) {ist = input_streams[i];for (j = 0; j < ist->nb_filters; j++) {if (!filtergraph_is_simple(ist->filters[j]->graph)) {av_log(NULL, AV_LOG_INFO, " Stream #%d:%d (%s) -> %s",ist->file_index, ist->st->index, ist->dec ? ist->dec->name : "?",ist->filters[j]->name);if (nb_filtergraphs > 1)av_log(NULL, AV_LOG_INFO, " (graph %d)", ist->filters[j]->graph->index);av_log(NULL, AV_LOG_INFO, "\n");}}}for (i = 0; i < nb_output_streams; i++) {ost = output_streams[i];if (ost->attachment_filename) {/* an attached file */av_log(NULL, AV_LOG_INFO, " File %s -> Stream #%d:%d\n",ost->attachment_filename, ost->file_index, ost->index);continue;}if (ost->filter && !filtergraph_is_simple(ost->filter->graph)) {/* output from a complex graph */av_log(NULL, AV_LOG_INFO, " %s", ost->filter->name);if (nb_filtergraphs > 1)av_log(NULL, AV_LOG_INFO, " (graph %d)", ost->filter->graph->index);av_log(NULL, AV_LOG_INFO, " -> Stream #%d:%d (%s)\n", ost->file_index,ost->index, ost->enc ? ost->enc->name : "?");continue;}av_log(NULL, AV_LOG_INFO, " Stream #%d:%d -> #%d:%d",input_streams[ost->source_index]->file_index,input_streams[ost->source_index]->st->index,ost->file_index,ost->index);if (ost->sync_ist != input_streams[ost->source_index])av_log(NULL, AV_LOG_INFO, " [sync #%d:%d]",ost->sync_ist->file_index,ost->sync_ist->st->index);if (ost->stream_copy)av_log(NULL, AV_LOG_INFO, " (copy)");else {const AVCodec *in_codec = input_streams[ost->source_index]->dec;const AVCodec *out_codec = ost->enc;const char *decoder_name = "?";const char *in_codec_name = "?";const char *encoder_name = "?";const char *out_codec_name = "?";const AVCodecDescriptor *desc;if (in_codec) {decoder_name = in_codec->name;desc = avcodec_descriptor_get(in_codec->id);if (desc)in_codec_name = desc->name;if (!strcmp(decoder_name, in_codec_name))decoder_name = "native";}if (out_codec) {encoder_name = out_codec->name;desc = avcodec_descriptor_get(out_codec->id);if (desc)out_codec_name = desc->name;if (!strcmp(encoder_name, out_codec_name))encoder_name = "native";}av_log(NULL, AV_LOG_INFO, " (%s (%s) -> %s (%s))",in_codec_name, decoder_name,out_codec_name, encoder_name);}av_log(NULL, AV_LOG_INFO, "\n");}if (ret) {av_log(NULL, AV_LOG_ERROR, "%s\n", error);return ret;}atomic_store(&transcode_init_done, 1);return 0; }/* Return 1 if there remain streams where more output is wanted, 0 otherwise. */ static int need_output(void) {int i;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];OutputFile *of = output_files[ost->file_index];AVFormatContext *os = output_files[ost->file_index]->ctx;if (ost->finished ||(os->pb && avio_tell(os->pb) >= of->limit_filesize))continue;if (ost->frame_number >= ost->max_frames) {int j;for (j = 0; j < of->ctx->nb_streams; j++)close_output_stream(output_streams[of->ost_index + j]);continue;}return 1;}return 0; }/*** Select the output stream to process.** @return selected output stream, or NULL if none available*/ static OutputStream *choose_output(void) {int i;int64_t opts_min = INT64_MAX;OutputStream *ost_min = NULL;for (i = 0; i < nb_output_streams; i++) {OutputStream *ost = output_streams[i];int64_t opts = ost->st->cur_dts == AV_NOPTS_VALUE ? INT64_MIN :av_rescale_q(ost->st->cur_dts, ost->st->time_base,AV_TIME_BASE_Q);if (ost->st->cur_dts == AV_NOPTS_VALUE)av_log(NULL, AV_LOG_DEBUG, "cur_dts is invalid (this is harmless if it occurs once at the start per stream)\n");if (!ost->initialized && !ost->inputs_done)return ost;if (!ost->finished && opts < opts_min) {opts_min = opts;ost_min = ost->unavailable ? NULL : ost;}}return ost_min; }static void set_tty_echo(int on) { #if HAVE_TERMIOS_Hstruct termios tty;if (tcgetattr(0, &tty) == 0) {if (on) tty.c_lflag |= ECHO;else tty.c_lflag &= ~ECHO;tcsetattr(0, TCSANOW, &tty);} #endif }static int check_keyboard_interaction(int64_t cur_time) {int i, ret, key;static int64_t last_time;if (received_nb_signals)return AVERROR_EXIT;/* read_key() returns 0 on EOF */if(cur_time - last_time >= 100000 && !run_as_daemon){key = read_key();last_time = cur_time;}elsekey = -1;if (key == 'q')return AVERROR_EXIT;if (key == '+') av_log_set_level(av_log_get_level()+10);if (key == '-') av_log_set_level(av_log_get_level()-10);if (key == 's') qp_hist ^= 1;if (key == 'h'){if (do_hex_dump){do_hex_dump = do_pkt_dump = 0;} else if(do_pkt_dump){do_hex_dump = 1;} elsedo_pkt_dump = 1;av_log_set_level(AV_LOG_DEBUG);}if (key == 'c' || key == 'C'){char buf[4096], target[64], command[256], arg[256] = {0};double time;int k, n = 0;fprintf(stderr, "\nEnter command: <target>|all <time>|-1 <command>[ <argument>]\n");i = 0;set_tty_echo(1);while ((k = read_key()) != '\n' && k != '\r' && i < sizeof(buf)-1)if (k > 0)buf[i++] = k;buf[i] = 0;set_tty_echo(0);fprintf(stderr, "\n");if (k > 0 &&(n = sscanf(buf, "%63[^ ] %lf %255[^ ] %255[^\n]", target, &time, command, arg)) >= 3) {av_log(NULL, AV_LOG_DEBUG, "Processing command target:%s time:%f command:%s arg:%s",target, time, command, arg);for (i = 0; i < nb_filtergraphs; i++) {FilterGraph *fg = filtergraphs[i];if (fg->graph) {if (time < 0) {ret = avfilter_graph_send_command(fg->graph, target, command, arg, buf, sizeof(buf),key == 'c' ? AVFILTER_CMD_FLAG_ONE : 0);fprintf(stderr, "Command reply for stream %d: ret:%d res:\n%s", i, ret, buf);} else if (key == 'c') {fprintf(stderr, "Queuing commands only on filters supporting the specific command is unsupported\n");ret = AVERROR_PATCHWELCOME;} else {ret = avfilter_graph_queue_command(fg->graph, target, command, arg, 0, time);if (ret < 0)fprintf(stderr, "Queuing command failed with error %s\n", av_err2str(ret));}}}} else {av_log(NULL, AV_LOG_ERROR,"Parse error, at least 3 arguments were expected, ""only %d given in string '%s'\n", n, buf);}}if (key == 'd' || key == 'D'){int debug=0;if(key == 'D') {debug = input_streams[0]->st->codec->debug<<1;if(!debug) debug = 1;while(debug & (FF_DEBUG_DCT_COEFF|FF_DEBUG_VIS_QP|FF_DEBUG_VIS_MB_TYPE)) //unsupported, would just crashdebug += debug;}else{char buf[32];int k = 0;i = 0;set_tty_echo(1);while ((k = read_key()) != '\n' && k != '\r' && i < sizeof(buf)-1)if (k > 0)buf[i++] = k;buf[i] = 0;set_tty_echo(0);fprintf(stderr, "\n");if (k <= 0 || sscanf(buf, "%d", &debug)!=1)fprintf(stderr,"error parsing debug value\n");}for(i=0;i<nb_input_streams;i++) {input_streams[i]->st->codec->debug = debug;}for(i=0;i<nb_output_streams;i++) {OutputStream *ost = output_streams[i];ost->enc_ctx->debug = debug;}if(debug) av_log_set_level(AV_LOG_DEBUG);fprintf(stderr,"debug=%d\n", debug);}if (key == '?'){fprintf(stderr, "key function\n""? show this help\n""+ increase verbosity\n""- decrease verbosity\n""c Send command to first matching filter supporting it\n""C Send/Queue command to all matching filters\n""D cycle through available debug modes\n""h dump packets/hex press to cycle through the 3 states\n""q quit\n""s Show QP histogram\n");}return 0; }#if HAVE_PTHREADS static void *input_thread(void *arg) {InputFile *f = arg;unsigned flags = f->non_blocking ? AV_THREAD_MESSAGE_NONBLOCK : 0;int ret = 0;while (1) {AVPacket pkt;ret = av_read_frame(f->ctx, &pkt);if (ret == AVERROR(EAGAIN)) {av_usleep(10000);continue;}if (ret < 0) {av_thread_message_queue_set_err_recv(f->in_thread_queue, ret);break;}ret = av_thread_message_queue_send(f->in_thread_queue, &pkt, flags);if (flags && ret == AVERROR(EAGAIN)) {flags = 0;ret = av_thread_message_queue_send(f->in_thread_queue, &pkt, flags);av_log(f->ctx, AV_LOG_WARNING,"Thread message queue blocking; consider raising the ""thread_queue_size option (current value: %d)\n",f->thread_queue_size);}if (ret < 0) {if (ret != AVERROR_EOF)av_log(f->ctx, AV_LOG_ERROR,"Unable to send packet to main thread: %s\n",av_err2str(ret));av_packet_unref(&pkt);av_thread_message_queue_set_err_recv(f->in_thread_queue, ret);break;}}return NULL; }static void free_input_threads(void) {int i;for (i = 0; i < nb_input_files; i++) {InputFile *f = input_files[i];AVPacket pkt;if (!f || !f->in_thread_queue)continue;av_thread_message_queue_set_err_send(f->in_thread_queue, AVERROR_EOF);while (av_thread_message_queue_recv(f->in_thread_queue, &pkt, 0) >= 0)av_packet_unref(&pkt);pthread_join(f->thread, NULL);f->joined = 1;av_thread_message_queue_free(&f->in_thread_queue);} }static int init_input_threads(void) {int i, ret;if (nb_input_files == 1)return 0;for (i = 0; i < nb_input_files; i++) {InputFile *f = input_files[i];if (f->ctx->pb ? !f->ctx->pb->seekable :strcmp(f->ctx->iformat->name, "lavfi"))f->non_blocking = 1;ret = av_thread_message_queue_alloc(&f->in_thread_queue,f->thread_queue_size, sizeof(AVPacket));if (ret < 0)return ret;if ((ret = pthread_create(&f->thread, NULL, input_thread, f))) {av_log(NULL, AV_LOG_ERROR, "pthread_create failed: %s. Try to increase `ulimit -v` or decrease `ulimit -s`.\n", strerror(ret));av_thread_message_queue_free(&f->in_thread_queue);return AVERROR(ret);}}return 0; }static int get_input_packet_mt(InputFile *f, AVPacket *pkt) {return av_thread_message_queue_recv(f->in_thread_queue, pkt,f->non_blocking ?AV_THREAD_MESSAGE_NONBLOCK : 0); } #endifstatic int get_input_packet(InputFile *f, AVPacket *pkt) {if (f->rate_emu) {int i;for (i = 0; i < f->nb_streams; i++) {InputStream *ist = input_streams[f->ist_index + i];int64_t pts = av_rescale(ist->dts, 1000000, AV_TIME_BASE);int64_t now = av_gettime_relative() - ist->start;if (pts > now)return AVERROR(EAGAIN);}}#if HAVE_PTHREADSif (nb_input_files > 1)return get_input_packet_mt(f, pkt); #endifreturn av_read_frame(f->ctx, pkt); }static int got_eagain(void) {int i;for (i = 0; i < nb_output_streams; i++)if (output_streams[i]->unavailable)return 1;return 0; }static void reset_eagain(void) {int i;for (i = 0; i < nb_input_files; i++)input_files[i]->eagain = 0;for (i = 0; i < nb_output_streams; i++)output_streams[i]->unavailable = 0; }// set duration to max(tmp, duration) in a proper time base and return duration's time_base static AVRational duration_max(int64_t tmp, int64_t *duration, AVRational tmp_time_base,AVRational time_base) {int ret;if (!*duration) {*duration = tmp;return tmp_time_base;}ret = av_compare_ts(*duration, time_base, tmp, tmp_time_base);if (ret < 0) {*duration = tmp;return tmp_time_base;}return time_base; }static int seek_to_start(InputFile *ifile, AVFormatContext *is) {InputStream *ist;AVCodecContext *avctx;int i, ret, has_audio = 0;int64_t duration = 0;ret = av_seek_frame(is, -1, is->start_time, 0);if (ret < 0)return ret;for (i = 0; i < ifile->nb_streams; i++) {ist = input_streams[ifile->ist_index + i];avctx = ist->dec_ctx;// flush decodersif (ist->decoding_needed) {process_input_packet(ist, NULL, 1);avcodec_flush_buffers(avctx);}/* duration is the length of the last frame in a stream* when audio stream is present we don't care about* last video frame length because it's not defined exactly */if (avctx->codec_type == AVMEDIA_TYPE_AUDIO && ist->nb_samples)has_audio = 1;}for (i = 0; i < ifile->nb_streams; i++) {ist = input_streams[ifile->ist_index + i];avctx = ist->dec_ctx;if (has_audio) {if (avctx->codec_type == AVMEDIA_TYPE_AUDIO && ist->nb_samples) {AVRational sample_rate = {1, avctx->sample_rate};duration = av_rescale_q(ist->nb_samples, sample_rate, ist->st->time_base);} elsecontinue;} else {if (ist->framerate.num) {duration = av_rescale_q(1, ist->framerate, ist->st->time_base);} else if (ist->st->avg_frame_rate.num) {duration = av_rescale_q(1, ist->st->avg_frame_rate, ist->st->time_base);} else duration = 1;}if (!ifile->duration)ifile->time_base = ist->st->time_base;/* the total duration of the stream, max_pts - min_pts is* the duration of the stream without the last frame */duration += ist->max_pts - ist->min_pts;ifile->time_base = duration_max(duration, &ifile->duration, ist->st->time_base,ifile->time_base);}if (ifile->loop > 0)ifile->loop--;return ret; }/** Return* - 0 -- one packet was read and processed* - AVERROR(EAGAIN) -- no packets were available for selected file,* this function should be called again* - AVERROR_EOF -- this function should not be called again*/ static int process_input(int file_index) {InputFile *ifile = input_files[file_index];AVFormatContext *is;InputStream *ist;AVPacket pkt;int ret, i, j;int64_t duration;int64_t pkt_dts;is = ifile->ctx;ret = get_input_packet(ifile, &pkt);if (ret == AVERROR(EAGAIN)) {ifile->eagain = 1;return ret;}if (ret < 0 && ifile->loop) {if ((ret = seek_to_start(ifile, is)) < 0)return ret;ret = get_input_packet(ifile, &pkt);if (ret == AVERROR(EAGAIN)) {ifile->eagain = 1;return ret;}}if (ret < 0) {if (ret != AVERROR_EOF) {print_error(is->filename, ret);if (exit_on_error)exit_program(1);}for (i = 0; i < ifile->nb_streams; i++) {ist = input_streams[ifile->ist_index + i];if (ist->decoding_needed) {ret = process_input_packet(ist, NULL, 0);if (ret>0)return 0;}/* mark all outputs that don't go through lavfi as finished */for (j = 0; j < nb_output_streams; j++) {OutputStream *ost = output_streams[j];if (ost->source_index == ifile->ist_index + i &&(ost->stream_copy || ost->enc->type == AVMEDIA_TYPE_SUBTITLE))finish_output_stream(ost);}}ifile->eof_reached = 1;return AVERROR(EAGAIN);}reset_eagain();if (do_pkt_dump) {av_pkt_dump_log2(NULL, AV_LOG_INFO, &pkt, do_hex_dump,is->streams[pkt.stream_index]);}/* the following test is needed in case new streams appeardynamically in stream : we ignore them */if (pkt.stream_index >= ifile->nb_streams) {report_new_stream(file_index, &pkt);goto discard_packet;}ist = input_streams[ifile->ist_index + pkt.stream_index];ist->data_size += pkt.size;ist->nb_packets++;if (ist->discard)goto discard_packet;if (exit_on_error && (pkt.flags & AV_PKT_FLAG_CORRUPT)) {av_log(NULL, AV_LOG_FATAL, "%s: corrupt input packet in stream %d\n", is->filename, pkt.stream_index);exit_program(1);}if (debug_ts) {av_log(NULL, AV_LOG_INFO, "demuxer -> ist_index:%d type:%s ""next_dts:%s next_dts_time:%s next_pts:%s next_pts_time:%s pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s off:%s off_time:%s\n",ifile->ist_index + pkt.stream_index, av_get_media_type_string(ist->dec_ctx->codec_type),av_ts2str(ist->next_dts), av_ts2timestr(ist->next_dts, &AV_TIME_BASE_Q),av_ts2str(ist->next_pts), av_ts2timestr(ist->next_pts, &AV_TIME_BASE_Q),av_ts2str(pkt.pts), av_ts2timestr(pkt.pts, &ist->st->time_base),av_ts2str(pkt.dts), av_ts2timestr(pkt.dts, &ist->st->time_base),av_ts2str(input_files[ist->file_index]->ts_offset),av_ts2timestr(input_files[ist->file_index]->ts_offset, &AV_TIME_BASE_Q));}if(!ist->wrap_correction_done && is->start_time != AV_NOPTS_VALUE && ist->st->pts_wrap_bits < 64){int64_t stime, stime2;// Correcting starttime based on the enabled streams// FIXME this ideally should be done before the first use of starttime but we do not know which are the enabled streams at that point.// so we instead do it here as part of discontinuity handlingif ( ist->next_dts == AV_NOPTS_VALUE&& ifile->ts_offset == -is->start_time&& (is->iformat->flags & AVFMT_TS_DISCONT)) {int64_t new_start_time = INT64_MAX;for (i=0; i<is->nb_streams; i++) {AVStream *st = is->streams[i];if(st->discard == AVDISCARD_ALL || st->start_time == AV_NOPTS_VALUE)continue;new_start_time = FFMIN(new_start_time, av_rescale_q(st->start_time, st->time_base, AV_TIME_BASE_Q));}if (new_start_time > is->start_time) {av_log(is, AV_LOG_VERBOSE, "Correcting start time by %"PRId64"\n", new_start_time - is->start_time);ifile->ts_offset = -new_start_time;}}stime = av_rescale_q(is->start_time, AV_TIME_BASE_Q, ist->st->time_base);stime2= stime + (1ULL<<ist->st->pts_wrap_bits);ist->wrap_correction_done = 1;if(stime2 > stime && pkt.dts != AV_NOPTS_VALUE && pkt.dts > stime + (1LL<<(ist->st->pts_wrap_bits-1))) {pkt.dts -= 1ULL<<ist->st->pts_wrap_bits;ist->wrap_correction_done = 0;}if(stime2 > stime && pkt.pts != AV_NOPTS_VALUE && pkt.pts > stime + (1LL<<(ist->st->pts_wrap_bits-1))) {pkt.pts -= 1ULL<<ist->st->pts_wrap_bits;ist->wrap_correction_done = 0;}}/* add the stream-global side data to the first packet */if (ist->nb_packets == 1) {for (i = 0; i < ist->st->nb_side_data; i++) {AVPacketSideData *src_sd = &ist->st->side_data[i];uint8_t *dst_data;if (src_sd->type == AV_PKT_DATA_DISPLAYMATRIX)continue;if (av_packet_get_side_data(&pkt, src_sd->type, NULL))continue;dst_data = av_packet_new_side_data(&pkt, src_sd->type, src_sd->size);if (!dst_data)exit_program(1);memcpy(dst_data, src_sd->data, src_sd->size);}}if (pkt.dts != AV_NOPTS_VALUE)pkt.dts += av_rescale_q(ifile->ts_offset, AV_TIME_BASE_Q, ist->st->time_base);if (pkt.pts != AV_NOPTS_VALUE)pkt.pts += av_rescale_q(ifile->ts_offset, AV_TIME_BASE_Q, ist->st->time_base);if (pkt.pts != AV_NOPTS_VALUE)pkt.pts *= ist->ts_scale;if (pkt.dts != AV_NOPTS_VALUE)pkt.dts *= ist->ts_scale;pkt_dts = av_rescale_q_rnd(pkt.dts, ist->st->time_base, AV_TIME_BASE_Q, AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX);if ((ist->dec_ctx->codec_type == AVMEDIA_TYPE_VIDEO ||ist->dec_ctx->codec_type == AVMEDIA_TYPE_AUDIO) &&pkt_dts != AV_NOPTS_VALUE && ist->next_dts == AV_NOPTS_VALUE && !copy_ts&& (is->iformat->flags & AVFMT_TS_DISCONT) && ifile->last_ts != AV_NOPTS_VALUE) {int64_t delta = pkt_dts - ifile->last_ts;if (delta < -1LL*dts_delta_threshold*AV_TIME_BASE ||delta > 1LL*dts_delta_threshold*AV_TIME_BASE){ifile->ts_offset -= delta;av_log(NULL, AV_LOG_DEBUG,"Inter stream timestamp discontinuity %"PRId64", new offset= %"PRId64"\n",delta, ifile->ts_offset);pkt.dts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);if (pkt.pts != AV_NOPTS_VALUE)pkt.pts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);}}duration = av_rescale_q(ifile->duration, ifile->time_base, ist->st->time_base);if (pkt.pts != AV_NOPTS_VALUE) {pkt.pts += duration;ist->max_pts = FFMAX(pkt.pts, ist->max_pts);ist->min_pts = FFMIN(pkt.pts, ist->min_pts);}if (pkt.dts != AV_NOPTS_VALUE)pkt.dts += duration;pkt_dts = av_rescale_q_rnd(pkt.dts, ist->st->time_base, AV_TIME_BASE_Q, AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX);if ((ist->dec_ctx->codec_type == AVMEDIA_TYPE_VIDEO ||ist->dec_ctx->codec_type == AVMEDIA_TYPE_AUDIO) &&pkt_dts != AV_NOPTS_VALUE && ist->next_dts != AV_NOPTS_VALUE &&!copy_ts) {int64_t delta = pkt_dts - ist->next_dts;if (is->iformat->flags & AVFMT_TS_DISCONT) {if (delta < -1LL*dts_delta_threshold*AV_TIME_BASE ||delta > 1LL*dts_delta_threshold*AV_TIME_BASE ||pkt_dts + AV_TIME_BASE/10 < FFMAX(ist->pts, ist->dts)) {ifile->ts_offset -= delta;av_log(NULL, AV_LOG_DEBUG,"timestamp discontinuity %"PRId64", new offset= %"PRId64"\n",delta, ifile->ts_offset);pkt.dts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);if (pkt.pts != AV_NOPTS_VALUE)pkt.pts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);}} else {if ( delta < -1LL*dts_error_threshold*AV_TIME_BASE ||delta > 1LL*dts_error_threshold*AV_TIME_BASE) {av_log(NULL, AV_LOG_WARNING, "DTS %"PRId64", next:%"PRId64" st:%d invalid dropping\n", pkt.dts, ist->next_dts, pkt.stream_index);pkt.dts = AV_NOPTS_VALUE;}if (pkt.pts != AV_NOPTS_VALUE){int64_t pkt_pts = av_rescale_q(pkt.pts, ist->st->time_base, AV_TIME_BASE_Q);delta = pkt_pts - ist->next_dts;if ( delta < -1LL*dts_error_threshold*AV_TIME_BASE ||delta > 1LL*dts_error_threshold*AV_TIME_BASE) {av_log(NULL, AV_LOG_WARNING, "PTS %"PRId64", next:%"PRId64" invalid dropping st:%d\n", pkt.pts, ist->next_dts, pkt.stream_index);pkt.pts = AV_NOPTS_VALUE;}}}}if (pkt.dts != AV_NOPTS_VALUE)ifile->last_ts = av_rescale_q(pkt.dts, ist->st->time_base, AV_TIME_BASE_Q);if (debug_ts) {av_log(NULL, AV_LOG_INFO, "demuxer+ffmpeg -> ist_index:%d type:%s pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s off:%s off_time:%s\n",ifile->ist_index + pkt.stream_index, av_get_media_type_string(ist->dec_ctx->codec_type),av_ts2str(pkt.pts), av_ts2timestr(pkt.pts, &ist->st->time_base),av_ts2str(pkt.dts), av_ts2timestr(pkt.dts, &ist->st->time_base),av_ts2str(input_files[ist->file_index]->ts_offset),av_ts2timestr(input_files[ist->file_index]->ts_offset, &AV_TIME_BASE_Q));}sub2video_heartbeat(ist, pkt.pts);process_input_packet(ist, &pkt, 0);discard_packet:av_packet_unref(&pkt);return 0; }/*** Perform a step of transcoding for the specified filter graph.** @param[in] graph filter graph to consider* @param[out] best_ist input stream where a frame would allow to continue* @return 0 for success, <0 for error*/ static int transcode_from_filter(FilterGraph *graph, InputStream **best_ist) {int i, ret;int nb_requests, nb_requests_max = 0;InputFilter *ifilter;InputStream *ist;*best_ist = NULL;ret = avfilter_graph_request_oldest(graph->graph);if (ret >= 0)return reap_filters(0);if (ret == AVERROR_EOF) {ret = reap_filters(1);for (i = 0; i < graph->nb_outputs; i++)close_output_stream(graph->outputs[i]->ost);return ret;}if (ret != AVERROR(EAGAIN))return ret;for (i = 0; i < graph->nb_inputs; i++) {ifilter = graph->inputs[i];ist = ifilter->ist;if (input_files[ist->file_index]->eagain ||input_files[ist->file_index]->eof_reached)continue;nb_requests = av_buffersrc_get_nb_failed_requests(ifilter->filter);if (nb_requests > nb_requests_max) {nb_requests_max = nb_requests;*best_ist = ist;}}if (!*best_ist)for (i = 0; i < graph->nb_outputs; i++)graph->outputs[i]->ost->unavailable = 1;return 0; }/*** Run a single step of transcoding.** @return 0 for success, <0 for error*/ static int transcode_step(void) {OutputStream *ost;InputStream *ist = NULL;int ret;ost = choose_output();if (!ost) {if (got_eagain()) {reset_eagain();av_usleep(10000);return 0;}av_log(NULL, AV_LOG_VERBOSE, "No more inputs to read from, finishing.\n");return AVERROR_EOF;}if (ost->filter && !ost->filter->graph->graph) {if (ifilter_has_all_input_formats(ost->filter->graph)) {ret = configure_filtergraph(ost->filter->graph);if (ret < 0) {av_log(NULL, AV_LOG_ERROR, "Error reinitializing filters!\n");return ret;}}}if (ost->filter && ost->filter->graph->graph) {if ((ret = transcode_from_filter(ost->filter->graph, &ist)) < 0)return ret;if (!ist)return 0;} else if (ost->filter) {int i;for (i = 0; i < ost->filter->graph->nb_inputs; i++) {InputFilter *ifilter = ost->filter->graph->inputs[i];if (!ifilter->ist->got_output && !input_files[ifilter->ist->file_index]->eof_reached) {ist = ifilter->ist;break;}}if (!ist) {ost->inputs_done = 1;return 0;}} else {av_assert0(ost->source_index >= 0);ist = input_streams[ost->source_index];}ret = process_input(ist->file_index);if (ret == AVERROR(EAGAIN)) {if (input_files[ist->file_index]->eagain)ost->unavailable = 1;return 0;}if (ret < 0)return ret == AVERROR_EOF ? 0 : ret;return reap_filters(0); }/** The following code is the main loop of the file converter*/ static int transcode(void(call_back)(int current,int total)) {int ret, i;AVFormatContext *os;OutputStream *ost;InputStream *ist;int64_t timer_start;int64_t total_packets_written = 0;ret = transcode_init();if (ret < 0)goto fail;if (stdin_interaction) {av_log(NULL, AV_LOG_INFO, "Press [q] to stop, [?] for help\n");}timer_start = av_gettime_relative();#if HAVE_PTHREADSif ((ret = init_input_threads()) < 0)goto fail; #endifwhile (!received_sigterm) {int64_t cur_time= av_gettime_relative();/* if 'q' pressed, exits */if (stdin_interaction)if (check_keyboard_interaction(cur_time) < 0)break;/* check if there's any stream where output is still needed */if (!need_output()) {av_log(NULL, AV_LOG_VERBOSE, "No more output streams to write to, finishing.\n");break;}ret = transcode_step();if (ret < 0 && ret != AVERROR_EOF) {char errbuf[128];av_strerror(ret, errbuf, sizeof(errbuf));av_log(NULL, AV_LOG_ERROR, "Error while filtering: %s\n", errbuf);break;}/* dump report by using the output first video and audio streams */print_report(0, timer_start, cur_time,call_back);} #if HAVE_PTHREADSfree_input_threads(); #endif/* at the end of stream, we must flush the decoder buffers */for (i = 0; i < nb_input_streams; i++) {ist = input_streams[i];if (!input_files[ist->file_index]->eof_reached && ist->decoding_needed) {process_input_packet(ist, NULL, 0);}}flush_encoders();term_exit();/* write the trailer if needed and close file */for (i = 0; i < nb_output_files; i++) {os = output_files[i]->ctx;if (!output_files[i]->header_written) {av_log(NULL, AV_LOG_ERROR,"Nothing was written into output file %d (%s), because ""at least one of its streams received no packets.\n",i, os->filename);continue;}if ((ret = av_write_trailer(os)) < 0) {av_log(NULL, AV_LOG_ERROR, "Error writing trailer of %s: %s\n", os->filename, av_err2str(ret));if (exit_on_error)exit_program(1);}}/* dump report by using the first video and audio streams */print_report(1, timer_start, av_gettime_relative(),call_back);/* close each encoder */for (i = 0; i < nb_output_streams; i++) {ost = output_streams[i];if (ost->encoding_needed) {av_freep(&ost->enc_ctx->stats_in);}total_packets_written += ost->packets_written;}if (!total_packets_written && (abort_on_flags & ABORT_ON_FLAG_EMPTY_OUTPUT)) {av_log(NULL, AV_LOG_FATAL, "Empty output\n");exit_program(1);}/* close each decoder */for (i = 0; i < nb_input_streams; i++) {ist = input_streams[i];if (ist->decoding_needed) {avcodec_close(ist->dec_ctx);if (ist->hwaccel_uninit)ist->hwaccel_uninit(ist->dec_ctx);}}av_buffer_unref(&hw_device_ctx);/* finished ! */ret = 0;fail: #if HAVE_PTHREADSfree_input_threads(); #endifif (output_streams) {for (i = 0; i < nb_output_streams; i++) {ost = output_streams[i];if (ost) {if (ost->logfile) {if (fclose(ost->logfile))av_log(NULL, AV_LOG_ERROR,"Error closing logfile, loss of information possible: %s\n",av_err2str(AVERROR(errno)));ost->logfile = NULL;}av_freep(&ost->forced_kf_pts);av_freep(&ost->apad);av_freep(&ost->disposition);av_dict_free(&ost->encoder_opts);av_dict_free(&ost->sws_dict);av_dict_free(&ost->swr_opts);av_dict_free(&ost->resample_opts);}}}return ret; }static int64_t getutime(void) { #if HAVE_GETRUSAGEstruct rusage rusage;getrusage(RUSAGE_SELF, &rusage);return (rusage.ru_utime.tv_sec * 1000000LL) + rusage.ru_utime.tv_usec; #elif HAVE_GETPROCESSTIMESHANDLE proc;FILETIME c, e, k, u;proc = GetCurrentProcess();GetProcessTimes(proc, &c, &e, &k, &u);return ((int64_t) u.dwHighDateTime << 32 | u.dwLowDateTime) / 10; #elsereturn av_gettime_relative(); #endif }static int64_t getmaxrss(void) { #if HAVE_GETRUSAGE && HAVE_STRUCT_RUSAGE_RU_MAXRSSstruct rusage rusage;getrusage(RUSAGE_SELF, &rusage);return (int64_t)rusage.ru_maxrss * 1024; #elif HAVE_GETPROCESSMEMORYINFOHANDLE proc;PROCESS_MEMORY_COUNTERS memcounters;proc = GetCurrentProcess();memcounters.cb = sizeof(memcounters);GetProcessMemoryInfo(proc, &memcounters, sizeof(memcounters));return memcounters.PeakPagefileUsage; #elsereturn 0; #endif }static void log_callback_null(void *ptr, int level, const char *fmt, va_list vl) { } //命令函數(shù)的入口 int run_ffmpeg_command(int argc, char **argv,void(call_back)(int current,int total)) {int i, ret;int64_t ti;init_dynload();register_exit(ffmpeg_cleanup);setvbuf(stderr,NULL,_IONBF,0); /* win32 runtime needs this */av_log_set_flags(AV_LOG_SKIP_REPEATED);parse_loglevel(argc, argv, options);if(argc>1 && !strcmp(argv[1], "-d")){run_as_daemon=1;av_log_set_callback(log_callback_null);argc--;argv++;}avcodec_register_all(); #if CONFIG_AVDEVICEavdevice_register_all(); #endifavfilter_register_all();av_register_all();avformat_network_init();show_banner(argc, argv, options);/* parse options and open all input/output files */ret = ffmpeg_parse_options(argc, argv);if (ret < 0)exit_program(1);if (nb_output_files <= 0 && nb_input_files == 0) {show_usage();av_log(NULL, AV_LOG_WARNING, "Use -h to get full help or, even better, run 'man %s'\n", program_name);exit_program(1);}/* file converter / grab */if (nb_output_files <= 0) {av_log(NULL, AV_LOG_FATAL, "At least one output file must be specified\n");exit_program(1);}// if (nb_input_files == 0) { // av_log(NULL, AV_LOG_FATAL, "At least one input file must be specified\n"); // exit_program(1); // }for (i = 0; i < nb_output_files; i++) {if (strcmp(output_files[i]->ctx->oformat->name, "rtp"))want_sdp = 0;}current_time = ti = getutime();if (transcode(call_back) < 0)exit_program(1);ti = getutime() - ti;if (do_benchmark) {av_log(NULL, AV_LOG_INFO, "bench: utime=%0.3fs\n", ti / 1000000.0);}av_log(NULL, AV_LOG_DEBUG, "%"PRIu64" frames successfully decoded, %"PRIu64" decoding errors\n",decode_error_stat[0], decode_error_stat[1]);if ((decode_error_stat[0] + decode_error_stat[1]) * max_error_rate < decode_error_stat[1])exit_program(69);//注釋掉此行代碼 如果不注釋掉該行代碼,命令執(zhí)行完會(huì)導(dǎo)致app退出//exit_program(received_nb_signals ? 255 : main_return_code);//添加如下代碼如果不添加再次執(zhí)行run_ffmpeg_command 直接崩潰 因?yàn)闆]有賦初始值nb_filtergraphs = 0;nb_input_streams = 0;nb_input_files = 0;progress_avio = NULL;input_streams = NULL;nb_input_streams = 0;input_files = NULL;nb_input_files = 0;output_streams = NULL;nb_output_streams = 0;output_files = NULL;nb_output_files = 0;return main_return_code; }

修改native-lib.cpp內(nèi)容

?

?native-lib.cpp內(nèi)容如下:

#include <jni.h> #include <string> #include <android/log.h> #define TAG "JNI_TAG" #define LOGE(...) __android_log_print(ANDROID_LOG_ERROR,TAG,__VA_ARGS__); extern "C" { #include "libavutil/avutil.h"//聲明方法 argc命令的個(gè)數(shù) argv 二維數(shù)組 int run_ffmpeg_command(int argc, char **argv,void(call_back)(int current,int total)); //回調(diào)函數(shù) static jobject call_back_jobj;//call_back_jobj 對(duì)應(yīng)于VideoCompress類中的CompressCallback static JNIEnv *myEnv; void call_back(int current,int total){//LOGE("壓縮進(jìn)度:%d/%d",current,total);//把進(jìn)度回調(diào)出去,對(duì)象是jobject callbackif(myEnv!=NULL&&call_back_jobj!=NULL){//獲取j_mid是會(huì)被多次執(zhí)行jclass j_clazz=myEnv->GetObjectClass(call_back_jobj);//方法 onCompress對(duì)象的方法名 對(duì)應(yīng)VideoCompress類中的CompressCallback接口的onCompress方法//(II)V(方法簽名 利用javap命令也能打印) --(public void onCompress(int current,int total);onCompress的兩個(gè)參數(shù)都為int類型 onCompress方法的返回值為void)jmethodID j_mid=myEnv->GetMethodID(j_clazz,"onCompress","(II)V");//調(diào)用對(duì)象的方法 如果寫CallObjectMethod報(bào)錯(cuò),因?yàn)橹朗菬o返回值的方法,所以改成CallVoidMethodmyEnv->CallVoidMethod(call_back_jobj,j_mid,current,total);}} } extern "C" JNIEXPORT jstring JNICALL Java_com_suoer_ndk_ffmpegtestapplication_MainActivity_stringFromJNI(JNIEnv* env,jobject /* this */) {//std::string hello = "Hello from C++";return env->NewStringUTF(av_version_info()); }extern "C" JNIEXPORT void JNICALL Java_com_suoer_ndk_ffmpegtestapplication_VideoCompress_compressVideo(JNIEnv *env, jobject thiz,jobjectArray compress_command,jobject callback) {myEnv=env;call_back_jobj=env->NewGlobalRef(callback);//ffmpeg 處理視頻壓縮//arm這個(gè)里面的so都是用來處理音視頻的,include都是頭文件//還有幾個(gè)沒有被打包編譯成so,因?yàn)檫@些不算是音視頻的處理代碼,只是我們現(xiàn)在支持命令(封裝)//1.獲取命令個(gè)數(shù)int argc=env->GetArrayLength(compress_command);//2.給char **argv填充數(shù)據(jù)char **argv=(char **)malloc(sizeof(char*)*argc);for (int i = 0; i <argc ; ++i) {jstring j_param=(jstring)env->GetObjectArrayElement(compress_command,i);argv[i]= (char *)(env->GetStringUTFChars(j_param, NULL));LOGE("參數(shù):%s",argv[i]);}//3.調(diào)用命令函數(shù)去壓縮run_ffmpeg_command(argc,argv,call_back);//4.釋放內(nèi)存for (int i = 0; i <argc ; ++i) {free(argv[i]);}free(argv);env->DeleteGlobalRef(call_back_jobj); }

5.MainActivity界面優(yōu)化處理

首先需要把a(bǔ)ssets目錄中的test.mp4放置手機(jī)中

?

?修改MainActivity內(nèi)容:

壓縮之前首先判斷壓縮后的文件out.mp4是否存在,如果已經(jīng)存在,則刪除此文件。如果壓縮文件已經(jīng)存在再次壓縮會(huì)導(dǎo)致崩潰。

?

?

壓縮過程中展示進(jìn)度條,壓縮過程中不允許再次點(diǎn)擊壓縮按鈕

?

進(jìn)度彈窗背景:?bg_4radius.xml

<?xml version="1.0" encoding="utf-8"?> <selector xmlns:android="http://schemas.android.com/apk/res/android"><item><shape android:shape="rectangle"><solid android:color="@color/white" /><corners android:topLeftRadius="4dp" android:topRightRadius="4dp" android:bottomLeftRadius="4dp" android:bottomRightRadius="4dp"/></shape></item> </selector>

進(jìn)度條進(jìn)度背景:compress_progressbar.xml

<?xml version="1.0" encoding="utf-8"?> <layer-list xmlns:android="http://schemas.android.com/apk/res/android"><item android:id="@android:id/background"><!--形狀--><shape><!-- 角球--><corners android:radius="6dp" /><solid android:color="#ffeeeeee"></solid></shape></item><item android:id="@android:id/secondaryProgress"><clip><shape><corners android:radius="6dp" /><gradientandroid:type="linear"android:useLevel="true"android:angle="270"android:endColor="@color/purple_500"android:startColor="@color/purple_200" /></shape></clip></item><item android:id="@android:id/progress"><clip><shape><corners android:radius="6dp" /><gradientandroid:type="linear"android:useLevel="true"android:angle="270"android:endColor="@color/purple_500"android:startColor="@color/purple_200" /></shape></clip></item> </layer-list>

進(jìn)度彈窗布局文件dialog_progress.xml

<?xml version="1.0" encoding="utf-8"?> <RelativeLayout xmlns:android="http://schemas.android.com/apk/res/android"android:layout_width="wrap_content"android:layout_height="320dp"android:layout_gravity="center"android:background="@drawable/bg_4radius"><RelativeLayoutandroid:layout_width="match_parent"android:layout_height="match_parent"><LinearLayoutandroid:layout_centerInParent="true"android:layout_width="match_parent"android:layout_height="wrap_content"android:gravity="center_horizontal"android:orientation="vertical"><TextViewandroid:id="@+id/tv_dialog_compress_progress"android:layout_width="wrap_content"android:layout_height="wrap_content"android:textColor="@color/purple_500"android:textSize="24sp" /><ProgressBarandroid:id="@+id/compress_progressbar"android:layout_marginLeft="40dp"android:layout_marginRight="40dp"style="?android:attr/progressBarStyleHorizontal"android:layout_width="match_parent"android:max="100"android:min="0"android:progressDrawable="@drawable/compress_progressbar"android:layout_height="12dp" /><TextViewandroid:layout_marginBottom="20dp"android:text="視頻正在壓縮,請(qǐng)稍等…"android:layout_marginTop="40dp"android:textSize="14sp"android:textColor="#CCCCCC"android:layout_width="wrap_content"android:layout_height="wrap_content" /></LinearLayout></RelativeLayout></RelativeLayout>

彈窗樣式styles.xml

<?xml version="1.0" encoding="utf-8"?> <resources><style name="normal_dialog" parent="android:Theme.Dialog"><item name="android:background">@android:color/transparent</item><item name="android:windowBackground">@android:color/transparent</item><item name="android:windowNoTitle">true</item><item name="android:backgroundDimAmount">0.8</item></style> </resources>

彈窗ProgressDialog.java

package com.suoer.ndk.ffmpegtestapplication;import android.app.Dialog; import android.content.Context; import android.widget.ProgressBar; import android.widget.TextView;import androidx.annotation.NonNull;public class ProgressDialog extends Dialog {private TextView tvProgress;private ProgressBar compress_progressbar;public ProgressDialog(@NonNull Context context) {this(context, R.style.normal_dialog);setContentView(R.layout.dialog_progress);tvProgress = findViewById(R.id.tv_dialog_compress_progress);compress_progressbar =findViewById(R.id.compress_progressbar);}private ProgressDialog(@NonNull Context context, int themeResId) {super(context, themeResId);}public void setProgress(int progress) {if (progress < 0) {progress = 0;} else if (progress > 100) {progress = 100;}tvProgress.setText(progress + "%");compress_progressbar.setProgress(progress);} }

最終的MainActivity.java

package com.suoer.ndk.ffmpegtestapplication;import android.Manifest; import android.os.Bundle; import android.os.Environment; import android.util.Log; import android.view.View; import android.widget.TextView; import android.widget.Toast;import com.tbruyelle.rxpermissions3.RxPermissions;import java.io.File;import androidx.appcompat.app.AppCompatActivity; import io.reactivex.rxjava3.android.schedulers.AndroidSchedulers; import io.reactivex.rxjava3.core.Observable; import io.reactivex.rxjava3.functions.Consumer; import io.reactivex.rxjava3.functions.Function; import io.reactivex.rxjava3.schedulers.Schedulers;public class MainActivity extends AppCompatActivity {private File mInFile=new File(Environment.getExternalStorageDirectory(),"test.mp4");//mInFile 需要壓縮的文件路徑private File mOutFile=new File(Environment.getExternalStorageDirectory(),"out.mp4");//mOutFile壓縮后的文件路徑private ProgressDialog mProgressDialog;//進(jìn)度彈窗// Used to load the 'native-lib' library on application startup.static {System.loadLibrary("native-lib");}@Overrideprotected void onCreate(Bundle savedInstanceState) {super.onCreate(savedInstanceState);setContentView(R.layout.activity_main);if (mProgressDialog == null) {mProgressDialog = new ProgressDialog(this);}mProgressDialog.setCancelable(false);mProgressDialog.setProgress(0);// Example of a call to a native methodTextView tv = findViewById(R.id.sample_text);//tv.setText("ffmpeg版本:"+stringFromJNI());tv.setText("壓縮");//tv的點(diǎn)擊事件 點(diǎn)擊按鈕實(shí)現(xiàn)視頻壓縮tv.setOnClickListener(new View.OnClickListener() {@Overridepublic void onClick(View v) {// 壓縮文件 需要讀寫文件權(quán)限 申請(qǐng)權(quán)限RxPermissions rxPermissions=new RxPermissions(MainActivity.this);rxPermissions.request(Manifest.permission.READ_EXTERNAL_STORAGE, Manifest.permission.WRITE_EXTERNAL_STORAGE).subscribe(new Consumer<Boolean>() {@Overridepublic void accept(Boolean aBoolean) throws Throwable {if(aBoolean){//權(quán)限已經(jīng)獲取 壓縮視頻//如果手機(jī)中不存在test.mp4的文件則提示if(!mInFile.exists()){Toast.makeText(MainActivity.this,"請(qǐng)把a(bǔ)ssert目錄中的test.mp4拷貝至手機(jī)中!",Toast.LENGTH_LONG).show();return;}if(mOutFile.exists()){mOutFile.delete();}compressVideo();}}});}});}/*** 開啟子線程處理耗時(shí)壓縮問題*/private void compressVideo() {//ffmpeg的壓縮命令:ffmpeg -i test.mp4 -b:v 1024k out.mp4//ffmpeg -i test.mp4 -b:v 1024k out.mp4//-b:v 1024k 1024k為碼率 碼率越高視頻越清晰,而且視頻越大//test.mp4需要壓縮的文件//out.mp4 壓縮之后的文件if(mProgressDialog!=null&&!mProgressDialog.isShowing()){mProgressDialog.show();}String[] compressCommand={"ffmpeg","-i",mInFile.getAbsolutePath(),"-b:v","1024k",mOutFile.getAbsolutePath()};//壓縮是耗時(shí)的,需要子線程處理Observable.just(compressCommand).map(new Function<String[], File>() {@Overridepublic File apply(String[] compressCommand) throws Throwable {//子線程VideoCompress videoCompress=new VideoCompress();videoCompress.compressVideo(compressCommand, new VideoCompress.CompressCallback() {@Overridepublic void onCompress(int current, int total) {Log.e("TAG", "onCompress: 壓縮進(jìn)度:"+current+"/"+total);runOnUiThread(new Runnable() {@Overridepublic void run() {if(mProgressDialog!=null){mProgressDialog.setProgress(100 * current / total);}}});}});return mOutFile;}}).subscribeOn(Schedulers.io()).observeOn(AndroidSchedulers.mainThread()).subscribe(new Consumer<File>() {@Overridepublic void accept(File file) throws Throwable {// 主線程 壓縮完成Log.e("TAG", "accept: 壓縮完成!" );if(mProgressDialog!=null){mProgressDialog.dismiss();}}});}/*** A native method that is implemented by the 'native-lib' native library,* which is packaged with this application.*/public native String stringFromJNI();@Overrideprotected void onDestroy() {super.onDestroy();if (mProgressDialog != null) {mProgressDialog.cancel();mProgressDialog = null;}} }

總結(jié)

以上是生活随笔為你收集整理的Android 使用FFmpeg3.3.9基于命令实现视频压缩的全部內(nèi)容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網(wǎng)站內(nèi)容還不錯(cuò),歡迎將生活随笔推薦給好友。

天天曰视频 | 操操操影院 | 亚洲精品在线观看av | 91丝袜美腿 | 香蕉91视频 | 午夜精品在线看 | 久久午夜网 | 麻豆久久久 | 四虎国产精品永久在线国在线 | 狠狠狠色丁香婷婷综合激情 | 国产午夜精品一区二区三区在线观看 | 国产一在线精品一区在线观看 | 国产精品毛片久久久久久久久久99999999 | 国产精品黄色 | www.久久爱.cn | 国产丝袜一区二区三区 | 国产午夜精品福利视频 | 中文在线a∨在线 | 成年人免费在线观看 | 中文字幕欧美日韩va免费视频 | 精品久久久久久国产偷窥 | 狠狠色丁香婷婷综合 | 久久久久久久久久久久久久av | 欧美淫aaa免费观看 日韩激情免费视频 | 一级一级一片免费 | 99精品视频免费观看 | 福利视频| 欧美日在线观看 | 日韩极品在线 | 高清av影院 | 9幺看片 | 国产精品视频久久久 | 一区二区三区中文字幕在线观看 | 9992tv成人免费看片 | 日韩精品久久久 | 国内精品免费久久影院 | 亚洲国产资源 | 特级毛片aaa| 国产一区成人在线 | 狠狠色丁香 | 91超碰免费在线 | 国产一级做a | 国产欧美综合在线观看 | 中文av在线播放 | 免费亚洲精品视频 | 丁香视频在线观看 | 九九日韩 | 亚洲极色 | 91xav| 高潮久久久久久 | 日韩欧美v| 欧美婷婷综合 | 色婷婷 亚洲 | 日韩有码网站 | www.五月天激情 | 色综合婷婷久久 | 免费看片色 | 美女在线免费视频 | 成人福利在线播放 | 99精品亚洲| 最近中文国产在线视频 | 亚洲国产黄色片 | 免费看一级特黄a大片 | 欧美黑人巨大xxxxx | 亚州中文av | 在线中文字幕网站 | 欧美在线91 | 四虎成人免费观看 | 日韩三区在线 | 欧美性精品 | 操操操人人 | 久久久久日本精品一区二区三区 | 综合婷婷 | 日韩理论在线 | 91精品视频免费在线观看 | 超碰在线观看97 | 天天色天天操天天爽 | 欧美日韩高清国产 | 日本护士三级少妇三级999 | 午夜精品福利一区二区三区蜜桃 | 美女网站黄在线观看 | 五月激情六月丁香 | 91字幕 | 午夜国产在线 | 欧美午夜剧场 | 日韩视频在线观看视频 | 国产精品欧美久久久久天天影视 | 在线观看自拍 | 精品91视频 | 国产一级二级三级视频 | 草樱av | 999国内精品永久免费视频 | 国产精品尤物视频 | 亚洲综合五月 | 美女网色 | 亚洲最大成人免费网站 | 亚洲三级av | 亚洲成人精品 | 在线观看黄色 | 色爱成人网 | 91久久国产综合精品女同国语 | 五月婷婷激情 | 99视频偷窥在线精品国自产拍 | 亚洲视频在线免费观看 | 免费黄色a网站 | 一级电影免费在线观看 | 久久99久久99精品免观看粉嫩 | 美女av电影 | 天天操天天草 | 国产精品亚洲片夜色在线 | 亚洲综合日韩在线 | 欧美婷婷色| 亚洲人毛片 | 欧美日产一区 | 亚洲综合视频在线观看 | 国产日韩亚洲 | 美女在线黄 | 欧美少妇影院 | 成人一级电影在线观看 | 精品亚洲成a人在线观看 | 亚洲国产经典视频 | 五月天综合色 | 8x成人在线 | 久久久久久久久久久精 | 久久精品视频在线观看免费 | 久久99国产精品久久99 | 一区二区 不卡 | 欧美精品亚洲精品日韩精品 | 99爱在线 | 日本黄色免费看 | 夜夜爽88888免费视频4848 | 国产黄色一级片在线 | 国产无限资源在线观看 | 黄色大片免费网站 | 日韩在线观看视频在线 | 日韩激情影院 | 精品a级片 | 99在线观看精品 | 欧美日韩在线视频免费 | 五月天色网站 | 成人性生爱a∨ | 国产1区2区3区精品美女 | 91尤物国产尤物福利在线播放 | 亚洲精品美女久久17c | 黄av在线| 国产精品一区久久久久 | 国产一区视频在线 | 77国产精品| 亚洲免费av在线播放 | 99热官网 | 国产精品毛片一区二区 | 日韩高清av| 91精品综合| 精品国产1区2区 | 在线精品一区二区 | 成人久久网| 日韩免费一区二区 | 99久久综合精品五月天 | 国产一区视频在线观看免费 | 黄色毛片电影 | 天堂在线视频中文网 | 色综合五月| 亚洲激情在线播放 | 91插插插免费视频 | 国产一区欧美二区 | 中文字幕视频一区 | 精品二区久久 | 成人av在线网 | 亚洲视频久久 | 久久久久久久免费看 | 午夜精品久久久久久 | 99精品视频免费看 | 天天弄天天操 | 91黄色在线看 | 午夜精品电影 | 久久理论电影 | 精品电影一区二区 | 91理论片午午伦夜理片久久 | 美女久久久久 | 九九久久久久久久久激情 | www.天天操 | 午夜a区 | 丁香六月婷婷开心婷婷网 | 久一网站 | 激情五月看片 | 美女网站久久 | 在线不卡的av | 97在线观看免费观看高清 | 国产不卡免费 | 日夜夜精品视频 | 国产v亚洲v| 韩日在线一区 | 国产精品免费一区二区 | 欧美一二三视频 | 三级av在线免费观看 | 天天干天天射天天操 | 91成人免费看片 | 韩国一区二区在线观看 | 日韩黄色免费看 | 成人九九视频 | 日韩黄视频 | 日韩精品免费一区二区 | 国产精品video爽爽爽爽 | 99久久精品久久久久久清纯 | 国产一级性生活视频 | 在线观看中文字幕网站 | 91麻豆精品国产91久久久久久久久 | 天天摸天天操天天爽 | 午夜黄色大片 | 亚洲一级片免费观看 | 久久精品免费 | 日韩av在线看 | 96在线| 久草在线最新视频 | 国产视频1 | 日韩91在线 | 国产日韩欧美在线一区 | 久久96国产精品久久99软件 | 色五月激情五月 | 午夜视频在线观看网站 | 中文字幕在线网址 | 91精品国产乱码在线观看 | 全久久久久久久久久久电影 | 欧美色综合久久 | 亚洲人在线7777777精品 | av日韩中文 | 国产精品字幕 | 福利电影久久 | 亚洲影视九九影院在线观看 | 一级片免费视频 | 久久久久久久久黄色 | 国产日韩欧美在线观看视频 | 国产视频在线一区二区 | 成人h在线播放 | 亚洲日本精品视频 | 亚洲精品一区二区网址 | av字幕在线 | 在线综合 亚洲 欧美在线视频 | 亚洲成a人片77777kkkk1在线观看 | 美女天天操 | 久久精品这里都是精品 | 999久久 | 波多野结衣视频在线 | 1024久久 | 高清av在线免费观看 | 麻豆视频一区 | 亚洲三级毛片 | 国产精品免费久久久久久 | 日韩高清一区在线 | 欧美三级高清 | 亚洲人毛片 | 精品久久网站 | 黄色亚洲| 在线播放日韩av | 91精品国自产在线 | 国产一卡久久电影永久 | 五月婷婷综合激情 | 人人草人人草 | 亚洲第一中文字幕 | 五月婷婷丁香 | 高潮毛片无遮挡高清免费 | 欧美色婷 | www国产在线 | 国产精品久久久久999 | 国产精品资源在线观看 | 97操操| 欧美日韩国产色综合一二三四 | 成年人电影免费在线观看 | 一级精品视频在线观看宜春院 | 综合伊人av | 4438全国亚洲精品观看视频 | 欧美少妇的秘密 | 色综合网在线 | 国产精品一区二区久久精品爱微奶 | 麻豆播放 | 在线中文字幕网站 | 一级α片| 亚洲成av人片在线观看香蕉 | 九七人人干 | 亚洲中字幕 | 96av在线视频| 免费看av片网站 | 亚一亚二国产专区 | 午夜视频在线观看一区二区三区 | 香蕉视频国产在线观看 | 国产韩国日本高清视频 | 九九有精品 | 久久精品欧美一 | 激情丁香久久 | 99国产精品一区 | 天天操夜夜做 | 九九有精品 | 久久中文字幕视频 | 免费在线成人 | 在线观看www. | 亚洲精品午夜久久久久久久 | 亚洲h在线播放在线观看h | 99久久婷婷国产 | 婷婷av网站 | 久久99网站| 九九色在线观看 | 色综合久久中文字幕综合网 | 在线播放亚洲激情 | 97在线观视频免费观看 | 国产成人在线观看免费 | 国产免费又爽又刺激在线观看 | 国产美腿白丝袜足在线av | av电影在线免费 | 五月视频 | 久久精品99久久久久久2456 | 97夜夜澡人人双人人人喊 | 欧美国产日韩中文 | 综合伊人久久 | 成人91免费视频 | 91成人在线视频 | 美女福利视频 | 在线看片日韩 | 91成品视频 | 亚洲激精日韩激精欧美精品 | 丁香综合五月 | 国产原创91 | 91成人破解版 | 亚洲欧洲在线视频 | 狠狠色狠狠色综合日日小说 | 精品在线看 | 色婷婷天天干 | 三日本三级少妇三级99 | www.久热| 99精品在线免费 | 久久久久免费视频 | 日韩精品观看 | 国产精品视频最多的网站 | 久久国产免费看 | 婷婷伊人五月 | 免费观看性生交 | 美女国产免费 | 伊香蕉大综综综合久久啪 | 奇米网8888| 狠狠操夜夜操 | 国产成人99av超碰超爽 | www欧美日韩| 三级av小说 | 亚洲精品a区 | 国产一级二级在线播放 | 国产成人资源 | 国产特黄色片 | 亚洲成人黄色av | 国产色就色| 日韩一级电影在线 | 狠狠色香婷婷久久亚洲精品 | 亚洲成av人片一区二区梦乃 | 日本大尺码专区mv | 成人免费视频网站 | 夜夜躁日日躁 | 免费看片网站91 | 国产精品乱码久久久久久1区2区 | 久久久久女教师免费一区 | 亚洲综合黄色 | 久久精品国产免费看久久精品 | 国产最新91| 久久久国产精品人人片99精片欧美一 | 亚洲精品国产精品久久99热 | 久久久国产毛片 | 日韩毛片在线播放 | 91 在线视频播放 | 伊人永久在线 | 精品国产伦一区二区三区观看说明 | a级一a一级在线观看 | 亚洲国产精品成人综合 | 一区二区三区久久 | 免费中文字幕视频 | 激情综合五月天 | 久久精品79国产精品 | 久久精品视频免费观看 | 久久久久久久久爱 | www.狠狠干| 黄色三级免费 | 免费高清在线视频一区· | 国产经典av| 日韩免费电影一区二区 | 日操操| 日韩视频免费在线观看 | wwwwww黄| 99国产精品一区二区 | 五月亚洲综合 | 久久九九精品久久 | 日韩精品国产一区 | 欧美激情视频在线观看免费 | 亚洲色图27p | 久久男人中文字幕资源站 | 久久精品男人的天堂 | 波多野结衣一区二区三区中文字幕 | 黄色网址a| 国产一级视频免费看 | 欧美日韩性视频在线 | 国产精品美女久久久久久2018 | 久久福利综合 | 免费看91的网站 | 免费看黄的 | 色婷婷综合久久久中文字幕 | 日韩精品欧美精品 | 9幺看片| 久久久久五月天 | 69国产精品视频 | 夜添久久精品亚洲国产精品 | 国产亚洲人 | 亚洲狠狠丁香婷婷综合久久久 | 国产成人三级三级三级97 | 欧美国产大片 | 狠狠色丁香久久婷婷综合_中 | 久热这里有精品 | 亚洲精品国产精品国自产在线 | 亚洲一区免费在线 | 免费三级a | 黄色毛片在线 | 在线观看91网站 | 91九色精品国产 | 日韩黄色在线 | 久久电影中文字幕视频 | 日韩色在线 | 国产成人久久精品77777综合 | 欧美午夜激情网 | 欧美日韩视频一区二区三区 | 中文字幕在线久一本久 | 国产日产高清dvd碟片 | 久久综合五月婷婷 | 国产一级片久久 | av一本久道久久波多野结衣 | av电影 一区二区 | 国产成人区 | 欧美精品v国产精品v日韩精品 | 黄色毛片一级 | 国产视频在线观看一区 | 国产精品高清一区二区三区 | 国产精品成人免费一区久久羞羞 | 亚洲精品乱码久久 | 天天草天天操 | 不卡视频国产 | 久久精品国产亚洲精品2020 | 天天爱天天射天天干天天 | 日韩激情视频 | 午夜精品久久久 | 国产精品www| 国产精品毛片网 | 天天操天天射天天 | 超碰在线98 | 深爱开心激情网 | www.天堂av| 亚洲黑丝少妇 | av免费看电影 | 天堂视频中文在线 | 久久高清免费视频 | 国产香蕉视频在线观看 | 天天插一插 | 999久久久精品视频 日韩高清www | 在线免费观看涩涩 | 中文字幕一区二区三区在线观看 | 在线观看一区视频 | 国产精品久久久久久久久久久久午夜片 | 91夜夜夜 | 九九热1 | 国产婷婷精品av在线 | 91精彩视频 | 黄色特级毛片 | 欧美黑吊大战白妞欧美 | 精品少妇一区二区三区在线 | 狠狠干夜夜操天天爽 | 免费视频成人 | 久久一及片| 日韩在线视频看看 | 久久久麻豆视频 | 91福利视频免费观看 | 色噜噜在线观看视频 | 区一区二区三在线观看 | 国产精品不卡在线观看 | 婷婷激情综合 | 在线观看免费av网站 | 国产精品 亚洲精品 | 激情综合中文娱乐网 | 国产精品第2页 | 亚洲激情在线观看 | 免费在线观看毛片网站 | 久久精品国产第一区二区三区 | 中文字幕在线视频免费播放 | 国产精品久久久久毛片大屁完整版 | 在线亚洲欧美日韩 | 日韩欧美精品一区 | 亚洲国产97在线精品一区 | 久久综合五月天 | 国产拍揄自揄精品视频麻豆 | 一区二区三区四区在线 | 日韩三级免费 | 国产免费黄视频在线观看 | 中文永久字幕 | 日韩一二区在线 | 中国一级片在线播放 | 日韩中文字幕免费看 | 免费av片在线 | 久久国产露脸精品国产 | 中文字幕在线人 | 激情五月看片 | 亚洲天堂va | 欧美日韩1区| 蜜臀av性久久久久蜜臀aⅴ涩爱 | 国产精品系列在线 | 亚洲精品91天天久久人人 | 中文字幕 在线 一 二 | 色偷偷av男人天堂 | 一区二区三区电影大全 | 久久999久久 | 六月丁香久久 | 色噜噜狠狠色综合中国 | 久久激情视频 久久 | 999久久久国产精品 高清av免费观看 | 97国产精品亚洲精品 | 中文字幕色在线视频 | 一级久久久 | av网址最新 | 久久中国精品 | 国产一区二区三区黄 | 婷婷色综 | 亚洲成人第一区 | 免费观看全黄做爰大片国产 | 久久综合久久综合九色 | 婷婷丁香七月 | 91麻豆网| 国产亚洲精品久久久网站好莱 | 久热超碰 | 精品久久国产 | 久久夜色精品亚洲噜噜国4 午夜视频在线观看欧美 | 天天做天天爱夜夜爽 | 91理论片午午伦夜理片久久 | 91色在线观看视频 | 最近最新中文字幕 | 国产日韩在线播放 | 久久a免费视频 | av免费福利 | 国产精品理论片在线播放 | 91精品在线视频 | 亚洲激情综合 | 日日摸日日 | 国产麻豆精品一区 | 日韩在线免费看 | 亚洲91精品 | avav99| 天堂网中文在线 | 麻豆免费在线视频 | 色激情在线 | 激情视频亚洲 | av成人资源| 99久久超碰中文字幕伊人 | 91一区二区三区在线观看 | 久久久亚洲精华液 | 91麻豆传媒 | 国产成人精品av在线观 | 超碰在线99| 在线国产能看的 | 不卡av电影在线 | 久久66热这里只有精品 | 欧美日韩不卡一区 | 丁香婷婷激情 | 一级黄视频| 婷婷六月天丁香 | 国产视频在线观看一区 | 日日爱999 | 国产精品午夜久久久久久99热 | 欧美精品免费在线观看 | 又大又硬又黄又爽视频在线观看 | 一区二区三区四区五区在线视频 | 亚洲精品美女久久久久网站 | 91麻豆产精品久久久久久 | 日韩免费电影在线观看 | 黄色av影院 | 国产精品成人久久久久久久 | 最近日韩中文字幕中文 | 91成人黄色 | 偷拍精偷拍精品欧洲亚洲网站 | 国产亚洲免费的视频看 | 五月激情久久 | 久久99国产综合精品 | 一区二区三区不卡在线 | 久久久久久久久久电影 | 国产精品国产三级国产不产一地 | 人人舔人人舔 | 久久久久久久久毛片精品 | 美女免费黄视频网站 | 一色屋精品视频在线观看 | 中文字幕免费高清av | 亚洲欧美日韩中文在线 | 久久久精品 | av丝袜在线 | 手机看国产毛片 | 日韩在线免费播放 | 久久久久中文字幕 | 91高清完整版在线观看 | 欧美一区二区三区免费看 | 午夜精品一区二区三区在线 | 婷婷久久精品 | 国产精品久久久久av免费 | 高清av在线免费观看 | 日韩 在线a | 特级黄色视频毛片 | 国产成人精品999 | 久久首页 | 亚洲精品三级 | 91亚洲欧美激情 | 视频精品一区二区三区 | 久久国产精品99久久久久久老狼 | 国产99一区视频免费 | 国产伦精品一区二区三区无广告 | 亚洲香蕉视频 | 97视频免费在线观看 | 午夜视频一区二区三区 | 视频成人| 中文字幕第一页在线视频 | 97福利| 手机在线永久免费观看av片 | 在线日韩av | 91在线视频在线 | 久久久99精品免费观看 | 激情婷婷在线 | 麻豆视传媒官网免费观看 | 亚洲1区在线 | 激情欧美一区二区三区免费看 | 麻豆久久一区 | 日韩动漫免费观看高清完整版在线观看 | 国产剧情久久 | 日韩免费看片 | 久久亚洲电影 | 97精品国产一二三产区 | 久久人人爽| 亚洲精品美女在线 | 91热在线| 婷婷激情av | 国产无套精品久久久久久 | 99热手机在线 | 2023天天干 | 成人欧美一区二区三区黑人麻豆 | 国产色在线,com | 夜夜骑日日操 | 欧美日韩不卡在线视频 | 91视频免费网站 | 国产亚洲欧美日韩高清 | av九九九 | 久久草av | 日韩精品不卡在线观看 | 日日噜噜噜噜夜夜爽亚洲精品 | 日日夜夜国产 | 成人av高清在线 | 亚洲成人高清在线 | 欧美日比视频 | 日韩色爱| 一区二区三区免费 | 狠狠网站 | av电影不卡 | 91免费视频网站在线观看 | 中文字幕一区二区三区乱码不卡 | av激情五月 | 久久国产网 | 在线性视频日韩欧美 | 麻豆 videos| 亚洲 欧美 综合 在线 精品 | 久久国产美女视频 | 91尤物国产尤物福利在线播放 | 日日日日日 | 亚洲精品大片www | 成人一区二区在线 | 99久久精品免费一区 | 国产99在线免费 | 久久婷婷丁香 | 中文字幕有码在线播放 | 国产精品k频道 | 不卡av在线 | 日韩电影在线一区 | 国产精品久久久久久妇 | 中文字幕在线免费97 | 婷婷色 亚洲| 婷婷六月网 | 手机在线看永久av片免费 | 久久午夜鲁丝片 | 亚洲欧美日韩一二三区 | 免费视频久久久久久久 | 伊人亚洲综合网 | www.在线看片.com | 成人av电影在线观看 | av丝袜在线 | 精品国产乱码久久久久久浪潮 | 久久精品国产免费看久久精品 | 亚洲成a人片在线www | 中文字幕在线免费观看 | 月下香电影 | 91成人在线免费观看 | 在线a亚洲视频播放在线观看 | 免费观看黄色av | 日韩精品电影在线播放 | 99热超碰在线| 久久99精品久久久久久清纯直播 | 国产精品入口传媒 | 国产一二区在线观看 | 91麻豆精品国产91久久久无需广告 | 日韩欧美网站 | 国产精品美女久久久网av | 国产99久久久国产 | 久久久成人精品 | 日本中文字幕在线看 | 中文字幕视频观看 | 九九综合在线 | 色播六月天 | 国产免费观看视频 | 在线观看久久 | 一级黄色在线免费观看 | 国产亚洲一区 | 操久在线 | 91片在线观看 | 精品在线视频一区二区三区 | 精品久久久久久久久中文字幕 | 91精品国产九九九久久久亚洲 | 亚洲国产一区二区精品专区 | 五月激情天 | 免费在线国产 | 日韩网站一区 | 丁香 婷婷 激情 | 国产美女精品视频 | 热99久久精品 | 黄色特一级 | 中文字幕网站 | 国产91区| 一级黄色电影网站 | 婷婷色社区| 久久成人国产精品一区二区 | 久草在线视频网 | 精品久久1 | 亚洲欧洲精品在线 | 欧美色图亚洲图片 | 日韩精品一区二区电影 | 激情五月综合网 | 久久天天躁 | 久久精品视频3 | 日韩精品一区二区在线视频 | 久久久久久久久久久久亚洲 | 亚洲国产精品电影 | av免费电影在线观看 | 91视频免费看网站 | 有没有在线观看av | 国产精品久久久久久模特 | 97超碰成人在线 | av高清一区二区三区 | 麻豆网站免费观看 | 中文字幕免费看 | 国产原创中文在线 | 日韩高清av在线 | 久久国产电影院 | 99人久久精品视频最新地址 | 在线国产精品一区 | 欧美一二三专区 | 狠狠的操你 | 国产精品久久久久久久久久久久冷 | 日韩免费高清 | 日韩在线视频播放 | ww亚洲ww亚在线观看 | 国产一区二区三区高清播放 | 蜜臀av夜夜澡人人爽人人桃色 | 国产精品资源网 | 日韩一级黄色av | 在线观看国产一区二区 | 激情在线网址 | 国产精久久久久久妇女av | 婷婷深爱激情 | 午夜av色 | 正在播放一区 | 色婷婷av在线 | 911亚洲精品第一 | 亚洲午夜小视频 | 久久av一区二区三区亚洲 | 免费久久久 | 久久99精品久久久久婷婷 | 国产亚洲精品久久 | 51久久成人国产精品麻豆 | 久久伦理视频 | 婷婷伊人网 | 国产特级毛片aaaaaa毛片 | 久久成人午夜视频 | 天天天天干 | 色在线网站 | 日日夜夜精品免费 | 久久理论电影网 | 日韩高清无线码2023 | 毛片精品免费在线观看 | 欧美日本三级 | av丝袜在线 | 视频一区视频二区在线观看 | 国产精品1区2区3区在线观看 | 久草成人在线 | 日韩电影中文字幕在线 | 国内精品久久久久久久97牛牛 | 中文字幕资源网 | 五月婷久 | 在线播放国产一区二区三区 | av品善网 | 美女网站色 | av高清在线观看 | 久草精品在线观看 | 午夜av影院 | 亚洲男人天堂a | 国产精品自产拍在线观看 | 久久免费毛片视频 | 久久免费看 | 国产精品久久久视频 | 夜夜躁天天躁很躁波 | 在线中文字母电影观看 | 日本在线中文 | 国产午夜精品一区二区三区四区 | 99精品视频免费观看 | 亚洲丝袜一区 | 日韩最新av在线 | 伊人影院av| 午夜色大片在线观看 | 91免费看黄| 狠狠色丁香久久婷婷综合五月 | 国产亚洲人成网站在线观看 | 91精品中文字幕 | 在线观看视频h | 精品国偷自产国产一区 | 亚洲国产精品推荐 | 国产视频日韩视频欧美视频 | 国产一级黄 | 欧美国产一区二区 | 天堂av在线网址 | 精品福利视频在线 | 97超视频免费观看 | 亚洲老妇xxxxxx | wwwwww黄| 久久人人爽人人 | 天天射网 | 一区二区三区久久 | 欧美综合在线观看 | 久久精品亚洲国产 | 久久精品99国产国产 | 久草在线视频在线 | 永久免费毛片 | 国产日产亚洲精华av | 欧美a√大片| 国产精品久久一区二区三区, | 精品999在线观看 | 99国内精品久久久久久久 | 97日日碰人人模人人澡分享吧 | 国产老太婆免费交性大片 | 国产又粗又长又硬免费视频 | 亚洲片在线观看 | 国产无区一区二区三麻豆 | 久久视频免费 | 最近中文字幕免费观看 | 草草草影院 | 国产精品成人a免费观看 | 欧美一级片在线 | 国产第一页在线播放 | 亚洲婷婷在线视频 | 久久www免费视频 | 成人免费视频播放 | 成人蜜桃网 | 日韩免费专区 | 成人91在线观看 | 成人午夜电影在线播放 | 久久午夜国产精品 | 一区二区 不卡 | 国产精品毛片一区二区在线 | 久久久国产精华液 | 色在线免费| 激情综合网色播五月 | 91麻豆高清视频 | 免费欧美精品 | 超碰在线国产 | 久久久久久久久久久久久久电影 | h文在线观看免费 | 欧美一级免费高清 | 深夜国产在线 | 久久成人精品电影 | 伊人va | 九七视频在线 | 五月天久久久久 | 日日干美女 | 黄色一级动作片 | 国产精品18videosex性欧美 | 免费久久久久久 | 色婷婷激情电影 | www色网站 | 9在线观看免费高清完整版在线观看明 | 国内精品福利视频 | 成人黄色小说在线观看 | 91黄色影视 | 国产99久久久国产精品免费看 | 久久精品人| 国产拍在线 | 欧美日韩免费一区 | 五月天亚洲综合小说网 | 一区二区观看 | 91精品视频在线免费观看 | 国产精品成久久久久三级 | 国产一区二区三区高清播放 | 国产字幕在线看 | 六月丁香久久 | 亚洲一区欧美精品 | 亚洲香蕉视频 | 91精品成人 | 欧美孕交vivoestv另类 | a v在线视频| 狠狠色噜噜狠狠 | 欧美在线视频第一页 | 丁香婷婷综合五月 | 色多多污污在线观看 | 成人黄大片视频在线观看 | 亚洲黄色一级视频 | 久久久久久久久影院 | 夜夜夜夜猛噜噜噜噜噜初音未来 | a视频在线观看免费 | 亚洲成人一区 | 国产3p视频 | 国产精品视频不卡 | 激情图片区 | 夜夜躁日日躁狠狠躁 | 国产精品成人aaaaa网站 | 国产欧美精品在线观看 | 天天摸天天操天天舔 | 超碰公开在线 | 国产91丝袜在线播放动漫 | 91免费版在线观看 | 国产美女精品视频 | 免费看成人a | 国产喷水在线 | 九九九热精品免费视频观看 | 青青久草在线 | 超碰在线人人草 | 日日噜噜噜噜夜夜爽亚洲精品 | 99精品国产福利在线观看免费 | 日韩免费视频在线观看 | 99视频精品全部免费 在线 | 国产青草视频在线观看 | 香蕉在线观看 | 国产一区在线视频播放 | 国产精品v欧美精品 | 亚洲国产美女精品久久久久∴ | 91视频电影 | av在线免费在线观看 | 97高清视频| 中文字幕在线观看2018 | 久久久91精品国产一区二区精品 | 亚洲成人家庭影院 | 免费看色视频 | 久久午夜剧场 | 国产精品久久久久一区二区国产 | 香蕉97视频观看在线观看 | 国产精品第一页在线观看 | 国产精选在线观看 | 四虎国产精品免费观看视频优播 | 国产精品第 | 国产99精品 | 国产午夜精品福利视频 | 91视频一8mav| 91免费日韩| 久久66热这里只有精品 | 免费福利影院 | 国产三级久久久 | 成人久久久久 | 欧美黑人性爽 | 九九九九九精品 | 久久成人在线视频 | 免费在线观看不卡av | 精品免费视频 | 9免费视频 | 亚洲精品在线观看视频 | 日日婷婷夜日日天干 | 久久av中文字幕片 | 成年人三级网站 | 亚洲毛片久久 | 国产麻豆电影 | av一级在线 | 日本黄色免费电影网站 | 成人aⅴ视频 | 黄色成人在线观看 | 99精品免费网 | 成人毛片一区 | 在线观看免费高清视频大全追剧 | 亚洲国产精品一区二区尤物区 | www.黄色在线| 激情丁香5月 | 日韩三级在线 | 久久精彩 | 国产免费看 | 国产黄色免费在线观看 | 一区二区中文字幕在线 | 99精品视频免费看 | 国产精品久久久久久久久久久久午夜 | 韩国一区在线 | av在线永久免费观看 | 亚洲乱亚洲乱亚洲 | 中文字幕在线观看你懂的 | av爱干| 久久99深爱久久99精品 |