libmp3fixes

This commit is contained in:
shabinder 2021-09-02 16:33:28 +05:30
parent df6e969a56
commit 24c0002bb0
41 changed files with 824 additions and 1053 deletions

View File

@ -103,6 +103,7 @@ dependencies {
implementation(project(":common:data-models"))
implementation(project(":common:core-components"))
implementation(project(":common:providers"))
implementation(project(":ffmpeg:android-ffmpeg"))
// Koin
implementation(Koin.android)

View File

@ -77,6 +77,7 @@ import kotlinx.coroutines.flow.MutableSharedFlow
import kotlinx.coroutines.flow.conflate
import kotlinx.coroutines.flow.emitAll
import kotlinx.coroutines.launch
import nl.bravobit.ffmpeg.*
import org.koin.android.ext.android.inject
import org.koin.core.parameter.parametersOf
import java.io.File
@ -107,13 +108,37 @@ class MainActivity : ComponentActivity() {
// This app draws behind the system bars, so we want to handle fitting system windows
WindowCompat.setDecorFitsSystemWindows(window, false)
rootComponent = spotiFlyerRoot(defaultComponentContext())
val ffmpeg = FFmpeg.getInstance(this@MainActivity)
val ffprobe = FFprobe.getInstance(this@MainActivity)
lifecycleScope.launch {
Log.d("FFmpeg", "init")
AndroidMediaConverter().convertAudioFile("/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.mp3","/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.temp.mp3").fold({
Log.d("FFmpeg Success",it)
}){
it.printStackTrace()
}
FFmpegConfig.versionFFmpeg(this@MainActivity)
FFmpegConfig.codecsFFmpeg(this@MainActivity)
FFmpegConfig.versionFFprobe(this@MainActivity)
Log.d("FFmpeg Support", ffmpeg.isSupported.toString())
val inputFilePath = "/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.mp3"
val outputFilePath = "/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.temp.mp3"
val kbpsArg = "-b:a 192k"
ffmpeg.execute(arrayOf("-i", inputFilePath, /*"-acodec", "libmp3lame",*/ "-vn", outputFilePath),object : ExecuteBinaryResponseHandler() {
override fun onSuccess(message: String?) {
Log.d("FFmpeg Command", "Success $message")
}
override fun onProgress(message: String?) {
Log.d("FFmpeg Command", "Progress $message")
}
override fun onFailure(message: String?) {
Log.d("FFmpeg Command", "Failed $message")
}
})
/* AndroidMediaConverter().convertAudioFile("/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.mp3","/storage/emulated/0/Music/SpotiFlyer/Playlists/Sing-along_Punjabi/Kya_Baat_Ay.temp.mp3").fold({
Log.d("FFmpeg Success",it)
}){
it.printStackTrace()
}*/
}
/*FFmpeg.testInit()*/
setContent {

View File

@ -19,7 +19,7 @@ kotlin {
dependencies {
implementation(Extras.mp3agic)
implementation(Extras.Android.countly)
implementation(project(":ffmpeg:ffmpeg-kit-android-lib"))
implementation(project(":ffmpeg:android-ffmpeg"))
// implementation("com.arthenica:ffmpeg-kit-audio:4.4.LTS")
//api(files("$rootDir/libs/mobile-ffmpeg.aar"))
}

View File

@ -1,6 +1,5 @@
package com.shabinder.common.core_components.media_converter
import com.shabinder.spotiflyer.ffmpeg.AndroidFFmpeg.runTranscode
import com.shabinder.common.models.AudioQuality
import org.koin.dsl.bind
import org.koin.dsl.module
@ -14,8 +13,12 @@ class AndroidMediaConverter : MediaConverter() {
progressCallbacks: (Long) -> Unit,
) = executeSafelyInPool {
// 192 is Default
val audioBitrate = if (audioQuality == AudioQuality.UNKNOWN) 192 else audioQuality.kbps.toIntOrNull() ?: 192
runTranscode(inputFilePath,outputFilePath,audioBitrate).toString()
val audioBitrate =
if (audioQuality == AudioQuality.UNKNOWN) 192 else audioQuality.kbps.toIntOrNull()
?: 192
""
//runTranscode(inputFilePath,outputFilePath,audioBitrate).toString()
/*val kbpsArg = if (audioQuality == AudioQuality.UNKNOWN) {
val mediaInformation = FFprobeKit.getMediaInformation(inputFilePath)
val bitrate = ((mediaInformation.mediaInformation.bitrate).toFloat()/1000).roundToInt()

1
ffmpeg/android-ffmpeg/.gitignore vendored Normal file
View File

@ -0,0 +1 @@
/build

View File

@ -13,24 +13,12 @@ android {
minSdk = Versions.minSdkVersion
targetSdk = Versions.targetSdkVersion
// versionCode = Versions.versionCode
// versionName = Versions.versionName
/*versionCode = Versions.versionCode
versionName = Versions.versionName*/
ndk {
/*ndk {
abiFilters.addAll(setOf("x86", "x86_64", "armeabi-v7a", "arm64-v8a"))
}
}
sourceSets {
named("main") {
jniLibs.srcDir("../ffmpeg-android-maker/output/lib")
}
}
externalNativeBuild {
cmake {
path("CMakeLists.txt")
}
}*/
}
buildTypes {
@ -60,4 +48,5 @@ android {
}
}
dependencies { /**/ }
dependencies { /**/ }

View File

View File

@ -0,0 +1,21 @@
# Add project specific ProGuard rules here.
# You can control the set of applied configuration files using the
# proguardFiles setting in build.gradle.kts.
#
# For more details, see
# http://developer.android.com/guide/developing/tools/proguard.html
# If your project uses WebView with JS, uncomment the following
# and specify the fully qualified class name to the JavaScript interface
# class:
#-keepclassmembers class fqcn.of.javascript.interface.for.webview {
# public *;
#}
# Uncomment this to preserve the line number information for
# debugging stack traces.
#-keepattributes SourceFile,LineNumberTable
# If you keep the line number information, uncomment this to
# hide the original source file name.
#-renamesourcefileattribute SourceFile

View File

@ -0,0 +1 @@
<manifest package="bravobit.nl.ffmpegandroid" />

View File

@ -0,0 +1,30 @@
package nl.bravobit.ffmpeg;
class CommandResult {
final String output;
final boolean success;
CommandResult(boolean success, String output) {
this.success = success;
this.output = output;
}
static CommandResult getDummyFailureResponse() {
return new CommandResult(false, "");
}
static CommandResult getOutputFromProcess(Process process) {
String output;
if (success(process.exitValue())) {
output = Util.convertInputStreamToString(process.getInputStream());
} else {
output = Util.convertInputStreamToString(process.getErrorStream());
}
return new CommandResult(success(process.exitValue()), output);
}
static boolean success(Integer exitValue) {
return exitValue != null && exitValue == 0;
}
}

View File

@ -0,0 +1,29 @@
package nl.bravobit.ffmpeg;
public class ExecuteBinaryResponseHandler implements FFcommandExecuteResponseHandler {
@Override
public void onSuccess(String message) {
}
@Override
public void onProgress(String message) {
}
@Override
public void onFailure(String message) {
}
@Override
public void onStart() {
}
@Override
public void onFinish() {
}
}

View File

@ -0,0 +1,8 @@
package nl.bravobit.ffmpeg;
import android.content.Context;
public interface FFbinaryContextProvider {
Context provide();
}

View File

@ -0,0 +1,55 @@
package nl.bravobit.ffmpeg;
import java.util.Map;
interface FFbinaryInterface {
/**
* Executes a command
*
* @param environmentVars Environment variables
* @param cmd command to execute
* @param ffcommandExecuteResponseHandler {@link FFcommandExecuteResponseHandler}
* @return the task
*/
FFtask execute(Map<String, String> environmentVars, String[] cmd, FFcommandExecuteResponseHandler ffcommandExecuteResponseHandler);
/**
* Executes a command
*
* @param cmd command to execute
* @param ffcommandExecuteResponseHandler {@link FFcommandExecuteResponseHandler}
* @return the task
*/
FFtask execute(String[] cmd, FFcommandExecuteResponseHandler ffcommandExecuteResponseHandler);
/**
* Checks if FF binary is supported on this device
*
* @return true if FF binary is supported on this device
*/
boolean isSupported();
/**
* Checks if a command with given task is currently running
*
* @param task - the task that you want to check
* @return true if a command is running
*/
boolean isCommandRunning(FFtask task);
/**
* Kill given running process
*
* @param task - the task to kill
* @return true if process is killed successfully
*/
boolean killRunningProcesses(FFtask task);
/**
* Timeout for binary process, should be minimum of 10 seconds
*
* @param timeout in milliseconds
*/
void setTimeout(long timeout);
}

View File

@ -0,0 +1,6 @@
package nl.bravobit.ffmpeg;
public interface FFbinaryObserver extends Runnable {
void cancel();
}

View File

@ -0,0 +1,142 @@
package nl.bravobit.ffmpeg;
import android.os.AsyncTask;
import java.io.BufferedReader;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.OutputStream;
import java.util.Map;
import java.util.concurrent.TimeoutException;
class FFcommandExecuteAsyncTask extends AsyncTask<Void, String, CommandResult> implements FFtask {
private final String[] cmd;
private Map<String, String> environment;
private final FFcommandExecuteResponseHandler ffmpegExecuteResponseHandler;
private final ShellCommand shellCommand;
private final long timeout;
private long startTime;
private Process process;
private String output = "";
private boolean quitPending;
FFcommandExecuteAsyncTask(String[] cmd, Map<String, String> environment, long timeout, FFcommandExecuteResponseHandler ffmpegExecuteResponseHandler) {
this.cmd = cmd;
this.timeout = timeout;
this.environment = environment;
this.ffmpegExecuteResponseHandler = ffmpegExecuteResponseHandler;
this.shellCommand = new ShellCommand();
}
@Override
protected void onPreExecute() {
startTime = System.currentTimeMillis();
if (ffmpegExecuteResponseHandler != null) {
ffmpegExecuteResponseHandler.onStart();
}
}
@Override
protected CommandResult doInBackground(Void... params) {
try {
process = shellCommand.run(cmd, environment);
if (process == null) {
return CommandResult.getDummyFailureResponse();
}
Log.d("Running publishing updates method");
checkAndUpdateProcess();
return CommandResult.getOutputFromProcess(process);
} catch (TimeoutException e) {
Log.e("FFmpeg binary timed out", e);
return new CommandResult(false, e.getMessage());
} catch (Exception e) {
Log.e("Error running FFmpeg binary", e);
} finally {
Util.destroyProcess(process);
}
return CommandResult.getDummyFailureResponse();
}
@Override
protected void onProgressUpdate(String... values) {
if (values != null && values[0] != null && ffmpegExecuteResponseHandler != null) {
ffmpegExecuteResponseHandler.onProgress(values[0]);
}
}
@Override
protected void onPostExecute(CommandResult commandResult) {
if (ffmpegExecuteResponseHandler != null) {
output += commandResult.output;
if (commandResult.success) {
ffmpegExecuteResponseHandler.onSuccess(output);
} else {
ffmpegExecuteResponseHandler.onFailure(output);
}
ffmpegExecuteResponseHandler.onFinish();
}
}
private void checkAndUpdateProcess() throws TimeoutException, InterruptedException {
while (!Util.isProcessCompleted(process)) {
// checking if process is completed
if (Util.isProcessCompleted(process)) {
return;
}
// Handling timeout
if (timeout != Long.MAX_VALUE && System.currentTimeMillis() > startTime + timeout) {
throw new TimeoutException("FFmpeg binary timed out");
}
try {
String line;
BufferedReader reader = new BufferedReader(new InputStreamReader(process.getErrorStream()));
while ((line = reader.readLine()) != null) {
if (isCancelled()) {
process.destroy();
process.waitFor();
return;
}
if (quitPending) {
sendQ();
process = null;
return;
}
output += line + "\n";
publishProgress(line);
}
} catch (IOException e) {
e.printStackTrace();
}
}
}
public boolean isProcessCompleted() {
return Util.isProcessCompleted(process);
}
@Override
public boolean killRunningProcess() {
return Util.killAsync(this);
}
@Override
public void sendQuitSignal() {
quitPending = true;
}
private void sendQ() {
OutputStream outputStream = process.getOutputStream();
try {
outputStream.write("q\n".getBytes());
outputStream.flush();
} catch (IOException e) {
e.printStackTrace();
}
}
}

View File

@ -0,0 +1,26 @@
package nl.bravobit.ffmpeg;
public interface FFcommandExecuteResponseHandler extends ResponseHandler {
/**
* on Success
*
* @param message complete output of the binary command
*/
void onSuccess(String message);
/**
* on Progress
*
* @param message current output of binary command
*/
void onProgress(String message);
/**
* on Failure
*
* @param message complete output of the binary command
*/
void onFailure(String message);
}

View File

@ -0,0 +1,88 @@
package nl.bravobit.ffmpeg;
import android.content.Context;
import android.os.AsyncTask;
import java.io.File;
import java.util.Map;
public class FFmpeg implements FFbinaryInterface {
private final FFbinaryContextProvider context;
private static final long MINIMUM_TIMEOUT = 10 * 1000;
private long timeout = Long.MAX_VALUE;
private static FFmpeg instance = null;
private FFmpeg(FFbinaryContextProvider context) {
this.context = context;
Log.setDebug(Util.isDebug(this.context.provide()));
}
public static FFmpeg getInstance(final Context context) {
if (instance == null) {
instance = new FFmpeg(new FFbinaryContextProvider() {
@Override
public Context provide() {
return context;
}
});
}
return instance;
}
@Override
public boolean isSupported() {
// get ffmpeg file
File ffmpeg = FileUtils.getFFmpeg(context.provide());
// check if ffmpeg can be executed
if (!ffmpeg.canExecute()) {
// try to make executable
Log.e("ffmpeg cannot execute");
return false;
}
Log.d("ffmpeg is ready!");
return true;
}
@Override
public FFtask execute(Map<String, String> environvenmentVars, String[] cmd, FFcommandExecuteResponseHandler ffmpegExecuteResponseHandler) {
if (cmd.length != 0) {
final String[] command = new String[cmd.length + 1];
command[0] = FileUtils.getFFmpeg(context.provide()).getAbsolutePath();
System.arraycopy(cmd, 0, command, 1, cmd.length);
FFcommandExecuteAsyncTask task = new FFcommandExecuteAsyncTask(command, environvenmentVars, timeout, ffmpegExecuteResponseHandler);
task.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
return task;
} else {
throw new IllegalArgumentException("shell command cannot be empty");
}
}
@Override
public FFtask execute(String[] cmd, FFcommandExecuteResponseHandler ffmpegExecuteResponseHandler) {
return execute(null, cmd, ffmpegExecuteResponseHandler);
}
@Override
public boolean isCommandRunning(FFtask task) {
return task != null && !task.isProcessCompleted();
}
@Override
public boolean killRunningProcesses(FFtask task) {
return task != null && task.killRunningProcess();
}
@Override
public void setTimeout(long timeout) {
if (timeout >= MINIMUM_TIMEOUT) {
this.timeout = timeout;
}
}
}

View File

@ -0,0 +1,42 @@
package nl.bravobit.ffmpeg
import android.content.Context
object FFmpegConfig {
fun versionFFmpeg(context: Context) {
FFmpeg.getInstance(context).execute(arrayOf("-version"), object : ExecuteBinaryResponseHandler() {
override fun onSuccess(message: String) {
Log.d(message)
}
override fun onProgress(message: String) {
Log.d(message)
}
})
}
fun codecsFFmpeg(context: Context) {
FFmpeg.getInstance(context).execute(arrayOf("-codecs"), object : ExecuteBinaryResponseHandler() {
override fun onSuccess(message: String) {
Log.d(message)
}
override fun onProgress(message: String) {
Log.d(message)
}
})
}
fun versionFFprobe(context: Context) {
Log.d("version ffprobe")
FFprobe.getInstance(context).execute(arrayOf("-version"), object : ExecuteBinaryResponseHandler() {
override fun onSuccess(message: String) {
Log.d(message)
}
override fun onProgress(message: String) {
Log.d(message)
}
})
}
}

View File

@ -0,0 +1,85 @@
package nl.bravobit.ffmpeg;
import android.content.Context;
import android.os.AsyncTask;
import java.io.File;
import java.util.Map;
public class FFprobe implements FFbinaryInterface {
private final FFbinaryContextProvider context;
private static final long MINIMUM_TIMEOUT = 10 * 1000;
private long timeout = Long.MAX_VALUE;
private static FFprobe instance = null;
private FFprobe(FFbinaryContextProvider context) {
this.context = context;
Log.setDebug(Util.isDebug(this.context.provide()));
}
public static FFprobe getInstance(final Context context) {
if (instance == null) {
instance = new FFprobe(new FFbinaryContextProvider() {
@Override
public Context provide() {
return context;
}
});
}
return instance;
}
@Override
public boolean isSupported() {
// get ffprobe file
File ffprobe = FileUtils.getFFprobe(context.provide());
// check if ffprobe can be executed
if (!ffprobe.canExecute()) {
Log.e("ffprobe cannot execute");
return false;
}
Log.d("ffprobe is ready!");
return true;
}
@Override
public FFtask execute(Map<String, String> environvenmentVars, String[] cmd, FFcommandExecuteResponseHandler ffcommandExecuteResponseHandler) {
if (cmd.length != 0) {
final String[] command = new String[cmd.length + 1];
command[0] = FileUtils.getFFprobe(context.provide()).getAbsolutePath();
System.arraycopy(cmd, 0, command, 1, cmd.length);
FFcommandExecuteAsyncTask task = new FFcommandExecuteAsyncTask(command, environvenmentVars, timeout, ffcommandExecuteResponseHandler);
task.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
return task;
} else {
throw new IllegalArgumentException("shell command cannot be empty");
}
}
@Override
public FFtask execute(String[] cmd, FFcommandExecuteResponseHandler ffcommandExecuteResponseHandler) {
return execute(null, cmd, ffcommandExecuteResponseHandler);
}
public boolean isCommandRunning(FFtask task) {
return task != null && !task.isProcessCompleted();
}
@Override
public boolean killRunningProcesses(FFtask task) {
return task != null && task.killRunningProcess();
}
@Override
public void setTimeout(long timeout) {
if (timeout >= MINIMUM_TIMEOUT) {
this.timeout = timeout;
}
}
}

View File

@ -0,0 +1,22 @@
package nl.bravobit.ffmpeg;
public interface FFtask {
/**
* Sends 'q' to the ff binary running process asynchronously
*/
void sendQuitSignal();
/**
* Checks if process is completed
* @return <code>true</code> if a process is running
*/
boolean isProcessCompleted();
/**
* Kill given running process
*
* @return true if process is killed successfully
*/
boolean killRunningProcess();
}

View File

@ -0,0 +1,20 @@
package nl.bravobit.ffmpeg;
import android.content.Context;
import java.io.File;
class FileUtils {
private static final String FFMPEG_FILE_NAME = "ffmpeg";
private static final String FFPROBE_FILE_NAME = "ffprobe";
static File getFFmpeg(Context context) {
File folder = new File(context.getApplicationInfo().nativeLibraryDir);
return new File(folder, FFMPEG_FILE_NAME);
}
static File getFFprobe(Context context) {
File folder = new File(context.getApplicationInfo().nativeLibraryDir);
return new File(folder, FFPROBE_FILE_NAME);
}
}

View File

@ -0,0 +1,58 @@
package nl.bravobit.ffmpeg;
class Log {
private static String TAG = FFmpeg.class.getSimpleName();
private static boolean DEBUG = false;
public static void setDebug(boolean debug) {
Log.DEBUG = debug;
}
public static void setTag(String tag) {
Log.TAG = tag;
}
static void d(Object obj) {
if (DEBUG) {
android.util.Log.d(TAG, obj != null ? obj.toString() : "");
}
}
static void e(Object obj) {
if (DEBUG) {
android.util.Log.e(TAG, obj != null ? obj.toString() : "");
}
}
static void w(Object obj) {
if (DEBUG) {
android.util.Log.w(TAG, obj != null ? obj.toString() : "");
}
}
static void i(Object obj) {
if (DEBUG) {
android.util.Log.i(TAG, obj != null ? obj.toString() : "");
}
}
static void v(Object obj) {
if (DEBUG) {
android.util.Log.v(TAG, obj != null ? obj.toString() : "");
}
}
static void e(Object obj, Throwable throwable) {
if (DEBUG) {
android.util.Log.e(TAG, obj != null ? obj.toString() : "", throwable);
}
}
static void e(Throwable throwable) {
if (DEBUG) {
android.util.Log.e(TAG, "", throwable);
}
}
}

View File

@ -0,0 +1,15 @@
package nl.bravobit.ffmpeg;
public interface ResponseHandler {
/**
* on Start
*/
void onStart();
/**
* on Finish
*/
void onFinish();
}

View File

@ -0,0 +1,22 @@
package nl.bravobit.ffmpeg;
import java.util.Arrays;
import java.util.Map;
class ShellCommand {
Process run(String[] commandString, Map<String, String> environment) {
Process process = null;
try {
ProcessBuilder processBuilder = new ProcessBuilder(commandString);
if (environment != null) {
processBuilder.environment().putAll(environment);
}
process = processBuilder.start();
} catch (Throwable t) {
Log.e("Exception while trying to run: " + Arrays.toString(commandString), t);
}
return process;
}
}

View File

@ -0,0 +1,105 @@
package nl.bravobit.ffmpeg;
import android.content.Context;
import android.content.pm.ApplicationInfo;
import android.os.AsyncTask;
import java.io.BufferedReader;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
class Util {
static boolean isDebug(Context context) {
return (context.getApplicationContext().getApplicationInfo().flags & ApplicationInfo.FLAG_DEBUGGABLE) != 0;
}
static String convertInputStreamToString(InputStream inputStream) {
try {
BufferedReader r = new BufferedReader(new InputStreamReader(inputStream));
String str;
StringBuilder sb = new StringBuilder();
while ((str = r.readLine()) != null) {
sb.append(str);
}
return sb.toString();
} catch (IOException e) {
Log.e("error converting input stream to string", e);
}
return null;
}
static void destroyProcess(Process process) {
if (process != null) {
try {
process.destroy();
} catch (Exception e) {
Log.e("progress destroy error", e);
}
}
}
static boolean killAsync(AsyncTask asyncTask) {
return asyncTask != null && !asyncTask.isCancelled() && asyncTask.cancel(true);
}
static boolean isProcessCompleted(Process process) {
try {
if (process == null) return true;
process.exitValue();
return true;
} catch (IllegalThreadStateException e) {
// do nothing
}
return false;
}
public interface ObservePredicate {
Boolean isReadyToProceed();
}
static FFbinaryObserver observeOnce(final ObservePredicate predicate, final Runnable run, final int timeout) {
final android.os.Handler observer = new android.os.Handler();
final FFbinaryObserver observeAction = new FFbinaryObserver() {
private boolean canceled = false;
private int timeElapsed = 0;
@Override
public void run() {
if (timeElapsed + 40 > timeout) cancel();
timeElapsed += 40;
if (canceled) return;
boolean readyToProceed = false;
try {
readyToProceed = predicate.isReadyToProceed();
} catch (Exception e) {
Log.v("Observing " + e.getMessage());
observer.postDelayed(this, 40);
return;
}
if (readyToProceed) {
Log.v("Observed");
run.run();
} else {
Log.v("Observing");
observer.postDelayed(this, 40);
}
}
@Override
public void cancel() {
canceled = true;
}
};
observer.post(observeAction);
return observeAction;
}
}

View File

@ -0,0 +1,3 @@
<resources>
<string name="app_name">FFMpegAndroid</string>
</resources>

View File

@ -1,61 +0,0 @@
cmake_minimum_required(VERSION 3.4.1)
set(ffmpeg_dir ${CMAKE_SOURCE_DIR}/../ffmpeg-android-maker/output)
set(ffmpeg_libs ${ffmpeg_dir}/lib/${ANDROID_ABI})
include_directories(${ffmpeg_dir}/include/${ANDROID_ABI})
set(
# List variable name
ffmpeg_libs_names
# Values in the list
avutil avformat avcodec swresample avdevice avfilter swscale
)
foreach (ffmpeg_lib_name ${ffmpeg_libs_names})
add_library(
${ffmpeg_lib_name}
SHARED
IMPORTED
)
set_target_properties(
${ffmpeg_lib_name}
PROPERTIES
IMPORTED_LOCATION
${ffmpeg_libs}/lib${ffmpeg_lib_name}.so
)
endforeach ()
add_library(
# Name for a library to build
spotiflyer-ffmpeg
# Type of a library
SHARED
# All cpp files to compile
# mobile-ffmpeg
src/main/cpp/doc_examples_transcode_aac.c
# ffmpeg-kit
# src/main/cpp/ffmpegkit.c
# src/main/cpp/ffprobekit.c
# src/main/cpp/ffmpegkit_exception.c
# src/main/cpp/fftools_cmdutils.c
# src/main/cpp/fftools_ffmpeg.c
# src/main/cpp/fftools_ffprobe.c
# src/main/cpp/fftools_ffmpeg_opt.c
# src/main/cpp/fftools_ffmpeg_hw.c
# src/main/cpp/fftools_ffmpeg_filter.c
# src/main/cpp/saf_wrapper.c
)
target_link_libraries(
# Library to link
spotiflyer-ffmpeg
# List of libraries to link against:
# Library for writing messages in LogCat
log
# Library for processing Bitmap objects
jnigraphics
# FFmpeg libraries
${ffmpeg_libs_names}
)

View File

@ -1,17 +0,0 @@
# Add project specific ProGuard rules here.
# You can control the set of applied configuration files using the
# proguardFiles setting in build.gradle.kts.
#
# For more details, see
# http://developer.android.com/guide/developing/tools/proguard.html
-keep class com.arthenica.ffmpegkit.FFmpegKitConfig {
native <methods>;
void log(long, int, byte[]);
void statistics(long, int, float, float, long , int, double, double);
void closeParcelFileDescriptor(int);
}
-keep class com.arthenica.ffmpegkit.AbiDetect {
native <methods>;
}

View File

@ -1,4 +0,0 @@
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.shabinder.spotiflyer.ffmpeg">
</manifest>

View File

@ -1,2 +0,0 @@
/android_lts_support.o
/libandroidltssupport.a

View File

@ -1,915 +0,0 @@
/*
* Copyright (c) 2013-2018 Andreas Unterweger
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file
* Simple audio converter
*
* @example transcode_aac.c
* Convert an input audio file to AAC in an MP4 container using FFmpeg.
* Formats other than MP4 are supported based on the output file extension.
* @author Andreas Unterweger (dustsigns@gmail.com)
*/
#include <stdio.h>
#include <jni.h>
#include <android/log.h>
#include "libavformat/avformat.h"
#include "libavformat/avio.h"
#include "libavcodec/avcodec.h"
#include "libavutil/audio_fifo.h"
#include "libavutil/avassert.h"
#include "libavutil/avstring.h"
#include "libavutil/frame.h"
#include "libavutil/opt.h"
#include "libswresample/swresample.h"
#include <jni.h>
/* The number of output channels */
#define OUTPUT_CHANNELS 2
/* The index of audio stream that will be transcoded */
static int audio_stream_idx = -1;
/**
* Open an input file and the required decoder.
* @param filename File to be opened
* @param[out] input_format_context Format context of opened file
* @param[out] input_codec_context Codec context of opened file
* @return Error code (0 if successful)
*/
static int open_input_file(const char *filename,
AVFormatContext **input_format_context,
AVCodecContext **input_codec_context)
{
AVCodecContext *avctx;
AVCodec *input_codec;
int error;
/* Open the input file to read from it. */
if ((error = avformat_open_input(input_format_context, filename, NULL,
NULL)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open input file '%s' (error '%s')\n",
filename, av_err2str(error));
*input_format_context = NULL;
return error;
}
/* Get information on the input file (number of streams etc.). */
if ((error = avformat_find_stream_info(*input_format_context, NULL)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open find stream info (error '%s')\n",
av_err2str(error));
avformat_close_input(input_format_context);
return error;
}
for (audio_stream_idx = 0; audio_stream_idx < (*input_format_context)->nb_streams; audio_stream_idx++) {
if ((*input_format_context)->streams[audio_stream_idx]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO)
break;
__android_log_print(ANDROID_LOG_INFO, "transcode_aac", "Skip non-audio input stream %d\n", audio_stream_idx);
}
/* Make sure that there is at least one audio stream in the input file. */
if (audio_stream_idx >= (*input_format_context)->nb_streams) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not find an audio (error '%s')\n",
av_err2str(error));
avformat_close_input(input_format_context);
return AVERROR_EXIT;
}
/* Find a decoder for the audio stream. */
if (!(input_codec = avcodec_find_decoder((*input_format_context)->streams[audio_stream_idx]->codecpar->codec_id))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not find input codec\n");
avformat_close_input(input_format_context);
return AVERROR_EXIT;
}
/* Allocate a new decoding context. */
avctx = avcodec_alloc_context3(input_codec);
if (!avctx) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate a decoding context\n");
avformat_close_input(input_format_context);
return AVERROR(ENOMEM);
}
/* Initialize the stream parameters with demuxer information. */
error = avcodec_parameters_to_context(avctx, (*input_format_context)->streams[audio_stream_idx]->codecpar);
if (error < 0) {
avformat_close_input(input_format_context);
avcodec_free_context(&avctx);
return error;
}
/* Open the decoder for the audio stream to use it later. */
if ((error = avcodec_open2(avctx, input_codec, NULL)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open input codec (error '%s')\n",
av_err2str(error));
avcodec_free_context(&avctx);
avformat_close_input(input_format_context);
return error;
}
/* Save the decoder context for easier access later. */
*input_codec_context = avctx;
return 0;
}
/**
* Open an output file and the required encoder.
* Also set some basic encoder parameters.
* Some of these parameters are based on the input file's parameters.
* @param filename File to be opened
* @param input_codec_context Codec context of input file
* @param[out] output_format_context Format context of output file
* @param[out] output_codec_context Codec context of output file
* @return Error code (0 if successful)
*/
static int open_output_file(const char *filename,
AVCodecContext *input_codec_context,
AVFormatContext **output_format_context,
AVCodecContext **output_codec_context,
int audioBitrate
)
{
AVCodecContext *avctx = NULL;
AVIOContext *output_io_context = NULL;
AVStream *stream = NULL;
AVCodec *output_codec = NULL;
int error;
/* Open the output file to write to it. */
if ((error = avio_open(&output_io_context, filename,
AVIO_FLAG_WRITE)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open output file '%s' (error '%s')\n",
filename, av_err2str(error));
return error;
}
/* Create a new format context for the output container format. */
if (!(*output_format_context = avformat_alloc_context())) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate output format context\n");
return AVERROR(ENOMEM);
}
/* Associate the output file (pointer) with the container format context. */
(*output_format_context)->pb = output_io_context;
/* Guess the desired container format based on the file extension. */
if (!((*output_format_context)->oformat = av_guess_format(NULL, filename,
NULL))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not find output file format\n");
goto cleanup;
}
if (!((*output_format_context)->url = av_strdup(filename))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate url.\n");
error = AVERROR(ENOMEM);
goto cleanup;
}
/* Find the encoder to be used by its name. */
if (!(output_codec = avcodec_find_encoder((*output_format_context)->oformat->audio_codec))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not find an encoder for %s(%d).\n",
(*output_format_context)->oformat->long_name,
(*output_format_context)->oformat->audio_codec);
goto cleanup;
}
/* Create a new audio stream in the output file container. */
if (!(stream = avformat_new_stream(*output_format_context, NULL))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not create new stream\n");
error = AVERROR(ENOMEM);
goto cleanup;
}
avctx = avcodec_alloc_context3(output_codec);
if (!avctx) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate an encoding context\n");
error = AVERROR(ENOMEM);
goto cleanup;
}
/* Set the basic encoder parameters.
* The input file's sample rate is used to avoid a sample rate conversion. */
avctx->channels = OUTPUT_CHANNELS;
avctx->channel_layout = av_get_default_channel_layout(OUTPUT_CHANNELS);
avctx->sample_rate = input_codec_context->sample_rate;
avctx->sample_fmt = output_codec->sample_fmts[0];
avctx->bit_rate = audioBitrate;
/* Allow the use of the experimental AAC encoder. */
avctx->strict_std_compliance = FF_COMPLIANCE_EXPERIMENTAL;
/* Set the sample rate for the container. */
stream->time_base.den = input_codec_context->sample_rate;
stream->time_base.num = 1;
/* Some container formats (like MP4) require global headers to be present.
* Mark the encoder so that it behaves accordingly. */
if ((*output_format_context)->oformat->flags & AVFMT_GLOBALHEADER)
avctx->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
/* Open the encoder for the audio stream to use it later. */
if ((error = avcodec_open2(avctx, output_codec, NULL)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open output codec (error '%s')\n",
av_err2str(error));
goto cleanup;
}
error = avcodec_parameters_from_context(stream->codecpar, avctx);
if (error < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not initialize stream parameters\n");
goto cleanup;
}
/* Save the encoder context for easier access later. */
*output_codec_context = avctx;
return 0;
cleanup:
avcodec_free_context(&avctx);
avio_closep(&(*output_format_context)->pb);
avformat_free_context(*output_format_context);
*output_format_context = NULL;
return error < 0 ? error : AVERROR_EXIT;
}
/**
* Initialize one data packet for reading or writing.
* @param packet Packet to be initialized
*/
static void init_packet(AVPacket *packet)
{
av_init_packet(packet);
/* Set the packet data and size so that it is recognized as being empty. */
packet->data = NULL;
packet->size = 0;
}
/**
* Initialize one audio frame for reading from the input file.
* @param[out] frame Frame to be initialized
* @return Error code (0 if successful)
*/
static int init_input_frame(AVFrame **frame)
{
if (!(*frame = av_frame_alloc())) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate input frame\n");
return AVERROR(ENOMEM);
}
return 0;
}
/**
* Initialize the audio resampler based on the input and output codec settings.
* If the input and output sample formats differ, a conversion is required
* libswresample takes care of this, but requires initialization.
* @param input_codec_context Codec context of the input file
* @param output_codec_context Codec context of the output file
* @param[out] resample_context Resample context for the required conversion
* @return Error code (0 if successful)
*/
static int init_resampler(AVCodecContext *input_codec_context,
AVCodecContext *output_codec_context,
SwrContext **resample_context)
{
int error;
/*
* Create a resampler context for the conversion.
* Set the conversion parameters.
* Default channel layouts based on the number of channels
* are assumed for simplicity (they are sometimes not detected
* properly by the demuxer and/or decoder).
*/
*resample_context = swr_alloc_set_opts(NULL,
av_get_default_channel_layout(output_codec_context->channels),
output_codec_context->sample_fmt,
output_codec_context->sample_rate,
av_get_default_channel_layout(input_codec_context->channels),
input_codec_context->sample_fmt,
input_codec_context->sample_rate,
0, NULL);
if (!*resample_context) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate resample context\n");
return AVERROR(ENOMEM);
}
/*
* Perform a sanity check so that the number of converted samples is
* not greater than the number of samples to be converted.
* If the sample rates differ, this case has to be handled differently
*/
av_assert0(output_codec_context->sample_rate == input_codec_context->sample_rate);
/* Open the resampler with the specified parameters. */
if ((error = swr_init(*resample_context)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not open resample context\n");
swr_free(resample_context);
return error;
}
return 0;
}
/**
* Initialize a FIFO buffer for the audio samples to be encoded.
* @param[out] fifo Sample buffer
* @param output_codec_context Codec context of the output file
* @return Error code (0 if successful)
*/
static int init_fifo(AVAudioFifo **fifo, AVCodecContext *output_codec_context)
{
/* Create the FIFO buffer based on the specified output sample format. */
if (!(*fifo = av_audio_fifo_alloc(output_codec_context->sample_fmt,
output_codec_context->channels, 1))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate FIFO\n");
return AVERROR(ENOMEM);
}
return 0;
}
/**
* Write the header of the output file container.
* @param output_format_context Format context of the output file
* @return Error code (0 if successful)
*/
static int write_output_file_header(AVFormatContext *output_format_context)
{
int error;
if ((error = avformat_write_header(output_format_context, NULL)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not write output file header (error '%s')\n",
av_err2str(error));
return error;
}
return 0;
}
/**
* Decode one audio frame from the input file.
* @param frame Audio frame to be decoded
* @param input_format_context Format context of the input file
* @param input_codec_context Codec context of the input file
* @param[out] data_present Indicates whether data has been decoded
* @param[out] finished Indicates whether the end of file has
* been reached and all data has been
* decoded. If this flag is false, there
* is more data to be decoded, i.e., this
* function has to be called again.
* @return Error code (0 if successful)
*/
static int decode_audio_frame(AVFrame *frame,
AVFormatContext *input_format_context,
AVCodecContext *input_codec_context,
int *data_present, int *finished)
{
/* Packet used for temporary storage. */
AVPacket input_packet;
int error;
init_packet(&input_packet);
/* Read one audio frame from the input file into a temporary packet. */
if ((error = av_read_frame(input_format_context, &input_packet)) < 0) {
/* If we are at the end of the file, flush the decoder below. */
if (error == AVERROR_EOF)
*finished = 1;
else {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not read frame (error '%s')\n",
av_err2str(error));
return error;
}
}
if (error != AVERROR_EOF && input_packet.stream_index != audio_stream_idx) {
goto cleanup;
}
/* Send the audio frame stored in the temporary packet to the decoder.
* The input audio stream decoder is used to do this. */
if ((error = avcodec_send_packet(input_codec_context, &input_packet)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not send packet for decoding (error '%s')\n",
av_err2str(error));
return error;
}
/* Receive one frame from the decoder. */
error = avcodec_receive_frame(input_codec_context, frame);
/* If the decoder asks for more data to be able to decode a frame,
* return indicating that no data is present. */
if (error == AVERROR(EAGAIN)) {
error = 0;
goto cleanup;
/* If the end of the input file is reached, stop decoding. */
} else if (error == AVERROR_EOF) {
*finished = 1;
error = 0;
goto cleanup;
} else if (error < 0) {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not decode frame (error '%s')\n",
av_err2str(error));
goto cleanup;
/* Default case: Return decoded data. */
} else {
*data_present = 1;
goto cleanup;
}
cleanup:
av_packet_unref(&input_packet);
return error;
}
/**
* Initialize a temporary storage for the specified number of audio samples.
* The conversion requires temporary storage due to the different format.
* The number of audio samples to be allocated is specified in frame_size.
* @param[out] converted_input_samples Array of converted samples. The
* dimensions are reference, channel
* (for multi-channel audio), sample.
* @param output_codec_context Codec context of the output file
* @param frame_size Number of samples to be converted in
* each round
* @return Error code (0 if successful)
*/
static int init_converted_samples(uint8_t ***converted_input_samples,
AVCodecContext *output_codec_context,
int frame_size)
{
int error;
/* Allocate as many pointers as there are audio channels.
* Each pointer will later point to the audio samples of the corresponding
* channels (although it may be NULL for interleaved formats).
*/
if (!(*converted_input_samples = calloc(output_codec_context->channels,
sizeof(**converted_input_samples)))) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not allocate converted input sample pointers\n");
return AVERROR(ENOMEM);
}
/* Allocate memory for the samples of all channels in one consecutive
* block for convenience. */
if ((error = av_samples_alloc(*converted_input_samples, NULL,
output_codec_context->channels,
frame_size,
output_codec_context->sample_fmt, 0)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac",
"Could not allocate converted input samples (error '%s')\n",
av_err2str(error));
av_freep(&(*converted_input_samples)[0]);
free(*converted_input_samples);
return error;
}
return 0;
}
/**
* Convert the input audio samples into the output sample format.
* The conversion happens on a per-frame basis, the size of which is
* specified by frame_size.
* @param input_data Samples to be decoded. The dimensions are
* channel (for multi-channel audio), sample.
* @param[out] converted_data Converted samples. The dimensions are channel
* (for multi-channel audio), sample.
* @param frame_size Number of samples to be converted
* @param resample_context Resample context for the conversion
* @return Error code (0 if successful)
*/
static int convert_samples(const uint8_t **input_data,
uint8_t **converted_data, const int frame_size,
SwrContext *resample_context)
{
int error;
/* Convert the samples using the resampler. */
if ((error = swr_convert(resample_context,
converted_data, frame_size,
input_data , frame_size)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not convert input samples (error '%s')\n",
av_err2str(error));
return error;
}
return 0;
}
/**
* Add converted input audio samples to the FIFO buffer for later processing.
* @param fifo Buffer to add the samples to
* @param converted_input_samples Samples to be added. The dimensions are channel
* (for multi-channel audio), sample.
* @param frame_size Number of samples to be converted
* @return Error code (0 if successful)
*/
static int add_samples_to_fifo(AVAudioFifo *fifo,
uint8_t **converted_input_samples,
const int frame_size)
{
int error;
/* Make the FIFO as large as it needs to be to hold both,
* the old and the new samples. */
if ((error = av_audio_fifo_realloc(fifo, av_audio_fifo_size(fifo) + frame_size)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not reallocate FIFO\n");
return error;
}
/* Store the new samples in the FIFO buffer. */
if (av_audio_fifo_write(fifo, (void **)converted_input_samples,
frame_size) < frame_size) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not write data to FIFO\n");
return AVERROR_EXIT;
}
return 0;
}
/**
* Read one audio frame from the input file, decode, convert and store
* it in the FIFO buffer.
* @param fifo Buffer used for temporary storage
* @param input_format_context Format context of the input file
* @param input_codec_context Codec context of the input file
* @param output_codec_context Codec context of the output file
* @param resampler_context Resample context for the conversion
* @param[out] finished Indicates whether the end of file has
* been reached and all data has been
* decoded. If this flag is false,
* there is more data to be decoded,
* i.e., this function has to be called
* again.
* @return Error code (0 if successful)
*/
static int read_decode_convert_and_store(AVAudioFifo *fifo,
AVFormatContext *input_format_context,
AVCodecContext *input_codec_context,
AVCodecContext *output_codec_context,
SwrContext *resampler_context,
int *finished)
{
/* Temporary storage of the input samples of the frame read from the file. */
AVFrame *input_frame = NULL;
/* Temporary storage for the converted input samples. */
uint8_t **converted_input_samples = NULL;
int data_present = 0;
int ret = AVERROR_EXIT;
/* Initialize temporary storage for one input frame. */
if (init_input_frame(&input_frame))
goto cleanup;
/* Decode one frame worth of audio samples. */
if (decode_audio_frame(input_frame, input_format_context,
input_codec_context, &data_present, finished))
goto cleanup;
/* If we are at the end of the file and there are no more samples
* in the decoder which are delayed, we are actually finished.
* This must not be treated as an error. */
if (*finished) {
ret = 0;
goto cleanup;
}
/* If there is decoded data, convert and store it. */
if (data_present) {
/* Initialize the temporary storage for the converted input samples. */
if (init_converted_samples(&converted_input_samples, output_codec_context,
input_frame->nb_samples))
goto cleanup;
/* Convert the input samples to the desired output sample format.
* This requires a temporary storage provided by converted_input_samples. */
if (convert_samples((const uint8_t**)input_frame->extended_data, converted_input_samples,
input_frame->nb_samples, resampler_context))
goto cleanup;
/* Add the converted input samples to the FIFO buffer for later processing. */
if (add_samples_to_fifo(fifo, converted_input_samples,
input_frame->nb_samples))
goto cleanup;
ret = 0;
}
ret = 0;
cleanup:
if (converted_input_samples) {
av_freep(&converted_input_samples[0]);
free(converted_input_samples);
}
av_frame_free(&input_frame);
return ret;
}
/**
* Initialize one input frame for writing to the output file.
* The frame will be exactly frame_size samples large.
* @param[out] frame Frame to be initialized
* @param output_codec_context Codec context of the output file
* @param frame_size Size of the frame
* @return Error code (0 if successful)
*/
static int init_output_frame(AVFrame **frame,
AVCodecContext *output_codec_context,
int frame_size)
{
int error;
/* Create a new frame to store the audio samples. */
if (!(*frame = av_frame_alloc())) {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not allocate output frame\n");
return AVERROR_EXIT;
}
/* Set the frame's parameters, especially its size and format.
* av_frame_get_buffer needs this to allocate memory for the
* audio samples of the frame.
* Default channel layouts based on the number of channels
* are assumed for simplicity. */
(*frame)->nb_samples = frame_size;
(*frame)->channel_layout = output_codec_context->channel_layout;
(*frame)->format = output_codec_context->sample_fmt;
(*frame)->sample_rate = output_codec_context->sample_rate;
/* Allocate the samples of the created frame. This call will make
* sure that the audio frame can hold as many samples as specified. */
if ((error = av_frame_get_buffer(*frame, 0)) < 0) {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not allocate output frame samples (error '%s')\n",
av_err2str(error));
av_frame_free(frame);
return error;
}
return 0;
}
/* Global timestamp for the audio frames. */
static int64_t pts = 0;
/**
* Encode one frame worth of audio to the output file.
* @param frame Samples to be encoded
* @param output_format_context Format context of the output file
* @param output_codec_context Codec context of the output file
* @param[out] data_present Indicates whether data has been
* encoded
* @return Error code (0 if successful)
*/
static int encode_audio_frame(AVFrame *frame,
AVFormatContext *output_format_context,
AVCodecContext *output_codec_context,
int *data_present)
{
/* Packet used for temporary storage. */
AVPacket output_packet;
int error;
init_packet(&output_packet);
/* Set a timestamp based on the sample rate for the container. */
if (frame) {
frame->pts = pts;
pts += frame->nb_samples;
}
/* Send the audio frame stored in the temporary packet to the encoder.
* The output audio stream encoder is used to do this. */
error = avcodec_send_frame(output_codec_context, frame);
/* The encoder signals that it has nothing more to encode. */
if (error == AVERROR_EOF) {
error = 0;
goto cleanup;
} else if (error < 0) {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not send packet for encoding (error '%s')\n",
av_err2str(error));
return error;
}
/* Receive one encoded frame from the encoder. */
error = avcodec_receive_packet(output_codec_context, &output_packet);
/* If the encoder asks for more data to be able to provide an
* encoded frame, return indicating that no data is present. */
if (error == AVERROR(EAGAIN)) {
error = 0;
goto cleanup;
/* If the last frame has been encoded, stop encoding. */
} else if (error == AVERROR_EOF) {
error = 0;
goto cleanup;
} else if (error < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not encode frame (error '%s')\n",
av_err2str(error));
goto cleanup;
/* Default case: Return encoded data. */
} else {
*data_present = 1;
}
/* Write one audio frame from the temporary packet to the output file. */
if (*data_present &&
(error = av_write_frame(output_format_context, &output_packet)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not write frame (error '%s')\n",
av_err2str(error));
goto cleanup;
}
cleanup:
av_packet_unref(&output_packet);
return error;
}
/**
* Load one audio frame from the FIFO buffer, encode and write it to the
* output file.
* @param fifo Buffer used for temporary storage
* @param output_format_context Format context of the output file
* @param output_codec_context Codec context of the output file
* @return Error code (0 if successful)
*/
static int load_encode_and_write(AVAudioFifo *fifo,
AVFormatContext *output_format_context,
AVCodecContext *output_codec_context)
{
/* Temporary storage of the output samples of the frame written to the file. */
AVFrame *output_frame;
/* Use the maximum number of possible samples per frame.
* If there is less than the maximum possible frame size in the FIFO
* buffer use this number. Otherwise, use the maximum possible frame size. */
const int frame_size = FFMIN(av_audio_fifo_size(fifo),
output_codec_context->frame_size);
int data_written;
/* Initialize temporary storage for one output frame. */
if (init_output_frame(&output_frame, output_codec_context, frame_size))
return AVERROR_EXIT;
/* Read as many samples from the FIFO buffer as required to fill the frame.
* The samples are stored in the frame temporarily. */
if (av_audio_fifo_read(fifo, (void **)output_frame->data, frame_size) < frame_size) {
__android_log_print(ANDROID_LOG_WARN, "transcode_aac", "Could not read data from FIFO\n");
av_frame_free(&output_frame);
return AVERROR_EXIT;
}
/* Encode one frame worth of audio samples. */
if (encode_audio_frame(output_frame, output_format_context,
output_codec_context, &data_written)) {
av_frame_free(&output_frame);
return AVERROR_EXIT;
}
av_frame_free(&output_frame);
return 0;
}
/**
* Write the trailer of the output file container.
* @param output_format_context Format context of the output file
* @return Error code (0 if successful)
*/
static int write_output_file_trailer(AVFormatContext *output_format_context)
{
int error;
if ((error = av_write_trailer(output_format_context)) < 0) {
__android_log_print(ANDROID_LOG_ERROR, "transcode_aac", "Could not write output file trailer (error '%s')\n",
av_err2str(error));
return error;
}
return 0;
}
JNIEXPORT jint JNICALL Java_com_shabinder_spotiflyer_ffmpeg_AndroidFFmpeg_runTranscode(
JNIEnv *env, jobject c,
jstring inFilename,
jstring outFilename,
jint audioBitrate
) {
AVFormatContext *input_format_context = NULL, *output_format_context = NULL;
AVCodecContext *input_codec_context = NULL, *output_codec_context = NULL;
SwrContext *resample_context = NULL;
AVAudioFifo *fifo = NULL;
int ret = AVERROR_EXIT;
const char *in_filename = (*env)->GetStringUTFChars(env, inFilename, 0);
const char *out_filename = (*env)->GetStringUTFChars(env, outFilename, 0);
__android_log_print(ANDROID_LOG_INFO, "transcode_aac", "Bitrate:%d :: %s -> %s\n", audioBitrate, in_filename, out_filename);
/* Open the input file for reading. */
if (open_input_file(in_filename, &input_format_context,
&input_codec_context))
goto cleanup;
__android_log_print(ANDROID_LOG_INFO, "transcode_aac", "Input format: %s.\n",
input_format_context->iformat->long_name);
/* Open the output file for writing. */
if (open_output_file(out_filename, input_codec_context,
&output_format_context, &output_codec_context, (audioBitrate*1000)))
goto cleanup;
__android_log_print(ANDROID_LOG_INFO, "transcode_aac", "Output format: %s.\n",
output_format_context->oformat->long_name);
/* Initialize the resampler to be able to convert audio sample formats. */
if (init_resampler(input_codec_context, output_codec_context,
&resample_context))
goto cleanup;
/* Initialize the FIFO buffer to store audio samples to be encoded. */
if (init_fifo(&fifo, output_codec_context))
goto cleanup;
/* Write the header of the output file container. */
if (write_output_file_header(output_format_context))
goto cleanup;
/* Loop as long as we have input samples to read or output samples
* to write; abort as soon as we have neither. */
while (1) {
/* Use the encoder's desired frame size for processing. */
const int output_frame_size = output_codec_context->frame_size;
int finished = 0;
/* Make sure that there is one frame worth of samples in the FIFO
* buffer so that the encoder can do its work.
* Since the decoder's and the encoder's frame size may differ, we
* need to FIFO buffer to store as many frames worth of input samples
* that they make up at least one frame worth of output samples. */
while (av_audio_fifo_size(fifo) < output_frame_size) {
/* Decode one frame worth of audio samples, convert it to the
* output sample format and put it into the FIFO buffer. */
if (read_decode_convert_and_store(fifo, input_format_context,
input_codec_context,
output_codec_context,
resample_context, &finished))
goto cleanup;
/* If we are at the end of the input file, we continue
* encoding the remaining audio samples to the output file. */
if (finished)
break;
}
/* If we have enough samples for the encoder, we encode them.
* At the end of the file, we pass the remaining samples to
* the encoder. */
while (av_audio_fifo_size(fifo) >= output_frame_size ||
(finished && av_audio_fifo_size(fifo) > 0))
/* Take one frame worth of audio samples from the FIFO buffer,
* encode it and write it to the output file. */
if (load_encode_and_write(fifo, output_format_context,
output_codec_context))
goto cleanup;
/* If we are at the end of the input file and have encoded
* all remaining samples, we can exit this loop and finish. */
if (finished) {
int data_written;
/* Flush the encoder as it may have delayed frames. */
do {
data_written = 0;
if (encode_audio_frame(NULL, output_format_context,
output_codec_context, &data_written))
goto cleanup;
} while (data_written);
break;
}
}
/* Write the trailer of the output file container. */
if (write_output_file_trailer(output_format_context))
goto cleanup;
ret = 0;
cleanup:
if (fifo)
av_audio_fifo_free(fifo);
swr_free(&resample_context);
if (output_codec_context)
avcodec_free_context(&output_codec_context);
if (output_format_context) {
avio_closep(&output_format_context->pb);
avformat_free_context(output_format_context);
}
if (input_codec_context)
avcodec_free_context(&input_codec_context);
if (input_format_context)
avformat_close_input(&input_format_context);
return ret;
}

View File

@ -1,27 +0,0 @@
package com.shabinder.spotiflyer.ffmpeg
import android.util.Log
object AndroidFFmpeg {
/**
*
* Run transcode_aac from doc/examples.
*
* @return zero if transcoding was successful
*/
@JvmStatic
external fun runTranscode(inFilename: String?, outFilename: String?, audioBitrate: Int): Int
init {
Log.i("FFmpeg", "Loading mobile-ffmpeg.")
System.loadLibrary("avutil")
System.loadLibrary("swscale")
System.loadLibrary("swresample")
System.loadLibrary("avcodec")
System.loadLibrary("avformat")
System.loadLibrary("avfilter")
System.loadLibrary("avdevice")
//System.loadLibrary("avresample")
System.loadLibrary("spotiflyer-ffmpeg")
}
}

View File

@ -27,7 +27,7 @@ include(
":common:providers",
":common:core-components",
":common:dependency-injection",
":ffmpeg:ffmpeg-kit-android-lib",
":ffmpeg:android-ffmpeg",
":android",
":desktop",
":web-app",