mirror of https://github.com/mgba-emu/mgba.git
FFmpeg resampling
This commit is contained in:
parent
281f190ae6
commit
e9b26dda08
|
@ -81,7 +81,7 @@ add_definitions(-DBINARY_NAME="${BINARY_NAME}" -DPROJECT_NAME="${PROJECT_NAME}"
|
|||
|
||||
# Feature dependencies
|
||||
find_feature(USE_CLI_DEBUGGER "libedit")
|
||||
find_feature(USE_FFMPEG "libavcodec;libavformat;libavutil")
|
||||
find_feature(USE_FFMPEG "libavcodec;libavformat;libavresample;libavutil")
|
||||
find_feature(USE_PNG "ZLIB;PNG")
|
||||
find_feature(USE_LIBZIP "libzip")
|
||||
|
||||
|
@ -134,10 +134,10 @@ source_group("ARM debugger" FILES ${DEBUGGER_SRC})
|
|||
|
||||
if(USE_FFMPEG)
|
||||
add_definitions(-DUSE_FFMPEG)
|
||||
include_directories(AFTER ${LIBAVCODEC_INCLUDE_DIRS} ${LIBAVFORMAT_INCLUDE_DIRS} ${LIBAVUTIL_INCLUDE_DIRS})
|
||||
link_directories(${LIBAVCODEC_LIBRARY_DIRS} ${LIBAVFORMAT_LIBRARY_DIRS} ${LIBAVUTIL_LIBRARY_DIRS})
|
||||
include_directories(AFTER ${LIBAVCODEC_INCLUDE_DIRS} ${LIBAVFORMAT_INCLUDE_DIRS} ${LIBAVRESAMPLE_INCLUDE_DIRS} ${LIBAVUTIL_INCLUDE_DIRS})
|
||||
link_directories(${LIBAVCODEC_LIBRARY_DIRS} ${LIBAVFORMAT_LIBRARY_DIRS} ${LIBAVRESAMPLE_LIBRARY_DIRS} ${LIBAVUTIL_LIBRARY_DIRS})
|
||||
list(APPEND UTIL_SRC "${CMAKE_SOURCE_DIR}/src/platform/ffmpeg/ffmpeg-encoder.c")
|
||||
list(APPEND DEPENDENCY_LIB ${LIBAVCODEC_LIBRARIES} ${LIBAVFORMAT_LIBRARIES} ${LIBAVUTIL_LIBRARIES})
|
||||
list(APPEND DEPENDENCY_LIB ${LIBAVCODEC_LIBRARIES} ${LIBAVFORMAT_LIBRARIES} ${LIBAVRESAMPLE_LIBRARIES} ${LIBAVUTIL_LIBRARIES})
|
||||
endif()
|
||||
|
||||
if(USE_PNG)
|
||||
|
|
|
@ -3,10 +3,15 @@
|
|||
#include "gba-video.h"
|
||||
|
||||
#include <libavutil/imgutils.h>
|
||||
#include <libavutil/opt.h>
|
||||
|
||||
static void _ffmpegPostVideoFrame(struct GBAAVStream*, struct GBAVideoRenderer* renderer);
|
||||
static void _ffmpegPostAudioFrame(struct GBAAVStream*, int32_t left, int32_t right);
|
||||
|
||||
enum {
|
||||
PREFERRED_SAMPLE_RATE = 0x8000
|
||||
};
|
||||
|
||||
void FFmpegEncoderInit(struct FFmpegEncoder* encoder) {
|
||||
av_register_all();
|
||||
|
||||
|
@ -26,29 +31,70 @@ void FFmpegEncoderInit(struct FFmpegEncoder* encoder) {
|
|||
}
|
||||
|
||||
bool FFmpegEncoderSetAudio(struct FFmpegEncoder* encoder, const char* acodec, unsigned abr) {
|
||||
if (!avcodec_find_encoder_by_name(acodec)) {
|
||||
AVCodec* codec = avcodec_find_encoder_by_name(acodec);
|
||||
if (!codec) {
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!codec->sample_fmts) {
|
||||
return false;
|
||||
}
|
||||
size_t i;
|
||||
encoder->sampleFormat = AV_SAMPLE_FMT_NONE;
|
||||
for (i = 0; codec->sample_fmts[i] != AV_SAMPLE_FMT_NONE; ++i) {
|
||||
if (codec->sample_fmts[i] == AV_SAMPLE_FMT_S16 || codec->sample_fmts[i] == AV_SAMPLE_FMT_S16P) {
|
||||
encoder->sampleFormat = codec->sample_fmts[i];
|
||||
break;
|
||||
}
|
||||
}
|
||||
if (encoder->sampleFormat == AV_SAMPLE_FMT_NONE) {
|
||||
return false;
|
||||
}
|
||||
encoder->sampleRate = PREFERRED_SAMPLE_RATE;
|
||||
if (codec->supported_samplerates) {
|
||||
for (i = 0; codec->supported_samplerates[i]; ++i) {
|
||||
if (codec->supported_samplerates[i] < PREFERRED_SAMPLE_RATE) {
|
||||
continue;
|
||||
}
|
||||
if (encoder->sampleRate == PREFERRED_SAMPLE_RATE || encoder->sampleRate > codec->supported_samplerates[i]) {
|
||||
encoder->sampleRate = codec->supported_samplerates[i];
|
||||
}
|
||||
}
|
||||
} else if (codec->id == AV_CODEC_ID_AAC) {
|
||||
// HACK: AAC doesn't support 32768Hz (it rounds to 32000), but libfaac doesn't tell us that
|
||||
encoder->sampleRate = 44100;
|
||||
}
|
||||
encoder->audioCodec = acodec;
|
||||
encoder->audioBitrate = abr;
|
||||
return true;
|
||||
}
|
||||
|
||||
bool FFmpegEncoderSetVideo(struct FFmpegEncoder* encoder, const char* vcodec, unsigned vbr) {
|
||||
static struct {
|
||||
enum AVPixelFormat format;
|
||||
int priority;
|
||||
} priorities[] = {
|
||||
{ AV_PIX_FMT_RGB24, 0 },
|
||||
{ AV_PIX_FMT_BGR0, 1 },
|
||||
{ AV_PIX_FMT_YUV422P, 2 },
|
||||
{ AV_PIX_FMT_YUV444P, 3 },
|
||||
{ AV_PIX_FMT_YUV420P, 4 }
|
||||
};
|
||||
AVCodec* codec = avcodec_find_encoder_by_name(vcodec);
|
||||
if (!codec) {
|
||||
return false;
|
||||
}
|
||||
|
||||
size_t i;
|
||||
size_t j;
|
||||
int priority = INT_MAX;
|
||||
encoder->pixFormat = AV_PIX_FMT_NONE;
|
||||
for (i = 0; codec->pix_fmts[i] != AV_PIX_FMT_NONE; ++i) {
|
||||
if (codec->pix_fmts[i] == AV_PIX_FMT_RGB24) {
|
||||
encoder->pixFormat = AV_PIX_FMT_RGB24;
|
||||
break;
|
||||
}
|
||||
if (codec->pix_fmts[i] == AV_PIX_FMT_BGR0) {
|
||||
encoder->pixFormat = AV_PIX_FMT_BGR0;
|
||||
for (j = 0; j < sizeof(priorities) / sizeof(*priorities); ++j) {
|
||||
if (codec->pix_fmts[i] == priorities[j].format && priority > priorities[j].priority) {
|
||||
priority = priorities[j].priority;
|
||||
encoder->pixFormat = codec->pix_fmts[i];
|
||||
}
|
||||
}
|
||||
}
|
||||
if (encoder->pixFormat == AV_PIX_FMT_NONE) {
|
||||
|
@ -98,18 +144,34 @@ bool FFmpegEncoderOpen(struct FFmpegEncoder* encoder, const char* outfile) {
|
|||
encoder->audioStream = avformat_new_stream(encoder->context, acodec);
|
||||
encoder->audio = encoder->audioStream->codec;
|
||||
encoder->audio->bit_rate = encoder->audioBitrate;
|
||||
encoder->audio->sample_rate = 0x8000;
|
||||
encoder->audio->channels = 2;
|
||||
encoder->audio->channel_layout = AV_CH_LAYOUT_STEREO;
|
||||
encoder->audio->sample_fmt = AV_SAMPLE_FMT_S16;
|
||||
encoder->audio->sample_rate = encoder->sampleRate;
|
||||
encoder->audio->sample_fmt = encoder->sampleFormat;
|
||||
avcodec_open2(encoder->audio, acodec, 0);
|
||||
encoder->audioFrame = av_frame_alloc();
|
||||
encoder->audioFrame->nb_samples = encoder->audio->frame_size;
|
||||
encoder->audioFrame->format = encoder->audio->sample_fmt;
|
||||
encoder->audioFrame->pts = 0;
|
||||
encoder->audioBufferSize = av_samples_get_buffer_size(0, encoder->audio->channels, encoder->audio->frame_size, encoder->audio->sample_fmt, 0);
|
||||
encoder->audioBuffer = av_malloc(encoder->audioBufferSize);
|
||||
avcodec_fill_audio_frame(encoder->audioFrame, encoder->audio->channels, encoder->audio->sample_fmt, (const uint8_t*) encoder->audioBuffer, encoder->audioBufferSize, 0);
|
||||
if (encoder->sampleRate != PREFERRED_SAMPLE_RATE) {
|
||||
encoder->resampleContext = avresample_alloc_context();
|
||||
av_opt_set_int(encoder->resampleContext, "in_channel_layout", AV_CH_LAYOUT_STEREO, 0);
|
||||
av_opt_set_int(encoder->resampleContext, "out_channel_layout", AV_CH_LAYOUT_STEREO, 0);
|
||||
av_opt_set_int(encoder->resampleContext, "in_sample_rate", PREFERRED_SAMPLE_RATE, 0);
|
||||
av_opt_set_int(encoder->resampleContext, "out_sample_rate", encoder->sampleRate, 0);
|
||||
av_opt_set_int(encoder->resampleContext, "in_sample_fmt", encoder->sampleFormat, 0);
|
||||
av_opt_set_int(encoder->resampleContext, "out_sample_fmt", encoder->sampleFormat, 0);
|
||||
avresample_open(encoder->resampleContext);
|
||||
encoder->audioBufferSize = (encoder->audioFrame->nb_samples * PREFERRED_SAMPLE_RATE / encoder->sampleRate) * 4;
|
||||
encoder->audioBuffer = av_malloc(encoder->audioBufferSize);
|
||||
} else {
|
||||
encoder->resampleContext = 0;
|
||||
encoder->audioBufferSize = 0;
|
||||
encoder->audioBuffer = 0;
|
||||
}
|
||||
encoder->postaudioBufferSize = av_samples_get_buffer_size(0, encoder->audio->channels, encoder->audio->frame_size, encoder->audio->sample_fmt, 0);
|
||||
encoder->postaudioBuffer = av_malloc(encoder->postaudioBufferSize);
|
||||
avcodec_fill_audio_frame(encoder->audioFrame, encoder->audio->channels, encoder->audio->sample_fmt, (const uint8_t*) encoder->postaudioBuffer, encoder->postaudioBufferSize, 0);
|
||||
|
||||
encoder->videoStream = avformat_new_stream(encoder->context, vcodec);
|
||||
encoder->video = encoder->videoStream->codec;
|
||||
|
@ -146,14 +208,26 @@ void FFmpegEncoderClose(struct FFmpegEncoder* encoder) {
|
|||
av_write_trailer(encoder->context);
|
||||
avio_close(encoder->context->pb);
|
||||
|
||||
av_free(encoder->audioBuffer);
|
||||
av_free(encoder->postaudioBuffer);
|
||||
if (encoder->audioBuffer) {
|
||||
av_free(encoder->audioBuffer);
|
||||
}
|
||||
av_frame_free(&encoder->audioFrame);
|
||||
avcodec_close(encoder->audio);
|
||||
|
||||
av_frame_free(&encoder->videoFrame);
|
||||
avcodec_close(encoder->video);
|
||||
|
||||
if (encoder->resampleContext) {
|
||||
avresample_close(encoder->resampleContext);
|
||||
}
|
||||
|
||||
avformat_free_context(encoder->context);
|
||||
encoder->context = 0;
|
||||
|
||||
encoder->currentAudioSample = 0;
|
||||
encoder->currentAudioFrame = 0;
|
||||
encoder->currentVideoFrame = 0;
|
||||
}
|
||||
|
||||
void _ffmpegPostAudioFrame(struct GBAAVStream* stream, int32_t left, int32_t right) {
|
||||
|
@ -163,16 +237,56 @@ void _ffmpegPostAudioFrame(struct GBAAVStream* stream, int32_t left, int32_t rig
|
|||
}
|
||||
|
||||
av_frame_make_writable(encoder->audioFrame);
|
||||
encoder->audioBuffer[encoder->currentAudioSample * 2] = left;
|
||||
encoder->audioBuffer[encoder->currentAudioSample * 2 + 1] = right;
|
||||
encoder->audioFrame->pts = av_rescale_q(encoder->currentAudioFrame, encoder->audio->time_base, encoder->audioStream->time_base);
|
||||
uint16_t* buffers[2];
|
||||
int stride;
|
||||
bool planar = av_sample_fmt_is_planar(encoder->audio->sample_fmt);
|
||||
if (encoder->resampleContext) {
|
||||
buffers[0] = (uint16_t*) encoder->audioBuffer;
|
||||
if (planar) {
|
||||
stride = 1;
|
||||
buffers[1] = &buffers[0][encoder->audioBufferSize / 4];
|
||||
} else {
|
||||
stride = 2;
|
||||
buffers[1] = &buffers[0][1];
|
||||
}
|
||||
} else {
|
||||
buffers[0] = (uint16_t*) encoder->postaudioBuffer;
|
||||
if (planar) {
|
||||
stride = 1;
|
||||
buffers[1] = &buffers[0][encoder->postaudioBufferSize / 4];
|
||||
} else {
|
||||
stride = 2;
|
||||
buffers[1] = &buffers[0][1];
|
||||
}
|
||||
}
|
||||
buffers[0][encoder->currentAudioSample * stride] = left;
|
||||
buffers[1][encoder->currentAudioSample * stride] = right;
|
||||
|
||||
++encoder->currentAudioFrame;
|
||||
++encoder->currentAudioSample;
|
||||
|
||||
if ((encoder->currentAudioSample * 4) < encoder->audioBufferSize) {
|
||||
return;
|
||||
if (encoder->resampleContext) {
|
||||
if ((encoder->currentAudioSample * 4) < encoder->audioBufferSize) {
|
||||
return;
|
||||
}
|
||||
encoder->currentAudioSample = 0;
|
||||
|
||||
avresample_convert(encoder->resampleContext,
|
||||
0, 0, encoder->postaudioBufferSize / 4,
|
||||
(uint8_t**) buffers, 0, encoder->audioBufferSize / 4);
|
||||
if ((ssize_t) avresample_available(encoder->resampleContext) < (ssize_t) encoder->postaudioBufferSize / 4) {
|
||||
return;
|
||||
}
|
||||
avresample_read(encoder->resampleContext, encoder->audioFrame->data, encoder->postaudioBufferSize / 4);
|
||||
} else {
|
||||
if ((encoder->currentAudioSample * 4) < encoder->postaudioBufferSize) {
|
||||
return;
|
||||
}
|
||||
encoder->currentAudioSample = 0;
|
||||
}
|
||||
encoder->currentAudioSample = 0;
|
||||
|
||||
AVRational timeBase = { 1, PREFERRED_SAMPLE_RATE };
|
||||
encoder->audioFrame->pts = av_rescale_q(encoder->currentAudioFrame, timeBase, encoder->audioStream->time_base);
|
||||
|
||||
AVPacket packet;
|
||||
av_init_packet(&packet);
|
||||
|
|
|
@ -5,6 +5,7 @@
|
|||
|
||||
#include <libavcodec/avcodec.h>
|
||||
#include <libavformat/avformat.h>
|
||||
#include <libavresample/avresample.h>
|
||||
|
||||
struct FFmpegEncoder {
|
||||
struct GBAAVStream d;
|
||||
|
@ -19,11 +20,16 @@ struct FFmpegEncoder {
|
|||
const char* containerFormat;
|
||||
|
||||
AVCodecContext* audio;
|
||||
enum AVSampleFormat sampleFormat;
|
||||
int sampleRate;
|
||||
uint16_t* audioBuffer;
|
||||
size_t audioBufferSize;
|
||||
uint16_t* postaudioBuffer;
|
||||
size_t postaudioBufferSize;
|
||||
AVFrame* audioFrame;
|
||||
size_t currentAudioSample;
|
||||
int64_t currentAudioFrame;
|
||||
AVAudioResampleContext* resampleContext;
|
||||
AVStream* audioStream;
|
||||
|
||||
AVCodecContext* video;
|
||||
|
|
Loading…
Reference in New Issue