4e10eeda09
I can't believe I wasn't doing this. This is why file output was getting corrupted. Audio and video send in data from separate threads. I should be embarassed for not having considered that. Key lesson: Increase threading paranoia levels. Apparently my threading paranoid levels are lackluster.
699 lines
18 KiB
C
699 lines
18 KiB
C
/******************************************************************************
|
|
Copyright (C) 2014 by Hugh Bailey <obs.jim@gmail.com>
|
|
|
|
This program is free software: you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation, either version 2 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
******************************************************************************/
|
|
|
|
#include <obs.h>
|
|
#include <util/circlebuf.h>
|
|
#include <util/threading.h>
|
|
|
|
#include <libavutil/opt.h>
|
|
#include <libavformat/avformat.h>
|
|
#include <libswscale/swscale.h>
|
|
|
|
/* NOTE: much of this stuff is test stuff that was more or less copied from
|
|
* the muxing.c ffmpeg example */
|
|
|
|
struct ffmpeg_data {
|
|
AVStream *video;
|
|
AVStream *audio;
|
|
AVCodec *acodec;
|
|
AVCodec *vcodec;
|
|
AVFormatContext *output;
|
|
struct SwsContext *swscale;
|
|
|
|
AVPicture dst_picture;
|
|
AVFrame *vframe;
|
|
int frame_size;
|
|
int total_frames;
|
|
|
|
uint64_t start_timestamp;
|
|
|
|
uint32_t audio_samplerate;
|
|
enum audio_format audio_format;
|
|
size_t audio_planes;
|
|
size_t audio_size;
|
|
struct circlebuf excess_frames[MAX_AV_PLANES];
|
|
uint8_t *samples[MAX_AV_PLANES];
|
|
AVFrame *aframe;
|
|
int total_samples;
|
|
|
|
pthread_mutex_t write_mutex;
|
|
|
|
const char *filename_test;
|
|
|
|
bool initialized;
|
|
};
|
|
|
|
struct ffmpeg_output {
|
|
obs_output_t output;
|
|
volatile bool active;
|
|
struct ffmpeg_data ff_data;
|
|
};
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
|
|
static inline enum AVPixelFormat obs_to_ffmpeg_video_format(
|
|
enum video_format format)
|
|
{
|
|
switch (format) {
|
|
case VIDEO_FORMAT_NONE: return AV_PIX_FMT_NONE;
|
|
case VIDEO_FORMAT_I420: return AV_PIX_FMT_YUV420P;
|
|
case VIDEO_FORMAT_NV12: return AV_PIX_FMT_NV12;
|
|
case VIDEO_FORMAT_YVYU: return AV_PIX_FMT_NONE;
|
|
case VIDEO_FORMAT_YUY2: return AV_PIX_FMT_YUYV422;
|
|
case VIDEO_FORMAT_UYVY: return AV_PIX_FMT_UYVY422;
|
|
case VIDEO_FORMAT_RGBA: return AV_PIX_FMT_RGBA;
|
|
case VIDEO_FORMAT_BGRA: return AV_PIX_FMT_BGRA;
|
|
case VIDEO_FORMAT_BGRX: return AV_PIX_FMT_BGRA;
|
|
}
|
|
|
|
return AV_PIX_FMT_NONE;
|
|
}
|
|
|
|
static inline enum audio_format convert_ffmpeg_sample_format(
|
|
enum AVSampleFormat format)
|
|
{
|
|
switch ((uint32_t)format) {
|
|
case AV_SAMPLE_FMT_U8: return AUDIO_FORMAT_U8BIT;
|
|
case AV_SAMPLE_FMT_S16: return AUDIO_FORMAT_16BIT;
|
|
case AV_SAMPLE_FMT_S32: return AUDIO_FORMAT_32BIT;
|
|
case AV_SAMPLE_FMT_FLT: return AUDIO_FORMAT_FLOAT;
|
|
case AV_SAMPLE_FMT_U8P: return AUDIO_FORMAT_U8BIT_PLANAR;
|
|
case AV_SAMPLE_FMT_S16P: return AUDIO_FORMAT_16BIT_PLANAR;
|
|
case AV_SAMPLE_FMT_S32P: return AUDIO_FORMAT_32BIT_PLANAR;
|
|
case AV_SAMPLE_FMT_FLTP: return AUDIO_FORMAT_FLOAT_PLANAR;
|
|
}
|
|
|
|
/* shouldn't get here */
|
|
return AUDIO_FORMAT_16BIT;
|
|
}
|
|
|
|
static bool new_stream(struct ffmpeg_data *data, AVStream **stream,
|
|
AVCodec **codec, enum AVCodecID id)
|
|
{
|
|
*codec = avcodec_find_encoder(id);
|
|
if (!*codec) {
|
|
blog(LOG_ERROR, "Couldn't find encoder '%s'",
|
|
avcodec_get_name(id));
|
|
return false;
|
|
}
|
|
|
|
*stream = avformat_new_stream(data->output, *codec);
|
|
if (!*stream) {
|
|
blog(LOG_ERROR, "Couldn't create stream for encoder '%s'",
|
|
avcodec_get_name(id));
|
|
return false;
|
|
}
|
|
|
|
(*stream)->id = data->output->nb_streams-1;
|
|
return true;
|
|
}
|
|
|
|
static bool open_video_codec(struct ffmpeg_data *data)
|
|
{
|
|
AVCodecContext *context = data->video->codec;
|
|
int ret;
|
|
|
|
if (data->vcodec->id == AV_CODEC_ID_H264)
|
|
av_opt_set(context->priv_data, "preset", "veryfast", 0);
|
|
|
|
ret = avcodec_open2(context, data->vcodec, NULL);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Failed to open video codec: %s",
|
|
av_err2str(ret));
|
|
return false;
|
|
}
|
|
|
|
data->vframe = av_frame_alloc();
|
|
if (!data->vframe) {
|
|
blog(LOG_ERROR, "Failed to allocate video frame");
|
|
return false;
|
|
}
|
|
|
|
data->vframe->format = context->pix_fmt;
|
|
data->vframe->width = context->width;
|
|
data->vframe->height = context->height;
|
|
|
|
ret = avpicture_alloc(&data->dst_picture, context->pix_fmt,
|
|
context->width, context->height);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Failed to allocate dst_picture: %s",
|
|
av_err2str(ret));
|
|
return false;
|
|
}
|
|
|
|
*((AVPicture*)data->vframe) = data->dst_picture;
|
|
return true;
|
|
}
|
|
|
|
static bool init_swscale(struct ffmpeg_data *data, AVCodecContext *context)
|
|
{
|
|
data->swscale = sws_getContext(
|
|
context->width, context->height, AV_PIX_FMT_YUV420P,
|
|
context->width, context->height, context->pix_fmt,
|
|
SWS_BICUBIC, NULL, NULL, NULL);
|
|
|
|
if (!data->swscale) {
|
|
blog(LOG_ERROR, "Could not initialize swscale");
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool create_video_stream(struct ffmpeg_data *data)
|
|
{
|
|
AVCodecContext *context;
|
|
struct obs_video_info ovi;
|
|
|
|
if (!obs_get_video_info(&ovi)) {
|
|
blog(LOG_ERROR, "No active video");
|
|
return false;
|
|
}
|
|
|
|
if (!new_stream(data, &data->video, &data->vcodec,
|
|
data->output->oformat->video_codec))
|
|
return false;
|
|
|
|
context = data->video->codec;
|
|
context->codec_id = data->output->oformat->video_codec;
|
|
context->bit_rate = 6000000;
|
|
context->width = ovi.output_width;
|
|
context->height = ovi.output_height;
|
|
context->time_base.num = ovi.fps_den;
|
|
context->time_base.den = ovi.fps_num;
|
|
context->gop_size = 12;
|
|
context->pix_fmt = AV_PIX_FMT_YUV420P;
|
|
|
|
if (data->output->oformat->flags & AVFMT_GLOBALHEADER)
|
|
context->flags |= CODEC_FLAG_GLOBAL_HEADER;
|
|
|
|
if (!open_video_codec(data))
|
|
return false;
|
|
|
|
if (context->pix_fmt != AV_PIX_FMT_YUV420P)
|
|
if (!init_swscale(data, context))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool open_audio_codec(struct ffmpeg_data *data)
|
|
{
|
|
AVCodecContext *context = data->audio->codec;
|
|
int ret;
|
|
|
|
data->aframe = av_frame_alloc();
|
|
if (!data->aframe) {
|
|
blog(LOG_ERROR, "Failed to allocate audio frame");
|
|
return false;
|
|
}
|
|
|
|
context->strict_std_compliance = -2;
|
|
|
|
ret = avcodec_open2(context, data->acodec, NULL);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Failed to open audio codec: %s",
|
|
av_err2str(ret));
|
|
return false;
|
|
}
|
|
|
|
data->frame_size = context->frame_size ? context->frame_size : 1024;
|
|
|
|
ret = av_samples_alloc(data->samples, NULL, context->channels,
|
|
data->frame_size, context->sample_fmt, 0);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Failed to create audio buffer: %s",
|
|
av_err2str(ret));
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool create_audio_stream(struct ffmpeg_data *data)
|
|
{
|
|
AVCodecContext *context;
|
|
struct audio_output_info aoi;
|
|
|
|
if (!obs_get_audio_info(&aoi)) {
|
|
blog(LOG_ERROR, "No active audio");
|
|
return false;
|
|
}
|
|
|
|
if (!new_stream(data, &data->audio, &data->acodec,
|
|
data->output->oformat->audio_codec))
|
|
return false;
|
|
|
|
context = data->audio->codec;
|
|
context->bit_rate = 128000;
|
|
context->channels = get_audio_channels(aoi.speakers);
|
|
context->sample_rate = aoi.samples_per_sec;
|
|
context->sample_fmt = data->acodec->sample_fmts ?
|
|
data->acodec->sample_fmts[0] : AV_SAMPLE_FMT_FLTP;
|
|
|
|
data->audio_samplerate = aoi.samples_per_sec;
|
|
data->audio_format = convert_ffmpeg_sample_format(context->sample_fmt);
|
|
data->audio_planes = get_audio_planes(data->audio_format, aoi.speakers);
|
|
data->audio_size = get_audio_size(data->audio_format, aoi.speakers, 1);
|
|
|
|
if (data->output->oformat->flags & AVFMT_GLOBALHEADER)
|
|
context->flags |= CODEC_FLAG_GLOBAL_HEADER;
|
|
|
|
return open_audio_codec(data);
|
|
}
|
|
|
|
static inline bool init_streams(struct ffmpeg_data *data)
|
|
{
|
|
AVOutputFormat *format = data->output->oformat;
|
|
|
|
if (format->video_codec != AV_CODEC_ID_NONE)
|
|
if (!create_video_stream(data))
|
|
return false;
|
|
|
|
if (format->audio_codec != AV_CODEC_ID_NONE)
|
|
if (!create_audio_stream(data))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static inline bool open_output_file(struct ffmpeg_data *data)
|
|
{
|
|
AVOutputFormat *format = data->output->oformat;
|
|
int ret;
|
|
|
|
if ((format->flags & AVFMT_NOFILE) == 0) {
|
|
ret = avio_open(&data->output->pb, data->filename_test,
|
|
AVIO_FLAG_WRITE);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Couldn't open file '%s', %s",
|
|
data->filename_test, av_err2str(ret));
|
|
return false;
|
|
}
|
|
}
|
|
|
|
ret = avformat_write_header(data->output, NULL);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "Error opening file '%s': %s",
|
|
data->filename_test, av_err2str(ret));
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static void close_video(struct ffmpeg_data *data)
|
|
{
|
|
avcodec_close(data->video->codec);
|
|
avpicture_free(&data->dst_picture);
|
|
av_frame_free(&data->vframe);
|
|
}
|
|
|
|
static void close_audio(struct ffmpeg_data *data)
|
|
{
|
|
for (size_t i = 0; i < MAX_AV_PLANES; i++)
|
|
circlebuf_free(&data->excess_frames[i]);
|
|
|
|
av_freep(&data->samples[0]);
|
|
avcodec_close(data->audio->codec);
|
|
av_frame_free(&data->aframe);
|
|
}
|
|
|
|
static void ffmpeg_data_free(struct ffmpeg_data *data)
|
|
{
|
|
pthread_mutex_lock(&data->write_mutex);
|
|
if (data->initialized)
|
|
av_write_trailer(data->output);
|
|
|
|
if (data->video)
|
|
close_video(data);
|
|
if (data->audio)
|
|
close_audio(data);
|
|
if ((data->output->oformat->flags & AVFMT_NOFILE) == 0)
|
|
avio_close(data->output->pb);
|
|
|
|
avformat_free_context(data->output);
|
|
|
|
pthread_mutex_unlock(&data->write_mutex);
|
|
pthread_mutex_destroy(&data->write_mutex);
|
|
|
|
memset(data, 0, sizeof(struct ffmpeg_data));
|
|
}
|
|
|
|
static bool ffmpeg_data_init(struct ffmpeg_data *data, const char *filename)
|
|
{
|
|
memset(data, 0, sizeof(struct ffmpeg_data));
|
|
pthread_mutex_init_value(&data->write_mutex);
|
|
|
|
data->filename_test = filename;
|
|
|
|
if (!filename || !*filename)
|
|
return false;
|
|
|
|
if (pthread_mutex_init(&data->write_mutex, NULL) != 0)
|
|
return false;
|
|
|
|
av_register_all();
|
|
|
|
/* TODO: settings */
|
|
avformat_alloc_output_context2(&data->output, NULL, NULL,
|
|
data->filename_test);
|
|
if (!data->output) {
|
|
blog(LOG_ERROR, "Couldn't create avformat context");
|
|
goto fail;
|
|
}
|
|
|
|
if (!init_streams(data))
|
|
goto fail;
|
|
if (!open_output_file(data))
|
|
goto fail;
|
|
|
|
data->initialized = true;
|
|
return true;
|
|
|
|
fail:
|
|
blog(LOG_ERROR, "ffmpeg_data_init failed");
|
|
ffmpeg_data_free(data);
|
|
return false;
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
|
|
static const char *ffmpeg_output_getname(const char *locale)
|
|
{
|
|
UNUSED_PARAMETER(locale);
|
|
return "FFmpeg file output";
|
|
}
|
|
|
|
static void ffmpeg_log_callback(void *param, int level, const char *format,
|
|
va_list args)
|
|
{
|
|
blogva(LOG_DEBUG, format, args);
|
|
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
static void *ffmpeg_output_create(obs_data_t settings, obs_output_t output)
|
|
{
|
|
struct ffmpeg_output *data = bzalloc(sizeof(struct ffmpeg_output));
|
|
data->output = output;
|
|
|
|
av_log_set_callback(ffmpeg_log_callback);
|
|
|
|
UNUSED_PARAMETER(settings);
|
|
return data;
|
|
}
|
|
|
|
static void ffmpeg_output_destroy(void *data)
|
|
{
|
|
struct ffmpeg_output *output = data;
|
|
|
|
if (output) {
|
|
if (output->active)
|
|
ffmpeg_data_free(&output->ff_data);
|
|
bfree(data);
|
|
}
|
|
}
|
|
|
|
static inline int64_t rescale_ts(int64_t val, AVCodecContext *context,
|
|
AVStream *stream)
|
|
{
|
|
return av_rescale_q_rnd(val, context->time_base,
|
|
stream->time_base,
|
|
AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX);
|
|
}
|
|
|
|
#define YUV420_PLANES 3
|
|
|
|
static inline void copy_data(AVPicture *pic, const struct video_data *frame,
|
|
int height)
|
|
{
|
|
for (int plane = 0; plane < YUV420_PLANES; plane++) {
|
|
int frame_rowsize = (int)frame->linesize[plane];
|
|
int pic_rowsize = pic->linesize[plane];
|
|
int bytes = frame_rowsize < pic_rowsize ?
|
|
frame_rowsize : pic_rowsize;
|
|
int plane_height = plane == 0 ? height : height/2;
|
|
|
|
for (int y = 0; y < plane_height; y++) {
|
|
int pos_frame = y * frame_rowsize;
|
|
int pos_pic = y * pic_rowsize;
|
|
|
|
memcpy(pic->data[plane] + pos_pic,
|
|
frame->data[plane] + pos_frame,
|
|
bytes);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void receive_video(void *param, const struct video_data *frame)
|
|
{
|
|
struct ffmpeg_output *output = param;
|
|
struct ffmpeg_data *data = &output->ff_data;
|
|
AVCodecContext *context = data->video->codec;
|
|
AVPacket packet = {0};
|
|
int ret, got_packet;
|
|
|
|
av_init_packet(&packet);
|
|
|
|
if (!data->start_timestamp)
|
|
data->start_timestamp = frame->timestamp;
|
|
|
|
if (context->pix_fmt != AV_PIX_FMT_YUV420P)
|
|
sws_scale(data->swscale, frame->data,
|
|
(const int*)frame->linesize,
|
|
0, context->height, data->dst_picture.data,
|
|
data->dst_picture.linesize);
|
|
else
|
|
copy_data(&data->dst_picture, frame, context->height);
|
|
|
|
if (data->output->flags & AVFMT_RAWPICTURE) {
|
|
packet.flags |= AV_PKT_FLAG_KEY;
|
|
packet.stream_index = data->video->index;
|
|
packet.data = data->dst_picture.data[0];
|
|
packet.size = sizeof(AVPicture);
|
|
|
|
pthread_mutex_lock(&data->write_mutex);
|
|
ret = av_interleaved_write_frame(data->output, &packet);
|
|
pthread_mutex_unlock(&data->write_mutex);
|
|
|
|
} else {
|
|
data->vframe->pts = data->total_frames;
|
|
ret = avcodec_encode_video2(context, &packet, data->vframe,
|
|
&got_packet);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "receive_video: Error encoding "
|
|
"video: %s", av_err2str(ret));
|
|
return;
|
|
}
|
|
|
|
if (!ret && got_packet && packet.size) {
|
|
packet.pts = rescale_ts(packet.pts, context,
|
|
data->video);
|
|
packet.dts = rescale_ts(packet.dts, context,
|
|
data->video);
|
|
packet.duration = (int)av_rescale_q(packet.duration,
|
|
context->time_base,
|
|
data->video->time_base);
|
|
|
|
pthread_mutex_lock(&data->write_mutex);
|
|
ret = av_interleaved_write_frame(data->output,
|
|
&packet);
|
|
pthread_mutex_unlock(&data->write_mutex);
|
|
} else {
|
|
ret = 0;
|
|
}
|
|
}
|
|
|
|
if (ret != 0) {
|
|
blog(LOG_ERROR, "receive_video: Error writing video: %s",
|
|
av_err2str(ret));
|
|
}
|
|
|
|
data->total_frames++;
|
|
}
|
|
|
|
static inline void encode_audio(struct ffmpeg_data *output,
|
|
struct AVCodecContext *context, size_t block_size)
|
|
{
|
|
AVPacket packet = {0};
|
|
int ret, got_packet;
|
|
size_t total_size = output->frame_size * block_size * context->channels;
|
|
|
|
output->aframe->nb_samples = output->frame_size;
|
|
output->aframe->pts = av_rescale_q(output->total_samples,
|
|
(AVRational){1, context->sample_rate},
|
|
context->time_base);
|
|
|
|
ret = avcodec_fill_audio_frame(output->aframe, context->channels,
|
|
context->sample_fmt, output->samples[0],
|
|
(int)total_size, 1);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "receive_audio: avcodec_fill_audio_frame "
|
|
"failed: %s", av_err2str(ret));
|
|
return;
|
|
}
|
|
|
|
output->total_samples += output->frame_size;
|
|
|
|
ret = avcodec_encode_audio2(context, &packet, output->aframe,
|
|
&got_packet);
|
|
if (ret < 0) {
|
|
blog(LOG_ERROR, "receive_audio: Error encoding audio: %s",
|
|
av_err2str(ret));
|
|
return;
|
|
}
|
|
|
|
if (!got_packet)
|
|
return;
|
|
|
|
packet.pts = rescale_ts(packet.pts, context, output->audio);
|
|
packet.dts = rescale_ts(packet.dts, context, output->audio);
|
|
packet.duration = (int)av_rescale_q(packet.duration, context->time_base,
|
|
output->audio->time_base);
|
|
packet.stream_index = output->audio->index;
|
|
|
|
pthread_mutex_lock(&output->write_mutex);
|
|
ret = av_interleaved_write_frame(output->output, &packet);
|
|
if (ret != 0)
|
|
blog(LOG_ERROR, "receive_audio: Error writing audio: %s",
|
|
av_err2str(ret));
|
|
pthread_mutex_unlock(&output->write_mutex);
|
|
}
|
|
|
|
static bool prepare_audio(struct ffmpeg_data *data,
|
|
const struct audio_data *frame, struct audio_data *output)
|
|
{
|
|
*output = *frame;
|
|
|
|
if (frame->timestamp < data->start_timestamp) {
|
|
uint64_t duration = (uint64_t)frame->frames * 1000000000 /
|
|
(uint64_t)data->audio_samplerate;
|
|
uint64_t end_ts = (frame->timestamp + duration);
|
|
uint64_t cutoff;
|
|
|
|
if (end_ts <= data->start_timestamp)
|
|
return false;
|
|
|
|
cutoff = data->start_timestamp - frame->timestamp;
|
|
cutoff = cutoff * (uint64_t)data->audio_samplerate /
|
|
1000000000;
|
|
|
|
for (size_t i = 0; i < data->audio_planes; i++)
|
|
output->data[i] += data->audio_size * (uint32_t)cutoff;
|
|
output->frames -= (uint32_t)cutoff;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static void receive_audio(void *param, const struct audio_data *frame)
|
|
{
|
|
struct ffmpeg_output *output = param;
|
|
struct ffmpeg_data *data = &output->ff_data;
|
|
size_t frame_size_bytes;
|
|
struct audio_data in;
|
|
|
|
AVCodecContext *context = data->audio->codec;
|
|
|
|
if (!data->start_timestamp)
|
|
return;
|
|
if (!prepare_audio(data, frame, &in))
|
|
return;
|
|
|
|
frame_size_bytes = (size_t)data->frame_size * data->audio_size;
|
|
|
|
for (size_t i = 0; i < data->audio_planes; i++)
|
|
circlebuf_push_back(&data->excess_frames[i], in.data[i],
|
|
in.frames * data->audio_size);
|
|
|
|
while (data->excess_frames[0].size >= frame_size_bytes) {
|
|
for (size_t i = 0; i < data->audio_planes; i++)
|
|
circlebuf_pop_front(&data->excess_frames[i],
|
|
data->samples[i], frame_size_bytes);
|
|
|
|
encode_audio(data, context, data->audio_size);
|
|
}
|
|
}
|
|
|
|
static bool ffmpeg_output_start(void *data)
|
|
{
|
|
struct ffmpeg_output *output = data;
|
|
|
|
video_t video = obs_video();
|
|
audio_t audio = obs_audio();
|
|
|
|
if (!video || !audio) {
|
|
blog(LOG_ERROR, "ffmpeg_output_start: audio and video must "
|
|
"both be active (at least as of this writing)");
|
|
return false;
|
|
}
|
|
|
|
const char *filename_test;
|
|
obs_data_t settings = obs_output_get_settings(output->output);
|
|
filename_test = obs_data_getstring(settings, "filename");
|
|
obs_data_release(settings);
|
|
|
|
if (!filename_test || !*filename_test)
|
|
return false;
|
|
|
|
if (!ffmpeg_data_init(&output->ff_data, filename_test))
|
|
return false;
|
|
|
|
struct audio_convert_info aci = {
|
|
.format = output->ff_data.audio_format
|
|
};
|
|
|
|
struct video_scale_info vsi = {
|
|
.format = VIDEO_FORMAT_I420
|
|
};
|
|
|
|
video_output_connect(video, &vsi, receive_video, output);
|
|
audio_output_connect(audio, &aci, receive_audio, output);
|
|
output->active = true;
|
|
|
|
return true;
|
|
}
|
|
|
|
static void ffmpeg_output_stop(void *data)
|
|
{
|
|
struct ffmpeg_output *output = data;
|
|
|
|
if (output->active) {
|
|
output->active = false;
|
|
video_output_disconnect(obs_video(), receive_video, data);
|
|
audio_output_disconnect(obs_audio(), receive_audio, data);
|
|
ffmpeg_data_free(&output->ff_data);
|
|
}
|
|
}
|
|
|
|
static bool ffmpeg_output_active(void *data)
|
|
{
|
|
struct ffmpeg_output *output = data;
|
|
return output->active;
|
|
}
|
|
|
|
struct obs_output_info ffmpeg_output = {
|
|
.id = "ffmpeg_output",
|
|
.getname = ffmpeg_output_getname,
|
|
.create = ffmpeg_output_create,
|
|
.destroy = ffmpeg_output_destroy,
|
|
.start = ffmpeg_output_start,
|
|
.stop = ffmpeg_output_stop,
|
|
.active = ffmpeg_output_active
|
|
};
|