1994 lines
52 KiB
C
1994 lines
52 KiB
C
/******************************************************************************
|
|
Copyright (C) 2013-2014 by Hugh Bailey <obs.jim@gmail.com>
|
|
|
|
This program is free software: you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation, either version 2 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
******************************************************************************/
|
|
|
|
#include <inttypes.h>
|
|
|
|
#include "media-io/format-conversion.h"
|
|
#include "media-io/video-frame.h"
|
|
#include "media-io/audio-io.h"
|
|
#include "util/threading.h"
|
|
#include "util/platform.h"
|
|
#include "callback/calldata.h"
|
|
#include "graphics/matrix3.h"
|
|
#include "graphics/vec3.h"
|
|
|
|
#include "obs.h"
|
|
#include "obs-internal.h"
|
|
|
|
static inline bool source_valid(struct obs_source *source)
|
|
{
|
|
return source && source->context.data;
|
|
}
|
|
|
|
const struct obs_source_info *find_source(struct darray *list, const char *id)
|
|
{
|
|
size_t i;
|
|
struct obs_source_info *array = list->array;
|
|
|
|
for (i = 0; i < list->num; i++) {
|
|
struct obs_source_info *info = array+i;
|
|
if (strcmp(info->id, id) == 0)
|
|
return info;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const struct obs_source_info *get_source_info(enum obs_source_type type,
|
|
const char *id)
|
|
{
|
|
struct darray *list = NULL;
|
|
|
|
switch (type) {
|
|
case OBS_SOURCE_TYPE_INPUT:
|
|
list = &obs->input_types.da;
|
|
break;
|
|
|
|
case OBS_SOURCE_TYPE_FILTER:
|
|
list = &obs->filter_types.da;
|
|
break;
|
|
|
|
case OBS_SOURCE_TYPE_TRANSITION:
|
|
list = &obs->transition_types.da;
|
|
break;
|
|
}
|
|
|
|
return find_source(list, id);
|
|
}
|
|
|
|
static const char *source_signals[] = {
|
|
"void destroy(ptr source)",
|
|
"void add(ptr source)",
|
|
"void remove(ptr source)",
|
|
"void activate(ptr source)",
|
|
"void deactivate(ptr source)",
|
|
"void show(ptr source)",
|
|
"void hide(ptr source)",
|
|
"void rename(ptr source, string new_name, string prev_name)",
|
|
"void volume(ptr source, in out float volume)",
|
|
"void volume_level(ptr source, float level, float magnitude, "
|
|
"float peak)",
|
|
NULL
|
|
};
|
|
|
|
bool obs_source_init_context(struct obs_source *source,
|
|
obs_data_t settings, const char *name)
|
|
{
|
|
if (!obs_context_data_init(&source->context, settings, name))
|
|
return false;
|
|
|
|
return signal_handler_add_array(source->context.signals,
|
|
source_signals);
|
|
}
|
|
|
|
const char *obs_source_get_display_name(enum obs_source_type type,
|
|
const char *id)
|
|
{
|
|
const struct obs_source_info *info = get_source_info(type, id);
|
|
return (info != NULL) ? info->get_name() : NULL;
|
|
}
|
|
|
|
/* internal initialization */
|
|
bool obs_source_init(struct obs_source *source,
|
|
const struct obs_source_info *info)
|
|
{
|
|
source->refs = 1;
|
|
source->user_volume = 1.0f;
|
|
source->present_volume = 0.0f;
|
|
source->sync_offset = 0;
|
|
pthread_mutex_init_value(&source->filter_mutex);
|
|
pthread_mutex_init_value(&source->video_mutex);
|
|
pthread_mutex_init_value(&source->audio_mutex);
|
|
|
|
if (pthread_mutex_init(&source->filter_mutex, NULL) != 0)
|
|
return false;
|
|
if (pthread_mutex_init(&source->audio_mutex, NULL) != 0)
|
|
return false;
|
|
if (pthread_mutex_init(&source->video_mutex, NULL) != 0)
|
|
return false;
|
|
|
|
if (info && info->output_flags & OBS_SOURCE_AUDIO) {
|
|
source->audio_line = audio_output_create_line(obs->audio.audio,
|
|
source->context.name);
|
|
if (!source->audio_line) {
|
|
blog(LOG_ERROR, "Failed to create audio line for "
|
|
"source '%s'", source->context.name);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
obs_context_data_insert(&source->context,
|
|
&obs->data.sources_mutex,
|
|
&obs->data.first_source);
|
|
return true;
|
|
}
|
|
|
|
static inline void obs_source_dosignal(struct obs_source *source,
|
|
const char *signal_obs, const char *signal_source)
|
|
{
|
|
struct calldata data;
|
|
|
|
calldata_init(&data);
|
|
calldata_set_ptr(&data, "source", source);
|
|
if (signal_obs)
|
|
signal_handler_signal(obs->signals, signal_obs, &data);
|
|
if (signal_source)
|
|
signal_handler_signal(source->context.signals, signal_source,
|
|
&data);
|
|
calldata_free(&data);
|
|
}
|
|
|
|
obs_source_t obs_source_create(enum obs_source_type type, const char *id,
|
|
const char *name, obs_data_t settings)
|
|
{
|
|
struct obs_source *source = bzalloc(sizeof(struct obs_source));
|
|
|
|
const struct obs_source_info *info = get_source_info(type, id);
|
|
if (!info) {
|
|
blog(LOG_ERROR, "Source ID '%s' not found", id);
|
|
|
|
source->info.id = bstrdup(id);
|
|
source->info.type = type;
|
|
source->owns_info_id = true;
|
|
} else {
|
|
source->info = *info;
|
|
}
|
|
|
|
if (!obs_source_init_context(source, settings, name))
|
|
goto fail;
|
|
|
|
if (info && info->get_defaults)
|
|
info->get_defaults(source->context.settings);
|
|
|
|
/* allow the source to be created even if creation fails so that the
|
|
* user's data doesn't become lost */
|
|
if (info)
|
|
source->context.data = info->create(source->context.settings,
|
|
source);
|
|
if (!source->context.data)
|
|
blog(LOG_ERROR, "Failed to create source '%s'!", name);
|
|
|
|
if (!obs_source_init(source, info))
|
|
goto fail;
|
|
|
|
blog(LOG_INFO, "source '%s' (%s) created", name, id);
|
|
obs_source_dosignal(source, "source_create", NULL);
|
|
return source;
|
|
|
|
fail:
|
|
blog(LOG_ERROR, "obs_source_create failed");
|
|
obs_source_destroy(source);
|
|
return NULL;
|
|
}
|
|
|
|
void obs_source_frame_init(struct obs_source_frame *frame,
|
|
enum video_format format, uint32_t width, uint32_t height)
|
|
{
|
|
struct video_frame vid_frame;
|
|
|
|
if (!frame)
|
|
return;
|
|
|
|
video_frame_init(&vid_frame, format, width, height);
|
|
frame->format = format;
|
|
frame->width = width;
|
|
frame->height = height;
|
|
|
|
for (size_t i = 0; i < MAX_AV_PLANES; i++) {
|
|
frame->data[i] = vid_frame.data[i];
|
|
frame->linesize[i] = vid_frame.linesize[i];
|
|
}
|
|
}
|
|
|
|
void obs_source_destroy(struct obs_source *source)
|
|
{
|
|
size_t i;
|
|
|
|
if (!source)
|
|
return;
|
|
|
|
obs_context_data_remove(&source->context);
|
|
|
|
blog(LOG_INFO, "source '%s' destroyed", source->context.name);
|
|
|
|
obs_source_dosignal(source, "source_destroy", "destroy");
|
|
|
|
if (source->context.data) {
|
|
source->info.destroy(source->context.data);
|
|
source->context.data = NULL;
|
|
}
|
|
|
|
if (source->filter_parent)
|
|
obs_source_filter_remove(source->filter_parent, source);
|
|
|
|
for (i = 0; i < source->filters.num; i++)
|
|
obs_source_release(source->filters.array[i]);
|
|
|
|
for (i = 0; i < source->video_frames.num; i++)
|
|
obs_source_frame_destroy(source->video_frames.array[i]);
|
|
|
|
gs_enter_context(obs->video.graphics);
|
|
gs_texrender_destroy(source->async_convert_texrender);
|
|
gs_texture_destroy(source->async_texture);
|
|
gs_leave_context();
|
|
|
|
for (i = 0; i < MAX_AV_PLANES; i++)
|
|
bfree(source->audio_data.data[i]);
|
|
|
|
audio_line_destroy(source->audio_line);
|
|
audio_resampler_destroy(source->resampler);
|
|
|
|
gs_texrender_destroy(source->filter_texrender);
|
|
da_free(source->video_frames);
|
|
da_free(source->filters);
|
|
pthread_mutex_destroy(&source->filter_mutex);
|
|
pthread_mutex_destroy(&source->audio_mutex);
|
|
pthread_mutex_destroy(&source->video_mutex);
|
|
obs_context_data_free(&source->context);
|
|
|
|
if (source->owns_info_id)
|
|
bfree((void*)source->info.id);
|
|
|
|
bfree(source);
|
|
}
|
|
|
|
void obs_source_addref(obs_source_t source)
|
|
{
|
|
if (source)
|
|
os_atomic_inc_long(&source->refs);
|
|
}
|
|
|
|
void obs_source_release(obs_source_t source)
|
|
{
|
|
if (!source)
|
|
return;
|
|
|
|
if (os_atomic_dec_long(&source->refs) == 0)
|
|
obs_source_destroy(source);
|
|
}
|
|
|
|
void obs_source_remove(obs_source_t source)
|
|
{
|
|
struct obs_core_data *data = &obs->data;
|
|
size_t id;
|
|
bool exists;
|
|
|
|
pthread_mutex_lock(&data->sources_mutex);
|
|
|
|
if (!source || source->removed) {
|
|
pthread_mutex_unlock(&data->sources_mutex);
|
|
return;
|
|
}
|
|
|
|
source->removed = true;
|
|
|
|
obs_source_addref(source);
|
|
|
|
id = da_find(data->user_sources, &source, 0);
|
|
exists = (id != DARRAY_INVALID);
|
|
if (exists) {
|
|
da_erase(data->user_sources, id);
|
|
obs_source_release(source);
|
|
}
|
|
|
|
pthread_mutex_unlock(&data->sources_mutex);
|
|
|
|
if (exists)
|
|
obs_source_dosignal(source, "source_remove", "remove");
|
|
|
|
obs_source_release(source);
|
|
}
|
|
|
|
bool obs_source_removed(obs_source_t source)
|
|
{
|
|
return source ? source->removed : true;
|
|
}
|
|
|
|
static inline obs_data_t get_defaults(const struct obs_source_info *info)
|
|
{
|
|
obs_data_t settings = obs_data_create();
|
|
if (info->get_defaults)
|
|
info->get_defaults(settings);
|
|
return settings;
|
|
}
|
|
|
|
obs_data_t obs_source_settings(enum obs_source_type type, const char *id)
|
|
{
|
|
const struct obs_source_info *info = get_source_info(type, id);
|
|
return (info) ? get_defaults(info) : NULL;
|
|
}
|
|
|
|
obs_properties_t obs_get_source_properties(enum obs_source_type type,
|
|
const char *id)
|
|
{
|
|
const struct obs_source_info *info = get_source_info(type, id);
|
|
if (info && info->get_properties) {
|
|
obs_data_t defaults = get_defaults(info);
|
|
obs_properties_t properties;
|
|
|
|
properties = info->get_properties();
|
|
obs_properties_apply_settings(properties, defaults);
|
|
obs_data_release(defaults);
|
|
return properties;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
obs_properties_t obs_source_properties(obs_source_t source)
|
|
{
|
|
if (source_valid(source) && source->info.get_properties) {
|
|
obs_properties_t props;
|
|
props = source->info.get_properties();
|
|
obs_properties_apply_settings(props, source->context.settings);
|
|
return props;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
uint32_t obs_source_get_output_flags(obs_source_t source)
|
|
{
|
|
return source ? source->info.output_flags : 0;
|
|
}
|
|
|
|
static void obs_source_deferred_update(obs_source_t source)
|
|
{
|
|
if (source->context.data && source->info.update)
|
|
source->info.update(source->context.data,
|
|
source->context.settings);
|
|
|
|
source->defer_update = false;
|
|
}
|
|
|
|
void obs_source_update(obs_source_t source, obs_data_t settings)
|
|
{
|
|
if (!source) return;
|
|
|
|
if (settings)
|
|
obs_data_apply(source->context.settings, settings);
|
|
|
|
if (source->info.output_flags & OBS_SOURCE_VIDEO) {
|
|
source->defer_update = true;
|
|
} else if (source->context.data && source->info.update) {
|
|
source->info.update(source->context.data,
|
|
source->context.settings);
|
|
}
|
|
}
|
|
|
|
static void activate_source(obs_source_t source)
|
|
{
|
|
if (source->context.data && source->info.activate)
|
|
source->info.activate(source->context.data);
|
|
obs_source_dosignal(source, "source_activate", "activate");
|
|
}
|
|
|
|
static void deactivate_source(obs_source_t source)
|
|
{
|
|
if (source->context.data && source->info.deactivate)
|
|
source->info.deactivate(source->context.data);
|
|
obs_source_dosignal(source, "source_deactivate", "deactivate");
|
|
}
|
|
|
|
static void show_source(obs_source_t source)
|
|
{
|
|
if (source->context.data && source->info.show)
|
|
source->info.show(source->context.data);
|
|
obs_source_dosignal(source, "source_show", "show");
|
|
}
|
|
|
|
static void hide_source(obs_source_t source)
|
|
{
|
|
if (source->context.data && source->info.hide)
|
|
source->info.hide(source->context.data);
|
|
obs_source_dosignal(source, "source_hide", "hide");
|
|
}
|
|
|
|
static void activate_tree(obs_source_t parent, obs_source_t child, void *param)
|
|
{
|
|
if (os_atomic_inc_long(&child->activate_refs) == 1)
|
|
activate_source(child);
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
static void deactivate_tree(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
if (os_atomic_dec_long(&child->activate_refs) == 0)
|
|
deactivate_source(child);
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
static void show_tree(obs_source_t parent, obs_source_t child, void *param)
|
|
{
|
|
if (os_atomic_inc_long(&child->show_refs) == 1)
|
|
show_source(child);
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
static void hide_tree(obs_source_t parent, obs_source_t child, void *param)
|
|
{
|
|
if (os_atomic_dec_long(&child->show_refs) == 0)
|
|
hide_source(child);
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
void obs_source_activate(obs_source_t source, enum view_type type)
|
|
{
|
|
if (!source) return;
|
|
|
|
if (os_atomic_inc_long(&source->show_refs) == 1) {
|
|
show_source(source);
|
|
obs_source_enum_tree(source, show_tree, NULL);
|
|
}
|
|
|
|
if (type == MAIN_VIEW) {
|
|
if (os_atomic_inc_long(&source->activate_refs) == 1) {
|
|
activate_source(source);
|
|
obs_source_enum_tree(source, activate_tree, NULL);
|
|
obs_source_set_present_volume(source, 1.0f);
|
|
}
|
|
}
|
|
}
|
|
|
|
void obs_source_deactivate(obs_source_t source, enum view_type type)
|
|
{
|
|
if (!source) return;
|
|
|
|
if (os_atomic_dec_long(&source->show_refs) == 0) {
|
|
hide_source(source);
|
|
obs_source_enum_tree(source, hide_tree, NULL);
|
|
}
|
|
|
|
if (type == MAIN_VIEW) {
|
|
if (os_atomic_dec_long(&source->activate_refs) == 0) {
|
|
deactivate_source(source);
|
|
obs_source_enum_tree(source, deactivate_tree, NULL);
|
|
obs_source_set_present_volume(source, 0.0f);
|
|
}
|
|
}
|
|
}
|
|
|
|
void obs_source_video_tick(obs_source_t source, float seconds)
|
|
{
|
|
if (!source) return;
|
|
|
|
if (source->defer_update)
|
|
obs_source_deferred_update(source);
|
|
|
|
/* reset the filter render texture information once every frame */
|
|
if (source->filter_texrender)
|
|
gs_texrender_reset(source->filter_texrender);
|
|
|
|
if (source->context.data && source->info.video_tick)
|
|
source->info.video_tick(source->context.data, seconds);
|
|
|
|
source->async_rendered = false;
|
|
}
|
|
|
|
/* unless the value is 3+ hours worth of frames, this won't overflow */
|
|
static inline uint64_t conv_frames_to_time(size_t frames)
|
|
{
|
|
const struct audio_output_info *info;
|
|
info = audio_output_get_info(obs->audio.audio);
|
|
|
|
return (uint64_t)frames * 1000000000ULL /
|
|
(uint64_t)info->samples_per_sec;
|
|
}
|
|
|
|
/* maximum "direct" timestamp variance in nanoseconds */
|
|
#define MAX_TS_VAR 5000000000ULL
|
|
/* maximum time that timestamp can jump in nanoseconds */
|
|
#define MAX_TIMESTAMP_JUMP 2000000000ULL
|
|
|
|
static inline void reset_audio_timing(obs_source_t source, uint64_t timetamp)
|
|
{
|
|
source->timing_set = true;
|
|
source->timing_adjust = os_gettime_ns() - timetamp;
|
|
}
|
|
|
|
static inline void handle_ts_jump(obs_source_t source, uint64_t expected,
|
|
uint64_t ts, uint64_t diff)
|
|
{
|
|
blog(LOG_DEBUG, "Timestamp for source '%s' jumped by '%"PRIu64"', "
|
|
"expected value %"PRIu64", input value %"PRIu64,
|
|
source->context.name, diff, expected, ts);
|
|
|
|
/* if has video, ignore audio data until reset */
|
|
if (!(source->info.output_flags & OBS_SOURCE_ASYNC))
|
|
reset_audio_timing(source, ts);
|
|
}
|
|
|
|
#define VOL_MIN -96.0f
|
|
#define VOL_MAX 0.0f
|
|
|
|
static inline float to_db(float val)
|
|
{
|
|
float db = 20.0f * log10f(val);
|
|
return isfinite(db) ? db : VOL_MIN;
|
|
}
|
|
|
|
static void calc_volume_levels(struct obs_source *source, float *array,
|
|
size_t frames, float volume)
|
|
{
|
|
float sum_val = 0.0f;
|
|
float max_val = 0.0f;
|
|
float rms_val = 0.0f;
|
|
|
|
audio_t audio = obs_get_audio();
|
|
const uint32_t sample_rate = audio_output_get_sample_rate(audio);
|
|
const size_t channels = audio_output_get_channels(audio);
|
|
const size_t count = frames * channels;
|
|
const size_t vol_peak_delay = sample_rate * 3;
|
|
const float alpha = 0.15f;
|
|
|
|
for (size_t i = 0; i < count; i++) {
|
|
float val = array[i];
|
|
float val_pow2 = val * val;
|
|
|
|
sum_val += val_pow2;
|
|
max_val = fmaxf(max_val, val_pow2);
|
|
}
|
|
|
|
/*
|
|
We want the volume meters scale linearly in respect to current
|
|
volume, so, no need to apply volume here.
|
|
*/
|
|
|
|
UNUSED_PARAMETER(volume);
|
|
|
|
rms_val = to_db(sqrtf(sum_val / (float)count));
|
|
max_val = to_db(sqrtf(max_val));
|
|
|
|
if (max_val > source->vol_max)
|
|
source->vol_max = max_val;
|
|
else
|
|
source->vol_max = alpha * source->vol_max +
|
|
(1.0f - alpha) * max_val;
|
|
|
|
if (source->vol_max > source->vol_peak ||
|
|
source->vol_update_count > vol_peak_delay) {
|
|
source->vol_peak = source->vol_max;
|
|
source->vol_update_count = 0;
|
|
} else {
|
|
source->vol_update_count += count;
|
|
}
|
|
|
|
source->vol_mag = alpha * rms_val + source->vol_mag * (1.0f - alpha);
|
|
}
|
|
|
|
/* TODO update peak/etc later */
|
|
static void obs_source_update_volume_level(obs_source_t source,
|
|
struct audio_data *in)
|
|
{
|
|
if (source && in) {
|
|
struct calldata data = {0};
|
|
|
|
calc_volume_levels(source, (float*)in->data[0], in->frames,
|
|
in->volume);
|
|
|
|
calldata_set_ptr (&data, "source", source);
|
|
calldata_set_float(&data, "level", source->vol_max);
|
|
calldata_set_float(&data, "magnitude", source->vol_mag);
|
|
calldata_set_float(&data, "peak", source->vol_peak);
|
|
|
|
signal_handler_signal(source->context.signals, "volume_level",
|
|
&data);
|
|
signal_handler_signal(obs->signals, "source_volume_level",
|
|
&data);
|
|
|
|
calldata_free(&data);
|
|
}
|
|
}
|
|
|
|
static void source_output_audio_line(obs_source_t source,
|
|
const struct audio_data *data)
|
|
{
|
|
struct audio_data in = *data;
|
|
uint64_t diff;
|
|
|
|
if (!source->timing_set) {
|
|
reset_audio_timing(source, in.timestamp);
|
|
|
|
/* detects 'directly' set timestamps as long as they're within
|
|
* a certain threshold */
|
|
if ((source->timing_adjust + MAX_TS_VAR) < MAX_TS_VAR * 2)
|
|
source->timing_adjust = 0;
|
|
|
|
} else if (source->next_audio_ts_min != 0) {
|
|
bool ts_under = (in.timestamp < source->next_audio_ts_min);
|
|
|
|
diff = ts_under ?
|
|
(source->next_audio_ts_min - in.timestamp) :
|
|
(in.timestamp - source->next_audio_ts_min);
|
|
|
|
/* smooth audio if lower or within threshold */
|
|
if (diff > MAX_TIMESTAMP_JUMP)
|
|
handle_ts_jump(source, source->next_audio_ts_min,
|
|
in.timestamp, diff);
|
|
else if (ts_under || diff < TS_SMOOTHING_THRESHOLD)
|
|
in.timestamp = source->next_audio_ts_min;
|
|
}
|
|
|
|
source->next_audio_ts_min = in.timestamp +
|
|
conv_frames_to_time(in.frames);
|
|
|
|
in.timestamp += source->timing_adjust + source->sync_offset;
|
|
in.volume = source->user_volume * source->present_volume *
|
|
obs->audio.user_volume * obs->audio.present_volume;
|
|
|
|
audio_line_output(source->audio_line, &in);
|
|
obs_source_update_volume_level(source, &in);
|
|
}
|
|
|
|
enum convert_type {
|
|
CONVERT_NONE,
|
|
CONVERT_NV12,
|
|
CONVERT_420,
|
|
CONVERT_422_U,
|
|
CONVERT_422_Y,
|
|
};
|
|
|
|
static inline enum convert_type get_convert_type(enum video_format format)
|
|
{
|
|
switch (format) {
|
|
case VIDEO_FORMAT_I420:
|
|
return CONVERT_420;
|
|
case VIDEO_FORMAT_NV12:
|
|
return CONVERT_NV12;
|
|
|
|
case VIDEO_FORMAT_YVYU:
|
|
case VIDEO_FORMAT_YUY2:
|
|
return CONVERT_422_Y;
|
|
case VIDEO_FORMAT_UYVY:
|
|
return CONVERT_422_U;
|
|
|
|
case VIDEO_FORMAT_NONE:
|
|
case VIDEO_FORMAT_RGBA:
|
|
case VIDEO_FORMAT_BGRA:
|
|
case VIDEO_FORMAT_BGRX:
|
|
return CONVERT_NONE;
|
|
}
|
|
|
|
return CONVERT_NONE;
|
|
}
|
|
|
|
static inline bool set_packed422_sizes(struct obs_source *source,
|
|
struct obs_source_frame *frame)
|
|
{
|
|
source->async_convert_height = frame->height;
|
|
source->async_convert_width = frame->width / 2;
|
|
source->async_texture_format = GS_BGRA;
|
|
return true;
|
|
}
|
|
|
|
static inline bool set_planar420_sizes(struct obs_source *source,
|
|
struct obs_source_frame *frame)
|
|
{
|
|
uint32_t size = frame->width * frame->height;
|
|
size += size/2;
|
|
|
|
source->async_convert_width = frame->width;
|
|
source->async_convert_height = (size / frame->width + 1) & 0xFFFFFFFE;
|
|
source->async_texture_format = GS_R8;
|
|
source->async_plane_offset[0] = frame->width * frame->height;
|
|
source->async_plane_offset[1] = source->async_plane_offset[0] +
|
|
frame->width * frame->height / 4;
|
|
return true;
|
|
}
|
|
|
|
static inline bool init_gpu_conversion(struct obs_source *source,
|
|
struct obs_source_frame *frame)
|
|
{
|
|
switch (get_convert_type(frame->format)) {
|
|
case CONVERT_422_Y:
|
|
case CONVERT_422_U:
|
|
return set_packed422_sizes(source, frame);
|
|
|
|
case CONVERT_420:
|
|
return set_planar420_sizes(source, frame);
|
|
|
|
case CONVERT_NV12:
|
|
assert(false && "NV12 not yet implemented");
|
|
/* TODO: implement conversion */
|
|
break;
|
|
|
|
case CONVERT_NONE:
|
|
assert(false && "No conversion requested");
|
|
break;
|
|
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static inline enum gs_color_format convert_video_format(
|
|
enum video_format format)
|
|
{
|
|
if (format == VIDEO_FORMAT_RGBA)
|
|
return GS_RGBA;
|
|
else if (format == VIDEO_FORMAT_BGRA)
|
|
return GS_BGRA;
|
|
|
|
return GS_BGRX;
|
|
}
|
|
|
|
static inline bool set_async_texture_size(struct obs_source *source,
|
|
struct obs_source_frame *frame)
|
|
{
|
|
enum convert_type prev, cur;
|
|
prev = get_convert_type(source->async_format);
|
|
cur = get_convert_type(frame->format);
|
|
if (source->async_texture) {
|
|
if (source->async_width == frame->width &&
|
|
source->async_height == frame->height &&
|
|
prev == cur)
|
|
return true;
|
|
}
|
|
|
|
gs_texture_destroy(source->async_texture);
|
|
gs_texrender_destroy(source->async_convert_texrender);
|
|
source->async_convert_texrender = NULL;
|
|
|
|
if (cur != CONVERT_NONE && init_gpu_conversion(source, frame)) {
|
|
source->async_gpu_conversion = true;
|
|
|
|
source->async_convert_texrender =
|
|
gs_texrender_create(GS_BGRX, GS_ZS_NONE);
|
|
|
|
source->async_texture = gs_texture_create(
|
|
source->async_convert_width,
|
|
source->async_convert_height,
|
|
source->async_texture_format,
|
|
1, NULL, GS_DYNAMIC);
|
|
|
|
} else {
|
|
enum gs_color_format format = convert_video_format(
|
|
frame->format);
|
|
source->async_gpu_conversion = false;
|
|
|
|
source->async_texture = gs_texture_create(
|
|
frame->width, frame->height,
|
|
format, 1, NULL, GS_DYNAMIC);
|
|
}
|
|
|
|
if (!source->async_texture)
|
|
return false;
|
|
|
|
source->async_width = frame->width;
|
|
source->async_height = frame->height;
|
|
return true;
|
|
}
|
|
|
|
static void upload_raw_frame(gs_texture_t tex,
|
|
const struct obs_source_frame *frame)
|
|
{
|
|
switch (get_convert_type(frame->format)) {
|
|
case CONVERT_422_U:
|
|
case CONVERT_422_Y:
|
|
gs_texture_set_image(tex, frame->data[0],
|
|
frame->linesize[0], false);
|
|
break;
|
|
|
|
case CONVERT_420:
|
|
gs_texture_set_image(tex, frame->data[0],
|
|
frame->width, false);
|
|
break;
|
|
|
|
case CONVERT_NV12:
|
|
assert(false && "Conversion not yet implemented");
|
|
break;
|
|
|
|
case CONVERT_NONE:
|
|
assert(false && "No conversion requested");
|
|
break;
|
|
}
|
|
}
|
|
|
|
static const char *select_conversion_technique(enum video_format format)
|
|
{
|
|
switch (format) {
|
|
case VIDEO_FORMAT_UYVY:
|
|
return "UYVY_Reverse";
|
|
|
|
case VIDEO_FORMAT_YUY2:
|
|
return "YUY2_Reverse";
|
|
|
|
case VIDEO_FORMAT_YVYU:
|
|
return "YVYU_Reverse";
|
|
|
|
case VIDEO_FORMAT_I420:
|
|
return "I420_Reverse";
|
|
|
|
case VIDEO_FORMAT_NV12:
|
|
assert(false && "Conversion not yet implemented");
|
|
break;
|
|
|
|
case VIDEO_FORMAT_BGRA:
|
|
case VIDEO_FORMAT_BGRX:
|
|
case VIDEO_FORMAT_RGBA:
|
|
case VIDEO_FORMAT_NONE:
|
|
assert(false && "No conversion requested");
|
|
break;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static inline void set_eparam(gs_effect_t effect, const char *name, float val)
|
|
{
|
|
gs_eparam_t param = gs_effect_get_param_by_name(effect, name);
|
|
gs_effect_set_float(param, val);
|
|
}
|
|
|
|
static bool update_async_texrender(struct obs_source *source,
|
|
const struct obs_source_frame *frame)
|
|
{
|
|
gs_texture_t tex = source->async_texture;
|
|
gs_texrender_t texrender = source->async_convert_texrender;
|
|
|
|
gs_texrender_reset(texrender);
|
|
|
|
upload_raw_frame(tex, frame);
|
|
|
|
uint32_t cx = source->async_width;
|
|
uint32_t cy = source->async_height;
|
|
|
|
float convert_width = (float)source->async_convert_width;
|
|
float convert_height = (float)source->async_convert_height;
|
|
|
|
gs_effect_t conv = obs->video.conversion_effect;
|
|
gs_technique_t tech = gs_effect_get_technique(conv,
|
|
select_conversion_technique(frame->format));
|
|
|
|
if (!gs_texrender_begin(texrender, cx, cy))
|
|
return false;
|
|
|
|
gs_technique_begin(tech);
|
|
gs_technique_begin_pass(tech, 0);
|
|
|
|
gs_effect_set_texture(gs_effect_get_param_by_name(conv, "image"), tex);
|
|
set_eparam(conv, "width", (float)cx);
|
|
set_eparam(conv, "height", (float)cy);
|
|
set_eparam(conv, "width_i", 1.0f / cx);
|
|
set_eparam(conv, "height_i", 1.0f / cy);
|
|
set_eparam(conv, "width_d2", cx * 0.5f);
|
|
set_eparam(conv, "height_d2", cy * 0.5f);
|
|
set_eparam(conv, "width_d2_i", 1.0f / (cx * 0.5f));
|
|
set_eparam(conv, "height_d2_i", 1.0f / (cy * 0.5f));
|
|
set_eparam(conv, "input_width", convert_width);
|
|
set_eparam(conv, "input_height", convert_height);
|
|
set_eparam(conv, "input_width_i", 1.0f / convert_width);
|
|
set_eparam(conv, "input_height_i", 1.0f / convert_height);
|
|
set_eparam(conv, "input_width_i_d2", (1.0f / convert_width) * 0.5f);
|
|
set_eparam(conv, "input_height_i_d2", (1.0f / convert_height) * 0.5f);
|
|
set_eparam(conv, "u_plane_offset",
|
|
(float)source->async_plane_offset[0]);
|
|
set_eparam(conv, "v_plane_offset",
|
|
(float)source->async_plane_offset[1]);
|
|
|
|
gs_ortho(0.f, (float)cx, 0.f, (float)cy, -100.f, 100.f);
|
|
|
|
gs_draw_sprite(tex, 0, cx, cy);
|
|
|
|
gs_technique_end_pass(tech);
|
|
gs_technique_end(tech);
|
|
|
|
gs_texrender_end(texrender);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool update_async_texture(struct obs_source *source,
|
|
const struct obs_source_frame *frame)
|
|
{
|
|
gs_texture_t tex = source->async_texture;
|
|
gs_texrender_t texrender = source->async_convert_texrender;
|
|
enum convert_type type = get_convert_type(frame->format);
|
|
uint8_t *ptr;
|
|
uint32_t linesize;
|
|
|
|
source->async_format = frame->format;
|
|
source->async_flip = frame->flip;
|
|
source->async_full_range = frame->full_range;
|
|
memcpy(source->async_color_matrix, frame->color_matrix,
|
|
sizeof(frame->color_matrix));
|
|
memcpy(source->async_color_range_min, frame->color_range_min,
|
|
sizeof frame->color_range_min);
|
|
memcpy(source->async_color_range_max, frame->color_range_max,
|
|
sizeof frame->color_range_max);
|
|
|
|
if (source->async_gpu_conversion && texrender)
|
|
return update_async_texrender(source, frame);
|
|
|
|
if (type == CONVERT_NONE) {
|
|
gs_texture_set_image(tex, frame->data[0], frame->linesize[0],
|
|
false);
|
|
return true;
|
|
}
|
|
|
|
if (!gs_texture_map(tex, &ptr, &linesize))
|
|
return false;
|
|
|
|
if (type == CONVERT_420)
|
|
decompress_420((const uint8_t* const*)frame->data,
|
|
frame->linesize,
|
|
0, frame->height, ptr, linesize);
|
|
|
|
else if (type == CONVERT_NV12)
|
|
decompress_nv12((const uint8_t* const*)frame->data,
|
|
frame->linesize,
|
|
0, frame->height, ptr, linesize);
|
|
|
|
else if (type == CONVERT_422_Y)
|
|
decompress_422(frame->data[0], frame->linesize[0],
|
|
0, frame->height, ptr, linesize, true);
|
|
|
|
else if (type == CONVERT_422_U)
|
|
decompress_422(frame->data[0], frame->linesize[0],
|
|
0, frame->height, ptr, linesize, false);
|
|
|
|
gs_texture_unmap(tex);
|
|
return true;
|
|
}
|
|
|
|
static inline void obs_source_draw_texture(struct obs_source *source,
|
|
gs_effect_t effect, float *color_matrix,
|
|
float const *color_range_min, float const *color_range_max)
|
|
{
|
|
gs_texture_t tex = source->async_texture;
|
|
gs_eparam_t param;
|
|
|
|
if (source->async_convert_texrender)
|
|
tex = gs_texrender_get_texture(source->async_convert_texrender);
|
|
|
|
if (color_range_min) {
|
|
size_t const size = sizeof(float) * 3;
|
|
param = gs_effect_get_param_by_name(effect, "color_range_min");
|
|
gs_effect_set_val(param, color_range_min, size);
|
|
}
|
|
|
|
if (color_range_max) {
|
|
size_t const size = sizeof(float) * 3;
|
|
param = gs_effect_get_param_by_name(effect, "color_range_max");
|
|
gs_effect_set_val(param, color_range_max, size);
|
|
}
|
|
|
|
if (color_matrix) {
|
|
param = gs_effect_get_param_by_name(effect, "color_matrix");
|
|
gs_effect_set_val(param, color_matrix, sizeof(float) * 16);
|
|
}
|
|
|
|
param = gs_effect_get_param_by_name(effect, "image");
|
|
gs_effect_set_texture(param, tex);
|
|
|
|
gs_draw_sprite(tex, source->async_flip ? GS_FLIP_V : 0, 0, 0);
|
|
}
|
|
|
|
static void obs_source_draw_async_texture(struct obs_source *source)
|
|
{
|
|
gs_effect_t effect = gs_get_effect();
|
|
bool yuv = format_is_yuv(source->async_format);
|
|
bool limited_range = yuv && !source->async_full_range;
|
|
const char *type = yuv ? "DrawMatrix" : "Draw";
|
|
bool def_draw = (!effect);
|
|
gs_technique_t tech = NULL;
|
|
|
|
if (def_draw) {
|
|
effect = obs_get_default_effect();
|
|
tech = gs_effect_get_technique(effect, type);
|
|
gs_technique_begin(tech);
|
|
gs_technique_begin_pass(tech, 0);
|
|
}
|
|
|
|
obs_source_draw_texture(source, effect,
|
|
yuv ? source->async_color_matrix : NULL,
|
|
limited_range ? source->async_color_range_min : NULL,
|
|
limited_range ? source->async_color_range_max : NULL);
|
|
|
|
if (def_draw) {
|
|
gs_technique_end_pass(tech);
|
|
gs_technique_end(tech);
|
|
}
|
|
}
|
|
|
|
static void obs_source_render_async_video(obs_source_t source)
|
|
{
|
|
if (!source->async_rendered) {
|
|
struct obs_source_frame *frame = obs_source_get_frame(source);
|
|
|
|
source->async_rendered = true;
|
|
if (frame) {
|
|
if (!set_async_texture_size(source, frame))
|
|
return;
|
|
if (!update_async_texture(source, frame))
|
|
return;
|
|
}
|
|
|
|
obs_source_release_frame(source, frame);
|
|
}
|
|
|
|
if (source->async_texture)
|
|
obs_source_draw_async_texture(source);
|
|
}
|
|
|
|
static inline void obs_source_render_filters(obs_source_t source)
|
|
{
|
|
source->rendering_filter = true;
|
|
obs_source_video_render(source->filters.array[0]);
|
|
source->rendering_filter = false;
|
|
}
|
|
|
|
static inline void obs_source_default_render(obs_source_t source,
|
|
bool color_matrix)
|
|
{
|
|
gs_effect_t effect = obs->video.default_effect;
|
|
const char *tech_name = color_matrix ? "DrawMatrix" : "Draw";
|
|
gs_technique_t tech = gs_effect_get_technique(effect, tech_name);
|
|
size_t passes, i;
|
|
|
|
passes = gs_technique_begin(tech);
|
|
for (i = 0; i < passes; i++) {
|
|
gs_technique_begin_pass(tech, i);
|
|
if (source->context.data)
|
|
source->info.video_render(source->context.data, effect);
|
|
gs_technique_end_pass(tech);
|
|
}
|
|
gs_technique_end(tech);
|
|
}
|
|
|
|
static inline void obs_source_main_render(obs_source_t source)
|
|
{
|
|
uint32_t flags = source->info.output_flags;
|
|
bool color_matrix = (flags & OBS_SOURCE_COLOR_MATRIX) != 0;
|
|
bool custom_draw = (flags & OBS_SOURCE_CUSTOM_DRAW) != 0;
|
|
bool default_effect = !source->filter_parent &&
|
|
source->filters.num == 0 &&
|
|
!custom_draw;
|
|
|
|
if (default_effect)
|
|
obs_source_default_render(source, color_matrix);
|
|
else if (source->context.data)
|
|
source->info.video_render(source->context.data,
|
|
custom_draw ? NULL : gs_get_effect());
|
|
}
|
|
|
|
void obs_source_video_render(obs_source_t source)
|
|
{
|
|
if (!source_valid(source)) return;
|
|
|
|
if (source->filters.num && !source->rendering_filter)
|
|
obs_source_render_filters(source);
|
|
|
|
else if (source->info.video_render)
|
|
obs_source_main_render(source);
|
|
|
|
else if (source->filter_target)
|
|
obs_source_video_render(source->filter_target);
|
|
|
|
else
|
|
obs_source_render_async_video(source);
|
|
}
|
|
|
|
uint32_t obs_source_get_width(obs_source_t source)
|
|
{
|
|
if (!source_valid(source)) return 0;
|
|
|
|
if (source->info.get_width)
|
|
return source->info.get_width(source->context.data);
|
|
return source->async_width;
|
|
}
|
|
|
|
uint32_t obs_source_get_height(obs_source_t source)
|
|
{
|
|
if (!source_valid(source)) return 0;
|
|
|
|
if (source->info.get_height)
|
|
return source->info.get_height(source->context.data);
|
|
return source->async_height;
|
|
}
|
|
|
|
obs_source_t obs_filter_get_parent(obs_source_t filter)
|
|
{
|
|
return filter ? filter->filter_parent : NULL;
|
|
}
|
|
|
|
obs_source_t obs_filter_get_target(obs_source_t filter)
|
|
{
|
|
return filter ? filter->filter_target : NULL;
|
|
}
|
|
|
|
void obs_source_filter_add(obs_source_t source, obs_source_t filter)
|
|
{
|
|
if (!source || !filter)
|
|
return;
|
|
|
|
pthread_mutex_lock(&source->filter_mutex);
|
|
|
|
if (da_find(source->filters, &filter, 0) != DARRAY_INVALID) {
|
|
blog(LOG_WARNING, "Tried to add a filter that was already "
|
|
"present on the source");
|
|
return;
|
|
}
|
|
|
|
if (source->filters.num) {
|
|
obs_source_t *back = da_end(source->filters);
|
|
(*back)->filter_target = filter;
|
|
}
|
|
|
|
da_push_back(source->filters, &filter);
|
|
|
|
pthread_mutex_unlock(&source->filter_mutex);
|
|
|
|
filter->filter_parent = source;
|
|
filter->filter_target = source;
|
|
}
|
|
|
|
void obs_source_filter_remove(obs_source_t source, obs_source_t filter)
|
|
{
|
|
size_t idx;
|
|
|
|
if (!source || !filter)
|
|
return;
|
|
|
|
pthread_mutex_lock(&source->filter_mutex);
|
|
|
|
idx = da_find(source->filters, &filter, 0);
|
|
if (idx == DARRAY_INVALID)
|
|
return;
|
|
|
|
if (idx > 0) {
|
|
obs_source_t prev = source->filters.array[idx-1];
|
|
prev->filter_target = filter->filter_target;
|
|
}
|
|
|
|
da_erase(source->filters, idx);
|
|
|
|
pthread_mutex_unlock(&source->filter_mutex);
|
|
|
|
filter->filter_parent = NULL;
|
|
filter->filter_target = NULL;
|
|
}
|
|
|
|
void obs_source_filter_set_order(obs_source_t source, obs_source_t filter,
|
|
enum obs_order_movement movement)
|
|
{
|
|
size_t idx, i;
|
|
|
|
if (!source || !filter)
|
|
return;
|
|
|
|
idx = da_find(source->filters, &filter, 0);
|
|
if (idx == DARRAY_INVALID)
|
|
return;
|
|
|
|
if (movement == OBS_ORDER_MOVE_UP) {
|
|
if (idx == source->filters.num-1)
|
|
return;
|
|
da_move_item(source->filters, idx, idx+1);
|
|
|
|
} else if (movement == OBS_ORDER_MOVE_DOWN) {
|
|
if (idx == 0)
|
|
return;
|
|
da_move_item(source->filters, idx, idx-1);
|
|
|
|
} else if (movement == OBS_ORDER_MOVE_TOP) {
|
|
if (idx == source->filters.num-1)
|
|
return;
|
|
da_move_item(source->filters, idx, source->filters.num-1);
|
|
|
|
} else if (movement == OBS_ORDER_MOVE_BOTTOM) {
|
|
if (idx == 0)
|
|
return;
|
|
da_move_item(source->filters, idx, 0);
|
|
}
|
|
|
|
/* reorder filter targets, not the nicest way of dealing with things */
|
|
for (i = 0; i < source->filters.num; i++) {
|
|
obs_source_t next_filter = (i == source->filters.num-1) ?
|
|
source : source->filters.array[idx+1];
|
|
source->filters.array[i]->filter_target = next_filter;
|
|
}
|
|
}
|
|
|
|
obs_data_t obs_source_get_settings(obs_source_t source)
|
|
{
|
|
if (!source) return NULL;
|
|
|
|
obs_data_addref(source->context.settings);
|
|
return source->context.settings;
|
|
}
|
|
|
|
static inline struct obs_source_frame *filter_async_video(obs_source_t source,
|
|
struct obs_source_frame *in)
|
|
{
|
|
size_t i;
|
|
for (i = source->filters.num; i > 0; i--) {
|
|
struct obs_source *filter = source->filters.array[i-1];
|
|
|
|
if (filter->context.data && filter->info.filter_video) {
|
|
in = filter->info.filter_video(filter->context.data,
|
|
in);
|
|
if (!in)
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return in;
|
|
}
|
|
|
|
static inline void copy_frame_data_line(struct obs_source_frame *dst,
|
|
const struct obs_source_frame *src, uint32_t plane, uint32_t y)
|
|
{
|
|
uint32_t pos_src = y * src->linesize[plane];
|
|
uint32_t pos_dst = y * dst->linesize[plane];
|
|
uint32_t bytes = dst->linesize[plane] < src->linesize[plane] ?
|
|
dst->linesize[plane] : src->linesize[plane];
|
|
|
|
memcpy(dst->data[plane] + pos_dst, src->data[plane] + pos_src, bytes);
|
|
}
|
|
|
|
static inline void copy_frame_data_plane(struct obs_source_frame *dst,
|
|
const struct obs_source_frame *src,
|
|
uint32_t plane, uint32_t lines)
|
|
{
|
|
if (dst->linesize[plane] != src->linesize[plane])
|
|
for (uint32_t y = 0; y < lines; y++)
|
|
copy_frame_data_line(dst, src, plane, y);
|
|
else
|
|
memcpy(dst->data[plane], src->data[plane],
|
|
dst->linesize[plane] * lines);
|
|
}
|
|
|
|
static void copy_frame_data(struct obs_source_frame *dst,
|
|
const struct obs_source_frame *src)
|
|
{
|
|
dst->flip = src->flip;
|
|
dst->full_range = src->full_range;
|
|
dst->timestamp = src->timestamp;
|
|
memcpy(dst->color_matrix, src->color_matrix, sizeof(float) * 16);
|
|
if (!dst->full_range) {
|
|
size_t const size = sizeof(float) * 3;
|
|
memcpy(dst->color_range_min, src->color_range_min, size);
|
|
memcpy(dst->color_range_max, src->color_range_max, size);
|
|
}
|
|
|
|
switch (dst->format) {
|
|
case VIDEO_FORMAT_I420:
|
|
copy_frame_data_plane(dst, src, 0, dst->height);
|
|
copy_frame_data_plane(dst, src, 1, dst->height/2);
|
|
copy_frame_data_plane(dst, src, 2, dst->height/2);
|
|
break;
|
|
|
|
case VIDEO_FORMAT_NV12:
|
|
copy_frame_data_plane(dst, src, 0, dst->height);
|
|
copy_frame_data_plane(dst, src, 1, dst->height/2);
|
|
break;
|
|
|
|
case VIDEO_FORMAT_YVYU:
|
|
case VIDEO_FORMAT_YUY2:
|
|
case VIDEO_FORMAT_UYVY:
|
|
case VIDEO_FORMAT_NONE:
|
|
case VIDEO_FORMAT_RGBA:
|
|
case VIDEO_FORMAT_BGRA:
|
|
case VIDEO_FORMAT_BGRX:
|
|
copy_frame_data_plane(dst, src, 0, dst->height);
|
|
}
|
|
}
|
|
|
|
static inline struct obs_source_frame *cache_video(
|
|
const struct obs_source_frame *frame)
|
|
{
|
|
/* TODO: use an actual cache */
|
|
struct obs_source_frame *new_frame = obs_source_frame_create(
|
|
frame->format, frame->width, frame->height);
|
|
|
|
copy_frame_data(new_frame, frame);
|
|
return new_frame;
|
|
}
|
|
|
|
static bool ready_async_frame(obs_source_t source, uint64_t sys_time);
|
|
|
|
static inline void cycle_frames(struct obs_source *source)
|
|
{
|
|
if (source->video_frames.num && !source->activate_refs)
|
|
ready_async_frame(source, os_gettime_ns());
|
|
}
|
|
|
|
void obs_source_output_video(obs_source_t source,
|
|
const struct obs_source_frame *frame)
|
|
{
|
|
if (!source || !frame)
|
|
return;
|
|
|
|
struct obs_source_frame *output = cache_video(frame);
|
|
|
|
pthread_mutex_lock(&source->filter_mutex);
|
|
output = filter_async_video(source, output);
|
|
pthread_mutex_unlock(&source->filter_mutex);
|
|
|
|
if (output) {
|
|
pthread_mutex_lock(&source->video_mutex);
|
|
cycle_frames(source);
|
|
da_push_back(source->video_frames, &output);
|
|
pthread_mutex_unlock(&source->video_mutex);
|
|
}
|
|
}
|
|
|
|
static inline struct obs_audio_data *filter_async_audio(obs_source_t source,
|
|
struct obs_audio_data *in)
|
|
{
|
|
size_t i;
|
|
for (i = source->filters.num; i > 0; i--) {
|
|
struct obs_source *filter = source->filters.array[i-1];
|
|
|
|
if (filter->context.data && filter->info.filter_audio) {
|
|
in = filter->info.filter_audio(filter->context.data,
|
|
in);
|
|
if (!in)
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return in;
|
|
}
|
|
|
|
static inline void reset_resampler(obs_source_t source,
|
|
const struct obs_source_audio *audio)
|
|
{
|
|
const struct audio_output_info *obs_info;
|
|
struct resample_info output_info;
|
|
|
|
obs_info = audio_output_get_info(obs->audio.audio);
|
|
|
|
output_info.format = obs_info->format;
|
|
output_info.samples_per_sec = obs_info->samples_per_sec;
|
|
output_info.speakers = obs_info->speakers;
|
|
|
|
source->sample_info.format = audio->format;
|
|
source->sample_info.samples_per_sec = audio->samples_per_sec;
|
|
source->sample_info.speakers = audio->speakers;
|
|
|
|
if (source->sample_info.samples_per_sec == obs_info->samples_per_sec &&
|
|
source->sample_info.format == obs_info->format &&
|
|
source->sample_info.speakers == obs_info->speakers) {
|
|
source->audio_failed = false;
|
|
return;
|
|
}
|
|
|
|
audio_resampler_destroy(source->resampler);
|
|
source->resampler = audio_resampler_create(&output_info,
|
|
&source->sample_info);
|
|
|
|
source->audio_failed = source->resampler == NULL;
|
|
if (source->resampler == NULL)
|
|
blog(LOG_ERROR, "creation of resampler failed");
|
|
}
|
|
|
|
static inline void copy_audio_data(obs_source_t source,
|
|
const uint8_t *const data[], uint32_t frames, uint64_t ts)
|
|
{
|
|
size_t planes = audio_output_get_planes(obs->audio.audio);
|
|
size_t blocksize = audio_output_get_block_size(obs->audio.audio);
|
|
size_t size = (size_t)frames * blocksize;
|
|
bool resize = source->audio_storage_size < size;
|
|
|
|
source->audio_data.frames = frames;
|
|
source->audio_data.timestamp = ts;
|
|
|
|
for (size_t i = 0; i < planes; i++) {
|
|
/* ensure audio storage capacity */
|
|
if (resize) {
|
|
bfree(source->audio_data.data[i]);
|
|
source->audio_data.data[i] = bmalloc(size);
|
|
}
|
|
|
|
memcpy(source->audio_data.data[i], data[i], size);
|
|
}
|
|
|
|
if (resize)
|
|
source->audio_storage_size = size;
|
|
}
|
|
|
|
/* resamples/remixes new audio to the designated main audio output format */
|
|
static void process_audio(obs_source_t source,
|
|
const struct obs_source_audio *audio)
|
|
{
|
|
if (source->sample_info.samples_per_sec != audio->samples_per_sec ||
|
|
source->sample_info.format != audio->format ||
|
|
source->sample_info.speakers != audio->speakers)
|
|
reset_resampler(source, audio);
|
|
|
|
if (source->audio_failed)
|
|
return;
|
|
|
|
if (source->resampler) {
|
|
uint8_t *output[MAX_AV_PLANES];
|
|
uint32_t frames;
|
|
uint64_t offset;
|
|
|
|
memset(output, 0, sizeof(output));
|
|
|
|
audio_resampler_resample(source->resampler,
|
|
output, &frames, &offset,
|
|
audio->data, audio->frames);
|
|
|
|
copy_audio_data(source, (const uint8_t *const *)output, frames,
|
|
audio->timestamp - offset);
|
|
} else {
|
|
copy_audio_data(source, audio->data, audio->frames,
|
|
audio->timestamp);
|
|
}
|
|
}
|
|
|
|
void obs_source_output_audio(obs_source_t source,
|
|
const struct obs_source_audio *audio)
|
|
{
|
|
uint32_t flags;
|
|
struct obs_audio_data *output;
|
|
|
|
if (!source || !audio)
|
|
return;
|
|
|
|
flags = source->info.output_flags;
|
|
process_audio(source, audio);
|
|
|
|
pthread_mutex_lock(&source->filter_mutex);
|
|
output = filter_async_audio(source, &source->audio_data);
|
|
|
|
if (output) {
|
|
bool async = (flags & OBS_SOURCE_ASYNC) != 0;
|
|
|
|
pthread_mutex_lock(&source->audio_mutex);
|
|
|
|
/* wait for video to start before outputting any audio so we
|
|
* have a base for sync */
|
|
if (source->timing_set || !async) {
|
|
struct audio_data data;
|
|
|
|
for (int i = 0; i < MAX_AV_PLANES; i++)
|
|
data.data[i] = output->data[i];
|
|
|
|
data.frames = output->frames;
|
|
data.timestamp = output->timestamp;
|
|
source_output_audio_line(source, &data);
|
|
}
|
|
|
|
pthread_mutex_unlock(&source->audio_mutex);
|
|
}
|
|
|
|
pthread_mutex_unlock(&source->filter_mutex);
|
|
}
|
|
|
|
static inline bool frame_out_of_bounds(obs_source_t source, uint64_t ts)
|
|
{
|
|
if (ts < source->last_frame_ts)
|
|
return ((source->last_frame_ts - ts) > MAX_TIMESTAMP_JUMP);
|
|
else
|
|
return ((ts - source->last_frame_ts) > MAX_TIMESTAMP_JUMP);
|
|
}
|
|
|
|
/* #define DEBUG_ASYNC_FRAMES 1 */
|
|
|
|
static bool ready_async_frame(obs_source_t source, uint64_t sys_time)
|
|
{
|
|
struct obs_source_frame *next_frame = source->video_frames.array[0];
|
|
struct obs_source_frame *frame = NULL;
|
|
uint64_t sys_offset = sys_time - source->last_sys_timestamp;
|
|
uint64_t frame_time = next_frame->timestamp;
|
|
uint64_t frame_offset = 0;
|
|
|
|
#if DEBUG_ASYNC_FRAMES
|
|
blog(LOG_DEBUG, "source->last_frame_ts: %llu, frame_time: %llu, "
|
|
"sys_offset: %llu, frame_offset: %llu, "
|
|
"number of frames: %lu",
|
|
source->last_frame_ts, frame_time, sys_offset,
|
|
frame_time - source->last_frame_ts,
|
|
(unsigned long)source->video_frames.num);
|
|
#endif
|
|
|
|
/* account for timestamp invalidation */
|
|
if (frame_out_of_bounds(source, frame_time)) {
|
|
#if DEBUG_ASYNC_FRAMES
|
|
blog(LOG_DEBUG, "timing jump");
|
|
#endif
|
|
source->last_frame_ts = next_frame->timestamp;
|
|
return true;
|
|
} else {
|
|
frame_offset = frame_time - source->last_frame_ts;
|
|
source->last_frame_ts += sys_offset;
|
|
}
|
|
|
|
while (source->last_frame_ts > next_frame->timestamp) {
|
|
|
|
/* this tries to reduce the needless frame duplication, also
|
|
* helps smooth out async rendering to frame boundaries. In
|
|
* other words, tries to keep the framerate as smooth as
|
|
* possible */
|
|
if ((source->last_frame_ts - next_frame->timestamp) < 1000000)
|
|
break;
|
|
|
|
if (frame)
|
|
da_erase(source->video_frames, 0);
|
|
|
|
#if DEBUG_ASYNC_FRAMES
|
|
blog(LOG_DEBUG, "new frame, "
|
|
"source->last_frame_ts: %llu, "
|
|
"next_frame->timestamp: %llu",
|
|
source->last_frame_ts,
|
|
next_frame->timestamp);
|
|
#endif
|
|
|
|
obs_source_frame_destroy(frame);
|
|
|
|
if (source->video_frames.num == 1)
|
|
return true;
|
|
|
|
frame = next_frame;
|
|
next_frame = source->video_frames.array[1];
|
|
|
|
/* more timestamp checking and compensating */
|
|
if ((next_frame->timestamp - frame_time) > MAX_TIMESTAMP_JUMP) {
|
|
#if DEBUG_ASYNC_FRAMES
|
|
blog(LOG_DEBUG, "timing jump");
|
|
#endif
|
|
source->last_frame_ts =
|
|
next_frame->timestamp - frame_offset;
|
|
}
|
|
|
|
frame_time = next_frame->timestamp;
|
|
frame_offset = frame_time - source->last_frame_ts;
|
|
}
|
|
|
|
#if DEBUG_ASYNC_FRAMES
|
|
if (!frame)
|
|
blog(LOG_DEBUG, "no frame!");
|
|
#endif
|
|
|
|
return frame != NULL;
|
|
}
|
|
|
|
static inline struct obs_source_frame *get_closest_frame(obs_source_t source,
|
|
uint64_t sys_time)
|
|
{
|
|
if (ready_async_frame(source, sys_time)) {
|
|
struct obs_source_frame *frame = source->video_frames.array[0];
|
|
da_erase(source->video_frames, 0);
|
|
return frame;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Ensures that cached frames are displayed on time. If multiple frames
|
|
* were cached between renders, then releases the unnecessary frames and uses
|
|
* the frame with the closest timing to ensure sync. Also ensures that timing
|
|
* with audio is synchronized.
|
|
*/
|
|
struct obs_source_frame *obs_source_get_frame(obs_source_t source)
|
|
{
|
|
struct obs_source_frame *frame = NULL;
|
|
uint64_t sys_time;
|
|
|
|
if (!source)
|
|
return NULL;
|
|
|
|
pthread_mutex_lock(&source->video_mutex);
|
|
|
|
sys_time = os_gettime_ns();
|
|
|
|
if (!source->video_frames.num)
|
|
goto unlock;
|
|
|
|
if (!source->last_frame_ts) {
|
|
frame = source->video_frames.array[0];
|
|
da_erase(source->video_frames, 0);
|
|
|
|
source->last_frame_ts = frame->timestamp;
|
|
} else {
|
|
frame = get_closest_frame(source, sys_time);
|
|
}
|
|
|
|
/* reset timing to current system time */
|
|
if (frame) {
|
|
uint64_t min_expected_sys_ts =
|
|
frame->timestamp + source->timing_adjust;
|
|
|
|
source->timing_adjust = sys_time - frame->timestamp;
|
|
source->timing_set = true;
|
|
}
|
|
|
|
unlock:
|
|
source->last_sys_timestamp = sys_time;
|
|
|
|
pthread_mutex_unlock(&source->video_mutex);
|
|
|
|
if (frame)
|
|
obs_source_addref(source);
|
|
|
|
return frame;
|
|
}
|
|
|
|
void obs_source_release_frame(obs_source_t source,
|
|
struct obs_source_frame *frame)
|
|
{
|
|
if (source && frame) {
|
|
obs_source_frame_destroy(frame);
|
|
obs_source_release(source);
|
|
}
|
|
}
|
|
|
|
const char *obs_source_get_name(obs_source_t source)
|
|
{
|
|
return source ? source->context.name : NULL;
|
|
}
|
|
|
|
void obs_source_set_name(obs_source_t source, const char *name)
|
|
{
|
|
if (!source) return;
|
|
|
|
if (!name || !*name || strcmp(name, source->context.name) != 0) {
|
|
struct calldata data;
|
|
char *prev_name = bstrdup(source->context.name);
|
|
obs_context_data_setname(&source->context, name);
|
|
|
|
calldata_init(&data);
|
|
calldata_set_ptr(&data, "source", source);
|
|
calldata_set_string(&data, "new_name", source->context.name);
|
|
calldata_set_string(&data, "prev_name", prev_name);
|
|
signal_handler_signal(obs->signals, "source_rename", &data);
|
|
signal_handler_signal(source->context.signals, "rename", &data);
|
|
calldata_free(&data);
|
|
bfree(prev_name);
|
|
}
|
|
}
|
|
|
|
enum obs_source_type obs_source_get_type(obs_source_t source)
|
|
{
|
|
return source ? source->info.type : OBS_SOURCE_TYPE_INPUT;
|
|
}
|
|
|
|
const char *obs_source_get_id(obs_source_t source)
|
|
{
|
|
return source ? source->info.id : NULL;
|
|
}
|
|
|
|
static inline void render_filter_bypass(obs_source_t target, gs_effect_t effect,
|
|
bool use_matrix)
|
|
{
|
|
const char *tech_name = use_matrix ? "DrawMatrix" : "Draw";
|
|
gs_technique_t tech = gs_effect_get_technique(effect, tech_name);
|
|
size_t passes, i;
|
|
|
|
passes = gs_technique_begin(tech);
|
|
for (i = 0; i < passes; i++) {
|
|
gs_technique_begin_pass(tech, i);
|
|
obs_source_video_render(target);
|
|
gs_technique_end_pass(tech);
|
|
}
|
|
gs_technique_end(tech);
|
|
}
|
|
|
|
static inline void render_filter_tex(gs_texture_t tex, gs_effect_t effect,
|
|
uint32_t width, uint32_t height, bool use_matrix)
|
|
{
|
|
const char *tech_name = use_matrix ? "DrawMatrix" : "Draw";
|
|
gs_technique_t tech = gs_effect_get_technique(effect, tech_name);
|
|
gs_eparam_t image = gs_effect_get_param_by_name(effect, "image");
|
|
size_t passes, i;
|
|
|
|
gs_effect_set_texture(image, tex);
|
|
|
|
passes = gs_technique_begin(tech);
|
|
for (i = 0; i < passes; i++) {
|
|
gs_technique_begin_pass(tech, i);
|
|
gs_draw_sprite(tex, width, height, 0);
|
|
gs_technique_end_pass(tech);
|
|
}
|
|
gs_technique_end(tech);
|
|
}
|
|
|
|
void obs_source_process_filter(obs_source_t filter, gs_effect_t effect,
|
|
uint32_t width, uint32_t height, enum gs_color_format format,
|
|
enum obs_allow_direct_render allow_direct)
|
|
{
|
|
obs_source_t target, parent;
|
|
uint32_t target_flags, parent_flags;
|
|
int cx, cy;
|
|
bool use_matrix, expects_def, can_directly;
|
|
|
|
if (!filter) return;
|
|
|
|
target = obs_filter_get_target(filter);
|
|
parent = obs_filter_get_parent(filter);
|
|
target_flags = target->info.output_flags;
|
|
parent_flags = parent->info.output_flags;
|
|
cx = obs_source_get_width(target);
|
|
cy = obs_source_get_height(target);
|
|
use_matrix = !!(target_flags & OBS_SOURCE_COLOR_MATRIX);
|
|
expects_def = !(parent_flags & OBS_SOURCE_CUSTOM_DRAW);
|
|
can_directly = allow_direct == OBS_ALLOW_DIRECT_RENDERING;
|
|
|
|
/* if the parent does not use any custom effects, and this is the last
|
|
* filter in the chain for the parent, then render the parent directly
|
|
* using the filter effect instead of rendering to texture to reduce
|
|
* the total number of passes */
|
|
if (can_directly && expects_def && target == parent) {
|
|
render_filter_bypass(target, effect, use_matrix);
|
|
return;
|
|
}
|
|
|
|
if (!filter->filter_texrender)
|
|
filter->filter_texrender = gs_texrender_create(format,
|
|
GS_ZS_NONE);
|
|
|
|
if (gs_texrender_begin(filter->filter_texrender, cx, cy)) {
|
|
gs_ortho(0.0f, (float)cx, 0.0f, (float)cy, -100.0f, 100.0f);
|
|
if (expects_def && parent == target)
|
|
obs_source_default_render(parent, use_matrix);
|
|
else
|
|
obs_source_video_render(target);
|
|
gs_texrender_end(filter->filter_texrender);
|
|
}
|
|
|
|
/* --------------------------- */
|
|
|
|
render_filter_tex(gs_texrender_get_texture(filter->filter_texrender),
|
|
effect, width, height, use_matrix);
|
|
}
|
|
|
|
signal_handler_t obs_source_get_signal_handler(obs_source_t source)
|
|
{
|
|
return source ? source->context.signals : NULL;
|
|
}
|
|
|
|
proc_handler_t obs_source_get_proc_handler(obs_source_t source)
|
|
{
|
|
return source ? source->context.procs : NULL;
|
|
}
|
|
|
|
void obs_source_set_volume(obs_source_t source, float volume)
|
|
{
|
|
if (source) {
|
|
struct calldata data = {0};
|
|
calldata_set_ptr(&data, "source", source);
|
|
calldata_set_float(&data, "volume", volume);
|
|
|
|
signal_handler_signal(source->context.signals, "volume", &data);
|
|
signal_handler_signal(obs->signals, "source_volume", &data);
|
|
|
|
volume = (float)calldata_float(&data, "volume");
|
|
calldata_free(&data);
|
|
|
|
source->user_volume = volume;
|
|
}
|
|
}
|
|
|
|
static void set_tree_preset_vol(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
float *vol = param;
|
|
child->present_volume = *vol;
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
}
|
|
|
|
void obs_source_set_present_volume(obs_source_t source, float volume)
|
|
{
|
|
if (source) {
|
|
source->present_volume = volume;
|
|
|
|
/* don't set the presentation volume of the tree if a
|
|
* transition source, let the transition handle presentation
|
|
* volume for the child sources itself. */
|
|
if (source->info.type != OBS_SOURCE_TYPE_TRANSITION)
|
|
obs_source_enum_tree(source, set_tree_preset_vol,
|
|
&volume);
|
|
}
|
|
}
|
|
|
|
float obs_source_get_volume(obs_source_t source)
|
|
{
|
|
return source ? source->user_volume : 0.0f;
|
|
}
|
|
|
|
float obs_source_get_present_volume(obs_source_t source)
|
|
{
|
|
return source ? source->present_volume : 0.0f;
|
|
}
|
|
|
|
void obs_source_set_sync_offset(obs_source_t source, int64_t offset)
|
|
{
|
|
if (source)
|
|
source->sync_offset = offset;
|
|
}
|
|
|
|
int64_t obs_source_get_sync_offset(obs_source_t source)
|
|
{
|
|
return source ? source->sync_offset : 0;
|
|
}
|
|
|
|
struct source_enum_data {
|
|
obs_source_enum_proc_t enum_callback;
|
|
void *param;
|
|
};
|
|
|
|
static void enum_source_tree_callback(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
struct source_enum_data *data = param;
|
|
|
|
if (child->info.enum_sources && !child->enum_refs) {
|
|
os_atomic_inc_long(&child->enum_refs);
|
|
|
|
if (child->context.data)
|
|
child->info.enum_sources(child->context.data,
|
|
enum_source_tree_callback, data);
|
|
|
|
os_atomic_dec_long(&child->enum_refs);
|
|
}
|
|
|
|
data->enum_callback(parent, child, data->param);
|
|
}
|
|
|
|
void obs_source_enum_sources(obs_source_t source,
|
|
obs_source_enum_proc_t enum_callback,
|
|
void *param)
|
|
{
|
|
if (!source_valid(source) ||
|
|
!source->info.enum_sources ||
|
|
source->enum_refs)
|
|
return;
|
|
|
|
obs_source_addref(source);
|
|
|
|
os_atomic_inc_long(&source->enum_refs);
|
|
source->info.enum_sources(source->context.data, enum_callback, param);
|
|
os_atomic_dec_long(&source->enum_refs);
|
|
|
|
obs_source_release(source);
|
|
}
|
|
|
|
void obs_source_enum_tree(obs_source_t source,
|
|
obs_source_enum_proc_t enum_callback,
|
|
void *param)
|
|
{
|
|
struct source_enum_data data = {enum_callback, param};
|
|
|
|
if (!source_valid(source) ||
|
|
!source->info.enum_sources ||
|
|
source->enum_refs)
|
|
return;
|
|
|
|
obs_source_addref(source);
|
|
|
|
os_atomic_inc_long(&source->enum_refs);
|
|
source->info.enum_sources(source->context.data,
|
|
enum_source_tree_callback,
|
|
&data);
|
|
os_atomic_dec_long(&source->enum_refs);
|
|
|
|
obs_source_release(source);
|
|
}
|
|
|
|
void obs_source_add_child(obs_source_t parent, obs_source_t child)
|
|
{
|
|
if (!parent || !child) return;
|
|
|
|
for (int i = 0; i < parent->show_refs; i++) {
|
|
enum view_type type;
|
|
type = (i < parent->activate_refs) ? MAIN_VIEW : AUX_VIEW;
|
|
obs_source_activate(child, type);
|
|
}
|
|
}
|
|
|
|
void obs_source_remove_child(obs_source_t parent, obs_source_t child)
|
|
{
|
|
if (!parent || !child) return;
|
|
|
|
for (int i = 0; i < parent->show_refs; i++) {
|
|
enum view_type type;
|
|
type = (i < parent->activate_refs) ? MAIN_VIEW : AUX_VIEW;
|
|
obs_source_deactivate(child, type);
|
|
}
|
|
}
|
|
|
|
static void reset_transition_vol(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
child->transition_volume = 0.0f;
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
static void add_transition_vol(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
float *vol = param;
|
|
child->transition_volume += *vol;
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
}
|
|
|
|
static void apply_transition_vol(obs_source_t parent, obs_source_t child,
|
|
void *param)
|
|
{
|
|
child->present_volume = child->transition_volume;
|
|
|
|
UNUSED_PARAMETER(parent);
|
|
UNUSED_PARAMETER(param);
|
|
}
|
|
|
|
void obs_transition_begin_frame(obs_source_t transition)
|
|
{
|
|
if (!transition) return;
|
|
obs_source_enum_tree(transition, reset_transition_vol, NULL);
|
|
}
|
|
|
|
void obs_source_set_transition_vol(obs_source_t source, float vol)
|
|
{
|
|
if (!source) return;
|
|
|
|
add_transition_vol(NULL, source, &vol);
|
|
obs_source_enum_tree(source, add_transition_vol, &vol);
|
|
}
|
|
|
|
void obs_transition_end_frame(obs_source_t transition)
|
|
{
|
|
if (!transition) return;
|
|
obs_source_enum_tree(transition, apply_transition_vol, NULL);
|
|
}
|
|
|
|
void obs_source_save(obs_source_t source)
|
|
{
|
|
if (!source_valid(source) || !source->info.save) return;
|
|
source->info.save(source->context.data, source->context.settings);
|
|
}
|
|
|
|
void obs_source_load(obs_source_t source)
|
|
{
|
|
if (!source_valid(source) || !source->info.load) return;
|
|
source->info.load(source->context.data, source->context.settings);
|
|
}
|