#import #import #import #import #include #include #include #import "AVCaptureInputPort+PreMavericksCompat.h" #define MILLI_TIMESCALE 1000 #define MICRO_TIMESCALE (MILLI_TIMESCALE * 1000) #define NANO_TIMESCALE (MICRO_TIMESCALE * 1000) #define AV_REV_FOURCC(x) \ (x & 255), ((x >> 8) & 255), ((x >> 16) & 255), (x >> 24) struct av_capture; #define AVLOG(level, format, ...) \ blog(level, "%s: " format, \ obs_source_getname(capture->source), ##__VA_ARGS__) #define AVFREE(x) {[x release]; x = nil;} @interface OBSAVCaptureDelegate : NSObject { @public struct av_capture *capture; } - (void)captureOutput:(AVCaptureOutput *)out didDropSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection; - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection; @end struct av_capture { AVCaptureSession *session; AVCaptureDevice *device; AVCaptureDeviceInput *device_input; AVCaptureVideoDataOutput *out; OBSAVCaptureDelegate *delegate; dispatch_queue_t queue; bool has_clock; NSString *uid; id connect_observer; id disconnect_observer; unsigned fourcc; enum video_format video_format; obs_source_t source; struct source_frame frame; }; static inline void update_frame_size(struct av_capture *capture, struct source_frame *frame, uint32_t width, uint32_t height) { if (width != frame->width) { AVLOG(LOG_DEBUG, "Changed width from %d to %d", frame->width, width); frame->width = width; frame->linesize[0] = width * 2; } if (height != frame->height) { AVLOG(LOG_DEBUG, "Changed height from %d to %d", frame->height, height); frame->height = height; } } @implementation OBSAVCaptureDelegate - (void)captureOutput:(AVCaptureOutput *)out didDropSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection { UNUSED_PARAMETER(out); UNUSED_PARAMETER(sampleBuffer); UNUSED_PARAMETER(connection); } - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection { UNUSED_PARAMETER(captureOutput); UNUSED_PARAMETER(connection); CMItemCount count = CMSampleBufferGetNumSamples(sampleBuffer); if (count < 1 || !capture) return; struct source_frame *frame = &capture->frame; CVImageBufferRef img = CMSampleBufferGetImageBuffer(sampleBuffer); update_frame_size(capture, frame, CVPixelBufferGetWidth(img), CVPixelBufferGetHeight(img)); CMSampleTimingInfo info; CMSampleBufferGetSampleTimingInfo(sampleBuffer, 0, &info); CMTime target_pts; if (capture->has_clock) { AVCaptureInputPort *port = capture->device_input.ports[0]; target_pts = CMSyncConvertTime(info.presentationTimeStamp, port.clock, CMClockGetHostTimeClock()); } else { target_pts = info.presentationTimeStamp; } CMTime target_pts_nano = CMTimeConvertScale(target_pts, NANO_TIMESCALE, kCMTimeRoundingMethod_Default); frame->timestamp = target_pts_nano.value; CVPixelBufferLockBaseAddress(img, kCVPixelBufferLock_ReadOnly); frame->data[0] = CVPixelBufferGetBaseAddress(img); obs_source_output_video(capture->source, frame); CVPixelBufferUnlockBaseAddress(img, kCVPixelBufferLock_ReadOnly); } @end static const char *av_capture_getname(void) { /* TODO: locale */ return "Video Capture Device"; } static void remove_device(struct av_capture *capture) { [capture->session stopRunning]; [capture->session removeInput:capture->device_input]; AVFREE(capture->device_input); AVFREE(capture->device); } static void av_capture_destroy(void *data) { struct av_capture *capture = data; if (!capture) return; NSNotificationCenter *nc = [NSNotificationCenter defaultCenter]; [nc removeObserver:capture->connect_observer]; [nc removeObserver:capture->disconnect_observer]; remove_device(capture); AVFREE(capture->out); if (capture->queue) dispatch_release(capture->queue); AVFREE(capture->delegate); AVFREE(capture->session); AVFREE(capture->uid); bfree(capture); } static NSString *get_string(obs_data_t data, char const *name) { return @(obs_data_getstring(data, name)); } static bool init_session(struct av_capture *capture) { capture->session = [[AVCaptureSession alloc] init]; if (!capture->session) { AVLOG(LOG_ERROR, "Could not create AVCaptureSession"); goto error; } capture->delegate = [[OBSAVCaptureDelegate alloc] init]; if (!capture->delegate) { AVLOG(LOG_ERROR, "Could not create OBSAVCaptureDelegate"); goto error; } capture->delegate->capture = capture; capture->out = [[AVCaptureVideoDataOutput alloc] init]; if (!capture->out) { AVLOG(LOG_ERROR, "Could not create AVCaptureVideoDataOutput"); goto error; } capture->queue = dispatch_queue_create(NULL, NULL); if (!capture->queue) { AVLOG(LOG_ERROR, "Could not create dispatch queue"); goto error; } [capture->session addOutput:capture->out]; [capture->out setSampleBufferDelegate:capture->delegate queue:capture->queue]; return true; error: AVFREE(capture->session); AVFREE(capture->delegate); AVFREE(capture->out); return false; } static bool init_device_input(struct av_capture *capture) { NSError *err = nil; capture->device_input = [[AVCaptureDeviceInput deviceInputWithDevice:capture->device error:&err] retain]; if (!capture->device_input) { AVLOG(LOG_ERROR, "Error while initializing device input: %s", err.localizedFailureReason.UTF8String); return false; } [capture->session addInput:capture->device_input]; return true; } static uint32_t uint_from_dict(NSDictionary *dict, CFStringRef key) { return ((NSNumber*)dict[(__bridge NSString*)key]).unsignedIntValue; } static bool init_format(struct av_capture *capture) { AVCaptureDeviceFormat *format = capture->device.activeFormat; CMMediaType mtype = CMFormatDescriptionGetMediaType( format.formatDescription); // TODO: support other media types if (mtype != kCMMediaType_Video) { AVLOG(LOG_ERROR, "CMMediaType '%c%c%c%c' is unsupported", AV_REV_FOURCC(mtype)); return false; } capture->out.videoSettings = nil; capture->fourcc = htonl(uint_from_dict(capture->out.videoSettings, kCVPixelBufferPixelFormatTypeKey)); capture->video_format = video_format_from_fourcc(capture->fourcc); if (capture->video_format == VIDEO_FORMAT_NONE) { AVLOG(LOG_ERROR, "FourCC '%c%c%c%c' unsupported by libobs", AV_REV_FOURCC(capture->fourcc)); return false; } AVLOG(LOG_DEBUG, "Using FourCC '%c%c%c%c'", AV_REV_FOURCC(capture->fourcc)); return true; } static bool init_frame(struct av_capture *capture) { AVCaptureDeviceFormat *format = capture->device.activeFormat; CMVideoDimensions size = CMVideoFormatDescriptionGetDimensions( format.formatDescription); capture->frame.width = size.width; capture->frame.linesize[0] = size.width * 2; capture->frame.height = size.height; capture->frame.format = capture->video_format; capture->frame.full_range = false; NSDictionary *exts = (__bridge NSDictionary*)CMFormatDescriptionGetExtensions( format.formatDescription); capture->frame.linesize[0] = uint_from_dict(exts, (__bridge CFStringRef)@"CVBytesPerRow"); NSString *matrix_key = (__bridge NSString*)kCVImageBufferYCbCrMatrixKey; enum video_colorspace colorspace = exts[matrix_key] == (id)kCVImageBufferYCbCrMatrix_ITU_R_709_2 ? VIDEO_CS_709 : VIDEO_CS_601; if (!video_format_get_parameters(colorspace, VIDEO_RANGE_PARTIAL, capture->frame.color_matrix, capture->frame.color_range_min, capture->frame.color_range_max)) { AVLOG(LOG_ERROR, "Failed to get video format parameters for " "video format %u", colorspace); return false; } return true; } static NSArray *presets(void); static NSString *preset_names(NSString *preset); static NSString *select_preset(AVCaptureDevice *dev, NSString *cur_preset) { NSString *new_preset = nil; bool found_previous_preset = false; for (NSString *preset in presets().reverseObjectEnumerator) { if (!found_previous_preset) found_previous_preset = [cur_preset isEqualToString:preset]; if (![dev supportsAVCaptureSessionPreset:preset]) continue; if (!new_preset || !found_previous_preset) new_preset = preset; } return new_preset; } static void capture_device(struct av_capture *capture, AVCaptureDevice *dev, obs_data_t settings) { capture->device = dev; AVLOG(LOG_INFO, "Selected device '%s' %p", capture->device.localizedName.UTF8String, capture); if (obs_data_getbool(settings, "use_preset")) { NSString *preset = get_string(settings, "preset"); if (![dev supportsAVCaptureSessionPreset:preset]) { AVLOG(LOG_ERROR, "Preset %s not available", preset_names(preset).UTF8String); preset = select_preset(dev, preset); } if (!preset) { AVLOG(LOG_ERROR, "Could not select a preset, " "initialization failed"); return; } capture->session.sessionPreset = preset; AVLOG(LOG_INFO, "Using preset %s", preset_names(preset).UTF8String); } if (!init_device_input(capture)) goto error_input; AVCaptureInputPort *port = capture->device_input.ports[0]; capture->has_clock = [port respondsToSelector:@selector(clock)]; if (!init_format(capture)) goto error; if (!init_frame(capture)) goto error; [capture->session startRunning]; return; error: [capture->session removeInput:capture->device_input]; AVFREE(capture->device_input); error_input: AVFREE(capture->device); } static inline void handle_disconnect(struct av_capture* capture, AVCaptureDevice *dev) { if (!dev) return; if (![dev.uniqueID isEqualTo:capture->uid]) return; if (!capture->device) { AVLOG(LOG_ERROR, "Received disconnect for unused device '%s'", capture->uid.UTF8String); return; } AVLOG(LOG_ERROR, "Device with unique ID '%s' disconnected", dev.uniqueID.UTF8String); remove_device(capture); } static inline void handle_connect(struct av_capture *capture, AVCaptureDevice *dev, obs_data_t settings) { if (!dev) return; if (![dev.uniqueID isEqualTo:capture->uid]) return; if (capture->device) { AVLOG(LOG_ERROR, "Received connect for in-use device '%s'", capture->uid.UTF8String); return; } AVLOG(LOG_INFO, "Device with unique ID '%s' connected, " "resuming capture", dev.uniqueID.UTF8String); capture_device(capture, [dev retain], settings); } static void av_capture_init(struct av_capture *capture, obs_data_t settings) { if (!init_session(capture)) return; capture->uid = [get_string(settings, "device") retain]; NSNotificationCenter *nc = [NSNotificationCenter defaultCenter]; capture->disconnect_observer = [nc addObserverForName:AVCaptureDeviceWasDisconnectedNotification object:nil queue:[NSOperationQueue mainQueue] usingBlock:^(NSNotification *note) { handle_disconnect(capture, note.object); } ]; capture->connect_observer = [nc addObserverForName:AVCaptureDeviceWasConnectedNotification object:nil queue:[NSOperationQueue mainQueue] usingBlock:^(NSNotification *note) { handle_connect(capture, note.object, settings); } ]; AVCaptureDevice *dev = [[AVCaptureDevice deviceWithUniqueID:capture->uid] retain]; if (!dev) { if (capture->uid.length < 1) AVLOG(LOG_ERROR, "No device selected"); else AVLOG(LOG_ERROR, "Could not initialize device " \ "with unique ID '%s'", capture->uid.UTF8String); return; } capture_device(capture, dev, settings); } static void *av_capture_create(obs_data_t settings, obs_source_t source) { UNUSED_PARAMETER(source); struct av_capture *capture = bzalloc(sizeof(struct av_capture)); capture->source = source; av_capture_init(capture, settings); if (!capture->session) { AVLOG(LOG_ERROR, "No valid session, returning NULL context"); av_capture_destroy(capture); bfree(capture); capture = NULL; } return capture; } static NSArray *presets(void) { return @[ //AVCaptureSessionPresetiFrame1280x720, //AVCaptureSessionPresetiFrame960x540, AVCaptureSessionPreset1280x720, AVCaptureSessionPreset960x540, AVCaptureSessionPreset640x480, AVCaptureSessionPreset352x288, AVCaptureSessionPreset320x240, //AVCaptureSessionPresetHigh, //AVCaptureSessionPresetMedium, //AVCaptureSessionPresetLow, //AVCaptureSessionPresetPhoto, ]; } static NSString *preset_names(NSString *preset) { NSDictionary *preset_names = @{ AVCaptureSessionPresetLow:@"Low", AVCaptureSessionPresetMedium:@"Medium", AVCaptureSessionPresetHigh:@"High", AVCaptureSessionPreset320x240:@"320x240", AVCaptureSessionPreset352x288:@"352x288", AVCaptureSessionPreset640x480:@"640x480", AVCaptureSessionPreset960x540:@"960x540", AVCaptureSessionPreset1280x720:@"1280x720", }; NSString *name = preset_names[preset]; if (name) return name; return [NSString stringWithFormat:@"Unknown (%@)", preset]; } static void av_capture_defaults(obs_data_t settings) { AVCaptureDevice *dev = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo]; if (!dev) return; NSString *highest = nil; for (NSString *preset in presets()) { if (![dev supportsAVCaptureSessionPreset:preset]) continue; highest = preset; } if (!highest) return; obs_data_set_default_string(settings, "device", dev.uniqueID.UTF8String); obs_data_set_default_bool(settings, "use_preset", true); obs_data_set_default_string(settings, "preset", highest.UTF8String); } static void fill_presets(AVCaptureDevice *dev, obs_property_t list, NSString *current_preset) { obs_property_list_clear(list); bool preset_found = false; for (NSString *preset in presets()) { bool supported = [dev supportsAVCaptureSessionPreset:preset]; bool is_current = [preset isEqualToString:current_preset]; if (is_current) preset_found = true; if (!supported && !is_current) continue; size_t idx = obs_property_list_add_string(list, preset_names(preset).UTF8String, preset.UTF8String); obs_property_list_item_disable(list, idx, !supported); } if (preset_found) return; size_t idx = obs_property_list_add_string(list, preset_names(current_preset).UTF8String, current_preset.UTF8String); obs_property_list_item_disable(list, idx, true); } static bool check_preset(AVCaptureDevice *dev, obs_property_t list, obs_data_t settings) { NSString *current_preset = get_string(settings, "preset"); size_t size = obs_property_list_item_count(list); NSMutableSet *listed = [NSMutableSet setWithCapacity:size]; for (size_t i = 0; i < size; i++) [listed addObject:@(obs_property_list_item_string(list, i))]; bool presets_changed = false; for (NSString *preset in presets()) { bool is_listed = [listed member:preset] != nil; bool supported = [dev supportsAVCaptureSessionPreset:preset]; if (supported == is_listed) continue; presets_changed = true; } if (!presets_changed && [listed member:current_preset] != nil) return false; fill_presets(dev, list, current_preset); return true; } static bool properties_device_changed(obs_properties_t props, obs_property_t p, obs_data_t settings) { UNUSED_PARAMETER(p); NSString *uid = get_string(settings, "device"); AVCaptureDevice *dev = [AVCaptureDevice deviceWithUniqueID:uid]; if (!dev) return false; return check_preset(dev, obs_properties_get(props, "preset"), settings); } static bool properties_preset_changed(obs_properties_t props, obs_property_t p, obs_data_t settings) { UNUSED_PARAMETER(props); NSString *uid = get_string(settings, "device"); AVCaptureDevice *dev = [AVCaptureDevice deviceWithUniqueID:uid]; if (!dev) return false; return check_preset(dev, p, settings); } static obs_properties_t av_capture_properties(void) { obs_properties_t props = obs_properties_create(); /* TODO: locale */ obs_property_t dev_list = obs_properties_add_list(props, "device", "Device", OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_STRING); for (AVCaptureDevice *dev in [AVCaptureDevice devicesWithMediaType:AVMediaTypeVideo]) { obs_property_list_add_string(dev_list, dev.localizedName.UTF8String, dev.uniqueID.UTF8String); } obs_property_set_modified_callback(dev_list, properties_device_changed); obs_property_t use_preset = obs_properties_add_bool(props, "use_preset", "Use preset"); // TODO: implement manual configuration obs_property_set_enabled(use_preset, false); obs_property_t preset_list = obs_properties_add_list(props, "preset", "Preset", OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_STRING); for (NSString *preset in presets()) obs_property_list_add_string(preset_list, preset_names(preset).UTF8String, preset.UTF8String); obs_property_set_modified_callback(preset_list, properties_preset_changed); return props; } static void switch_device(struct av_capture *capture, NSString *uid, obs_data_t settings) { if (!uid) return; if (capture->device) remove_device(capture); AVFREE(capture->uid); capture->uid = [uid retain]; AVCaptureDevice *dev = [AVCaptureDevice deviceWithUniqueID:uid]; if (!dev) { AVLOG(LOG_ERROR, "Device with unique id '%s' not found", uid.UTF8String); return; } capture_device(capture, [dev retain], settings); } static void av_capture_update(void *data, obs_data_t settings) { struct av_capture *capture = data; NSString *uid = get_string(settings, "device"); if (!capture->device || ![capture->device.uniqueID isEqualToString:uid]) return switch_device(capture, uid, settings); NSString *preset = get_string(settings, "preset"); NSLog(@"%@", preset); if (![capture->device supportsAVCaptureSessionPreset:preset]) { AVLOG(LOG_ERROR, "Preset %s not available", preset.UTF8String); preset = select_preset(capture->device, preset); } capture->session.sessionPreset = preset; AVLOG(LOG_INFO, "Selected preset %s", preset.UTF8String); } struct obs_source_info av_capture_info = { .id = "av_capture_input", .type = OBS_SOURCE_TYPE_INPUT, .output_flags = OBS_SOURCE_ASYNC_VIDEO, .getname = av_capture_getname, .create = av_capture_create, .destroy = av_capture_destroy, .defaults = av_capture_defaults, .properties = av_capture_properties, .update = av_capture_update, };