#import #import #import #import #include #include #include #import "AVCaptureInputPort+PreMavericksCompat.h" #define MILLI_TIMESCALE 1000 #define MICRO_TIMESCALE (MILLI_TIMESCALE * 1000) #define NANO_TIMESCALE (MICRO_TIMESCALE * 1000) #define AV_REV_FOURCC(x) \ (x & 255), ((x >> 8) & 255), ((x >> 16) & 255), (x >> 24) struct av_capture; #define AVLOG(level, format, ...) \ blog(level, "%s: " format, \ obs_source_getname(capture->source), ##__VA_ARGS__) #define AVFREE(x) {[x release]; x = nil;} @interface OBSAVCaptureDelegate : NSObject { @public struct av_capture *capture; } - (void)captureOutput:(AVCaptureOutput *)out didDropSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection; - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection; @end struct av_capture { AVCaptureSession *session; AVCaptureDevice *device; AVCaptureDeviceInput *device_input; AVCaptureVideoDataOutput *out; OBSAVCaptureDelegate *delegate; dispatch_queue_t queue; bool has_clock; NSString *uid; id connect_observer; id disconnect_observer; unsigned fourcc; enum video_format video_format; obs_source_t source; struct source_frame frame; }; static inline void update_frame_size(struct av_capture *capture, struct source_frame *frame, uint32_t width, uint32_t height) { if (width != frame->width) { AVLOG(LOG_DEBUG, "Changed width from %d to %d", frame->width, width); frame->width = width; frame->linesize[0] = width * 2; } if (height != frame->height) { AVLOG(LOG_DEBUG, "Changed height from %d to %d", frame->height, height); frame->height = height; } } @implementation OBSAVCaptureDelegate - (void)captureOutput:(AVCaptureOutput *)out didDropSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection { UNUSED_PARAMETER(out); UNUSED_PARAMETER(sampleBuffer); UNUSED_PARAMETER(connection); } - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection { UNUSED_PARAMETER(captureOutput); UNUSED_PARAMETER(connection); CMItemCount count = CMSampleBufferGetNumSamples(sampleBuffer); if (count < 1 || !capture) return; struct source_frame *frame = &capture->frame; CVImageBufferRef img = CMSampleBufferGetImageBuffer(sampleBuffer); update_frame_size(capture, frame, CVPixelBufferGetWidth(img), CVPixelBufferGetHeight(img)); CMSampleTimingInfo info; CMSampleBufferGetSampleTimingInfo(sampleBuffer, 0, &info); CMTime target_pts; if (capture->has_clock) { AVCaptureInputPort *port = capture->device_input.ports[0]; target_pts = CMSyncConvertTime(info.presentationTimeStamp, port.clock, CMClockGetHostTimeClock()); } else { target_pts = info.presentationTimeStamp; } CMTime target_pts_nano = CMTimeConvertScale(target_pts, NANO_TIMESCALE, kCMTimeRoundingMethod_Default); frame->timestamp = target_pts_nano.value; CVPixelBufferLockBaseAddress(img, kCVPixelBufferLock_ReadOnly); frame->data[0] = CVPixelBufferGetBaseAddress(img); obs_source_output_video(capture->source, frame); CVPixelBufferUnlockBaseAddress(img, kCVPixelBufferLock_ReadOnly); } @end static const char *av_capture_getname(void) { /* TODO: locale */ return "Video Capture Device"; } static void remove_device(struct av_capture *capture) { [capture->session stopRunning]; [capture->session removeInput:capture->device_input]; AVFREE(capture->device_input); AVFREE(capture->device); } static void av_capture_destroy(void *data) { struct av_capture *capture = data; if (!capture) return; NSNotificationCenter *nc = [NSNotificationCenter defaultCenter]; [nc removeObserver:capture->connect_observer]; [nc removeObserver:capture->disconnect_observer]; remove_device(capture); AVFREE(capture->out); if (capture->queue) dispatch_release(capture->queue); AVFREE(capture->delegate); AVFREE(capture->session); AVFREE(capture->uid); bfree(capture); } static NSString *get_string(obs_data_t data, char const *name) { return @(obs_data_getstring(data, name)); } static bool init_session(struct av_capture *capture) { capture->session = [[AVCaptureSession alloc] init]; if (!capture->session) { AVLOG(LOG_ERROR, "Could not create AVCaptureSession"); goto error; } capture->delegate = [[OBSAVCaptureDelegate alloc] init]; if (!capture->delegate) { AVLOG(LOG_ERROR, "Could not create OBSAVCaptureDelegate"); goto error; } capture->delegate->capture = capture; capture->out = [[AVCaptureVideoDataOutput alloc] init]; if (!capture->out) { AVLOG(LOG_ERROR, "Could not create AVCaptureVideoDataOutput"); goto error; } capture->queue = dispatch_queue_create(NULL, NULL); if (!capture->queue) { AVLOG(LOG_ERROR, "Could not create dispatch queue"); goto error; } [capture->session addOutput:capture->out]; [capture->out setSampleBufferDelegate:capture->delegate queue:capture->queue]; return true; error: AVFREE(capture->session); AVFREE(capture->delegate); AVFREE(capture->out); return false; } static bool init_device_input(struct av_capture *capture) { NSError *err = nil; capture->device_input = [[AVCaptureDeviceInput deviceInputWithDevice:capture->device error:&err] retain]; if (!capture->device_input) { AVLOG(LOG_ERROR, "Error while initializing device input: %s", err.localizedFailureReason.UTF8String); return false; } [capture->session addInput:capture->device_input]; return true; } static uint32_t uint_from_dict(NSDictionary *dict, CFStringRef key) { return ((NSNumber*)dict[(__bridge NSString*)key]).unsignedIntValue; } static bool init_format(struct av_capture *capture) { AVCaptureDeviceFormat *format = capture->device.activeFormat; CMMediaType mtype = CMFormatDescriptionGetMediaType( format.formatDescription); // TODO: support other media types if (mtype != kCMMediaType_Video) { AVLOG(LOG_ERROR, "CMMediaType '%c%c%c%c' is unsupported", AV_REV_FOURCC(mtype)); return false; } capture->out.videoSettings = nil; capture->fourcc = htonl(uint_from_dict(capture->out.videoSettings, kCVPixelBufferPixelFormatTypeKey)); capture->video_format = video_format_from_fourcc(capture->fourcc); if (capture->video_format == VIDEO_FORMAT_NONE) { AVLOG(LOG_ERROR, "FourCC '%c%c%c%c' unsupported by libobs", AV_REV_FOURCC(capture->fourcc)); return false; } AVLOG(LOG_DEBUG, "Using FourCC '%c%c%c%c'", AV_REV_FOURCC(capture->fourcc)); return true; } static bool init_frame(struct av_capture *capture) { AVCaptureDeviceFormat *format = capture->device.activeFormat; CMVideoDimensions size = CMVideoFormatDescriptionGetDimensions( format.formatDescription); capture->frame.width = size.width; capture->frame.linesize[0] = size.width * 2; capture->frame.height = size.height; capture->frame.format = capture->video_format; capture->frame.full_range = false; NSDictionary *exts = (__bridge NSDictionary*)CMFormatDescriptionGetExtensions( format.formatDescription); capture->frame.linesize[0] = uint_from_dict(exts, (__bridge CFStringRef)@"CVBytesPerRow"); NSString *matrix_key = (__bridge NSString*)kCVImageBufferYCbCrMatrixKey; enum video_colorspace colorspace = exts[matrix_key] == (id)kCVImageBufferYCbCrMatrix_ITU_R_709_2 ? VIDEO_CS_709 : VIDEO_CS_601; if (!video_format_get_parameters(colorspace, VIDEO_RANGE_PARTIAL, capture->frame.color_matrix, capture->frame.color_range_min, capture->frame.color_range_max)) { AVLOG(LOG_ERROR, "Failed to get video format parameters for " "video format %u", colorspace); return false; } return true; } static NSArray *presets(void); static NSString *preset_names(NSString *preset); static NSString *select_preset(AVCaptureDevice *dev, NSString *cur_preset) { NSString *new_preset = nil; bool found_previous_preset = false; for (NSString *preset in presets()) { if (!found_previous_preset) found_previous_preset = [cur_preset isEqualToString:preset]; if (![dev supportsAVCaptureSessionPreset:preset]) continue; if (!new_preset || !found_previous_preset) new_preset = preset; } return new_preset; } static void capture_device(struct av_capture *capture, AVCaptureDevice *dev, obs_data_t settings) { capture->device = dev; AVLOG(LOG_INFO, "Selected device '%s' %p", capture->device.localizedName.UTF8String, capture); if (obs_data_getbool(settings, "use_preset")) { NSString *preset = get_string(settings, "preset"); if (![dev supportsAVCaptureSessionPreset:preset]) { AVLOG(LOG_ERROR, "Preset %s not available", preset_names(preset).UTF8String); preset = select_preset(dev, preset); } if (!preset) { AVLOG(LOG_ERROR, "Could not select a preset, " "initialization failed"); return; } capture->session.sessionPreset = preset; AVLOG(LOG_INFO, "Using preset %s", preset_names(preset).UTF8String); } if (!init_device_input(capture)) goto error_input; AVCaptureInputPort *port = capture->device_input.ports[0]; capture->has_clock = [port respondsToSelector:@selector(clock)]; if (!init_format(capture)) goto error; if (!init_frame(capture)) goto error; [capture->session startRunning]; return; error: [capture->session removeInput:capture->device_input]; AVFREE(capture->device_input); error_input: AVFREE(capture->device); } static inline void handle_disconnect(struct av_capture* capture, AVCaptureDevice *dev) { if (!dev) return; if (![dev.uniqueID isEqualTo:capture->uid]) return; if (!capture->device) { AVLOG(LOG_ERROR, "Received disconnect for unused device '%s'", capture->uid.UTF8String); return; } AVLOG(LOG_ERROR, "Device with unique ID '%s' disconnected", dev.uniqueID.UTF8String); remove_device(capture); } static inline void handle_connect(struct av_capture *capture, AVCaptureDevice *dev, obs_data_t settings) { if (!dev) return; if (![dev.uniqueID isEqualTo:capture->uid]) return; if (capture->device) { AVLOG(LOG_ERROR, "Received connect for in-use device '%s'", capture->uid.UTF8String); return; } AVLOG(LOG_INFO, "Device with unique ID '%s' connected, " "resuming capture", dev.uniqueID.UTF8String); capture_device(capture, [dev retain], settings); } static void av_capture_init(struct av_capture *capture, obs_data_t settings) { if (!init_session(capture)) return; capture->uid = [get_string(settings, "device") retain]; NSNotificationCenter *nc = [NSNotificationCenter defaultCenter]; capture->disconnect_observer = [nc addObserverForName:AVCaptureDeviceWasDisconnectedNotification object:nil queue:[NSOperationQueue mainQueue] usingBlock:^(NSNotification *note) { handle_disconnect(capture, note.object); } ]; capture->connect_observer = [nc addObserverForName:AVCaptureDeviceWasConnectedNotification object:nil queue:[NSOperationQueue mainQueue] usingBlock:^(NSNotification *note) { handle_connect(capture, note.object, settings); } ]; AVCaptureDevice *dev = [[AVCaptureDevice deviceWithUniqueID:capture->uid] retain]; if (!dev) { if (capture->uid.length < 1) AVLOG(LOG_ERROR, "No device selected"); else AVLOG(LOG_ERROR, "Could not initialize device " \ "with unique ID '%s'", capture->uid.UTF8String); return; } capture_device(capture, dev, settings); } static void *av_capture_create(obs_data_t settings, obs_source_t source) { UNUSED_PARAMETER(source); struct av_capture *capture = bzalloc(sizeof(struct av_capture)); capture->source = source; av_capture_init(capture, settings); if (!capture->session) { AVLOG(LOG_ERROR, "No valid session, returning NULL context"); av_capture_destroy(capture); bfree(capture); capture = NULL; } return capture; } static NSArray *presets(void) { return @[ //AVCaptureSessionPresetPhoto, //AVCaptureSessionPresetLow, //AVCaptureSessionPresetMedium, //AVCaptureSessionPresetHigh, AVCaptureSessionPreset320x240, AVCaptureSessionPreset352x288, AVCaptureSessionPreset640x480, AVCaptureSessionPreset960x540, AVCaptureSessionPreset1280x720, //AVCaptureSessionPresetiFrame960x540, //AVCaptureSessionPresetiFrame1280x720, ]; } static NSString *preset_names(NSString *preset) { NSDictionary *preset_names = @{ AVCaptureSessionPresetLow:@"Low", AVCaptureSessionPresetMedium:@"Medium", AVCaptureSessionPresetHigh:@"High", AVCaptureSessionPreset320x240:@"320x240", AVCaptureSessionPreset352x288:@"352x288", AVCaptureSessionPreset640x480:@"640x480", AVCaptureSessionPreset960x540:@"960x540", AVCaptureSessionPreset1280x720:@"1280x720", }; return preset_names[preset]; } static void av_capture_defaults(obs_data_t settings) { AVCaptureDevice *dev = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo]; if (!dev) return; NSString *highest = nil; for (NSString *preset in presets()) { if (![dev supportsAVCaptureSessionPreset:preset]) continue; highest = preset; } if (!highest) return; obs_data_set_default_string(settings, "device", dev.uniqueID.UTF8String); obs_data_set_default_bool(settings, "use_preset", true); obs_data_set_default_string(settings, "preset", highest.UTF8String); } static bool properties_device_changed(obs_properties_t props, obs_property_t p, obs_data_t settings) { UNUSED_PARAMETER(p); NSString *uid = get_string(settings, "device"); AVCaptureDevice *dev = [AVCaptureDevice deviceWithUniqueID:uid]; if (!dev) return false; obs_property_t preset_list = obs_properties_get(props, "preset"); obs_property_list_clear(preset_list); for (NSString *preset in presets()) { if (![dev supportsAVCaptureSessionPreset:preset]) continue; obs_property_list_add_string(preset_list, preset_names(preset).UTF8String, preset.UTF8String); } NSString *preset = get_string(settings, "preset"); if (![dev supportsAVCaptureSessionPreset:preset]) obs_data_setstring(settings, "preset", select_preset(dev, preset).UTF8String); return true; } static obs_properties_t av_capture_properties(void) { obs_properties_t props = obs_properties_create(); /* TODO: locale */ obs_property_t dev_list = obs_properties_add_list(props, "device", "Device", OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_STRING); for (AVCaptureDevice *dev in [AVCaptureDevice devicesWithMediaType:AVMediaTypeVideo]) { obs_property_list_add_string(dev_list, dev.localizedName.UTF8String, dev.uniqueID.UTF8String); } obs_property_set_modified_callback(dev_list, properties_device_changed); obs_property_t use_preset = obs_properties_add_bool(props, "use_preset", "Use preset"); // TODO: implement manual configuration obs_property_set_enabled(use_preset, false); obs_properties_add_list(props, "preset", "Preset", OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_STRING); return props; } static void switch_device(struct av_capture *capture, NSString *uid, obs_data_t settings) { if (!uid) return; if (capture->device) remove_device(capture); AVFREE(capture->uid); capture->uid = [uid retain]; AVCaptureDevice *dev = [AVCaptureDevice deviceWithUniqueID:uid]; if (!dev) { AVLOG(LOG_ERROR, "Device with unique id '%s' not found", uid.UTF8String); return; } capture_device(capture, [dev retain], settings); } static void av_capture_update(void *data, obs_data_t settings) { struct av_capture *capture = data; NSString *uid = get_string(settings, "device"); if (!capture->device || ![capture->device.uniqueID isEqualToString:uid]) return switch_device(capture, uid, settings); capture->session.sessionPreset = get_string(settings, "preset"); } struct obs_source_info av_capture_info = { .id = "av_capture_input", .type = OBS_SOURCE_TYPE_INPUT, .output_flags = OBS_SOURCE_ASYNC_VIDEO, .getname = av_capture_getname, .create = av_capture_create, .destroy = av_capture_destroy, .defaults = av_capture_defaults, .properties = av_capture_properties, .update = av_capture_update, };