28 #import <AVFoundation/AVFoundation.h>
77 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
135 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
136 AVCaptureDeviceTransportControlsPlaybackMode observed_mode;
160 - (void) captureOutput:(AVCaptureOutput *)captureOutput
161 didOutputSampleBuffer:(CMSampleBufferRef)videoFrame
162 fromConnection:(AVCaptureConnection *)connection;
170 if (
self = [super
init]) {
174 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
176 NSString *keyPath = NSStringFromSelector(
@selector(transportControlsPlaybackMode));
177 NSKeyValueObservingOptions
options = NSKeyValueObservingOptionNew;
179 [
_context->observed_device addObserver: self
191 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
193 NSString *keyPath = NSStringFromSelector(
@selector(transportControlsPlaybackMode));
194 [_context->observed_device removeObserver: self forKeyPath: keyPath];
200 - (void)observeValueForKeyPath:(NSString *)keyPath
202 change:(NSDictionary *)change
203 context:(
void *)context {
205 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
206 AVCaptureDeviceTransportControlsPlaybackMode
mode =
207 [change[NSKeyValueChangeNewKey] integerValue];
210 if (
mode == AVCaptureDeviceTransportControlsNotPlayingMode) {
217 [
super observeValueForKeyPath: keyPath
224 - (void) captureOutput:(AVCaptureOutput *)captureOutput
225 didOutputSampleBuffer:(CMSampleBufferRef)videoFrame
226 fromConnection:(AVCaptureConnection *)connection
252 - (void) captureOutput:(AVCaptureOutput *)captureOutput
253 didOutputSampleBuffer:(CMSampleBufferRef)audioFrame
254 fromConnection:(AVCaptureConnection *)connection;
262 if (
self = [super
init]) {
268 - (void) captureOutput:(AVCaptureOutput *)captureOutput
269 didOutputSampleBuffer:(CMSampleBufferRef)audioFrame
270 fromConnection:(AVCaptureConnection *)connection
289 [ctx->capture_session stopRunning];
291 [ctx->capture_session release];
292 [ctx->video_output release];
293 [ctx->audio_output release];
294 [ctx->avf_delegate release];
295 [ctx->avf_audio_delegate release];
301 ctx->avf_audio_delegate =
NULL;
308 if (
ctx->current_frame) {
309 CFRelease(
ctx->current_frame);
346 NSObject *range = nil;
348 NSObject *selected_range = nil;
349 NSObject *selected_format = nil;
355 for (
format in [video_device valueForKey:
@"formats"]) {
356 CMFormatDescriptionRef formatDescription;
357 CMVideoDimensions dimensions;
359 formatDescription = (CMFormatDescriptionRef) [
format performSelector:
@selector(formatDescription)];
360 dimensions = CMVideoFormatDescriptionGetDimensions(formatDescription);
362 if ((
ctx->width == 0 &&
ctx->height == 0) ||
363 (dimensions.width ==
ctx->width && dimensions.height ==
ctx->height)) {
367 for (range in [
format valueForKey:
@"videoSupportedFrameRateRanges"]) {
368 double max_framerate;
370 [[range valueForKey:@"maxFrameRate"] getValue:&max_framerate];
372 selected_range = range;
379 if (!selected_format) {
382 goto unsupported_format;
385 if (!selected_range) {
388 if (
ctx->video_is_muxed) {
391 goto unsupported_format;
395 if ([video_device lockForConfiguration:
NULL] == YES) {
396 if (selected_format) {
397 [video_device setValue:selected_format forKey:@"activeFormat"];
399 if (selected_range) {
400 NSValue *min_frame_duration = [selected_range valueForKey:@"minFrameDuration"];
401 [video_device setValue:min_frame_duration forKey:@"activeVideoMinFrameDuration"];
402 [video_device setValue:min_frame_duration forKey:@"activeVideoMaxFrameDuration"];
408 }
@catch(NSException *e) {
417 for (
format in [video_device valueForKey:
@"formats"]) {
418 CMFormatDescriptionRef formatDescription;
419 CMVideoDimensions dimensions;
421 formatDescription = (CMFormatDescriptionRef) [
format performSelector:
@selector(formatDescription)];
422 dimensions = CMVideoFormatDescriptionGetDimensions(formatDescription);
424 for (range in [
format valueForKey:
@"videoSupportedFrameRateRanges"]) {
425 double min_framerate;
426 double max_framerate;
428 [[range valueForKey:@"minFrameRate"] getValue:&min_framerate];
429 [[range valueForKey:@"maxFrameRate"] getValue:&max_framerate];
431 dimensions.width, dimensions.height,
432 min_framerate, max_framerate);
442 NSError *
error = nil;
443 AVCaptureInput* capture_input = nil;
445 NSNumber *pixel_format;
446 NSDictionary *capture_dict;
447 dispatch_queue_t queue;
449 if (
ctx->video_device_index <
ctx->num_video_devices) {
450 capture_input = (AVCaptureInput*) [[[AVCaptureDeviceInput alloc] initWithDevice:video_device
error:&
error] autorelease];
452 capture_input = (AVCaptureInput*) video_device;
455 if (!capture_input) {
457 [[
error localizedDescription] UTF8String]);
461 if ([
ctx->capture_session canAddInput:capture_input]) {
462 [ctx->capture_session addInput:capture_input];
469 ctx->video_output = [[AVCaptureVideoDataOutput alloc] init];
471 if (!
ctx->video_output) {
481 }
@catch (NSException *exception) {
482 if (![[exception
name] isEqualToString:NSUndefinedKeyException]) {
506 if ([[
ctx->video_output availableVideoCVPixelFormatTypes] indexOfObject:[NSNumber numberWithInt:pxl_fmt_spec.avf_id]] == NSNotFound) {
507 av_log(
s,
AV_LOG_ERROR,
"Selected pixel format (%s) is not supported by the input device.\n",
513 for (NSNumber *pxl_fmt in [
ctx->video_output availableVideoCVPixelFormatTypes]) {
528 pxl_fmt_spec = pxl_fmt_dummy;
543 if (
ctx->capture_raw_data) {
544 ctx->pixel_format = pxl_fmt_spec.ff_id;
545 ctx->video_output.videoSettings = @{ };
547 ctx->pixel_format = pxl_fmt_spec.ff_id;
548 pixel_format = [NSNumber numberWithUnsignedInt:pxl_fmt_spec.avf_id];
549 capture_dict = [NSDictionary dictionaryWithObject:pixel_format
550 forKey:(id)kCVPixelBufferPixelFormatTypeKey];
552 [ctx->video_output setVideoSettings:capture_dict];
554 [ctx->video_output setAlwaysDiscardsLateVideoFrames:ctx->drop_late_frames];
556 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
558 if (!
ctx->video_is_screen) {
559 int trans_ctrl = [video_device transportControlsSupported];
560 AVCaptureDeviceTransportControlsPlaybackMode trans_mode = [video_device transportControlsPlaybackMode];
563 ctx->observed_mode = trans_mode;
564 ctx->observed_device = video_device;
571 queue = dispatch_queue_create(
"avf_queue",
NULL);
572 [ctx->video_output setSampleBufferDelegate:ctx->avf_delegate queue:queue];
573 dispatch_release(queue);
575 if ([
ctx->capture_session canAddOutput:
ctx->video_output]) {
576 [ctx->capture_session addOutput:ctx->video_output];
588 NSError *
error = nil;
589 AVCaptureDeviceInput* audio_dev_input = [[[AVCaptureDeviceInput alloc] initWithDevice:audio_device
error:&
error] autorelease];
590 dispatch_queue_t queue;
592 if (!audio_dev_input) {
594 [[
error localizedDescription] UTF8String]);
598 if ([
ctx->capture_session canAddInput:audio_dev_input]) {
599 [ctx->capture_session addInput:audio_dev_input];
606 ctx->audio_output = [[AVCaptureAudioDataOutput alloc] init];
608 if (!
ctx->audio_output) {
615 queue = dispatch_queue_create(
"avf_audio_queue",
NULL);
616 [ctx->audio_output setSampleBufferDelegate:ctx->avf_audio_delegate queue:queue];
617 dispatch_release(queue);
619 if ([
ctx->capture_session canAddOutput:
ctx->audio_output]) {
620 [ctx->capture_session addOutput:ctx->audio_output];
632 CVImageBufferRef image_buffer;
633 CMBlockBufferRef block_buffer;
634 CGSize image_buffer_size;
642 while (
ctx->frames_captured < 1) {
643 CFRunLoopRunInMode(kCFRunLoopDefaultMode, 0.1, YES);
648 ctx->video_stream_index = stream->index;
652 image_buffer = CMSampleBufferGetImageBuffer(
ctx->current_frame);
653 block_buffer = CMSampleBufferGetDataBuffer(
ctx->current_frame);
656 image_buffer_size = CVImageBufferGetEncodedSize(image_buffer);
660 stream->codecpar->width = (
int)image_buffer_size.width;
661 stream->codecpar->height = (
int)image_buffer_size.height;
662 stream->codecpar->format =
ctx->pixel_format;
666 stream->codecpar->format =
ctx->pixel_format;
669 CFRelease(
ctx->current_frame);
670 ctx->current_frame = nil;
680 CMFormatDescriptionRef format_desc;
688 while (
ctx->audio_frames_captured < 1) {
689 CFRunLoopRunInMode(kCFRunLoopDefaultMode, 0.1, YES);
694 ctx->audio_stream_index = stream->index;
698 format_desc = CMSampleBufferGetFormatDescription(
ctx->current_audio_frame);
699 const AudioStreamBasicDescription *basic_desc = CMAudioFormatDescriptionGetStreamBasicDescription(format_desc);
708 stream->codecpar->sample_rate = basic_desc->mSampleRate;
709 stream->codecpar->channels = basic_desc->mChannelsPerFrame;
712 ctx->audio_channels = basic_desc->mChannelsPerFrame;
713 ctx->audio_bits_per_sample = basic_desc->mBitsPerChannel;
714 ctx->audio_float = basic_desc->mFormatFlags & kAudioFormatFlagIsFloat;
715 ctx->audio_be = basic_desc->mFormatFlags & kAudioFormatFlagIsBigEndian;
716 ctx->audio_signed_integer = basic_desc->mFormatFlags & kAudioFormatFlagIsSignedInteger;
717 ctx->audio_packed = basic_desc->mFormatFlags & kAudioFormatFlagIsPacked;
718 ctx->audio_non_interleaved = basic_desc->mFormatFlags & kAudioFormatFlagIsNonInterleaved;
720 if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
722 ctx->audio_bits_per_sample == 32 &&
725 }
else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
726 ctx->audio_signed_integer &&
727 ctx->audio_bits_per_sample == 16 &&
730 }
else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
731 ctx->audio_signed_integer &&
732 ctx->audio_bits_per_sample == 24 &&
735 }
else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
736 ctx->audio_signed_integer &&
737 ctx->audio_bits_per_sample == 32 &&
746 if (
ctx->audio_non_interleaved) {
747 CMBlockBufferRef block_buffer = CMSampleBufferGetDataBuffer(
ctx->current_audio_frame);
748 ctx->audio_buffer_size = CMBlockBufferGetDataLength(block_buffer);
750 if (!
ctx->audio_buffer) {
757 CFRelease(
ctx->current_audio_frame);
758 ctx->current_audio_frame = nil;
768 NSAutoreleasePool *pool = [[NSAutoreleasePool alloc] init];
769 uint32_t num_screens = 0;
771 AVCaptureDevice *video_device = nil;
772 AVCaptureDevice *audio_device = nil;
774 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeVideo];
775 NSArray *devices_muxed = [AVCaptureDevice devicesWithMediaType:AVMediaTypeMuxed];
777 ctx->num_video_devices = [devices count] + [devices_muxed count];
781 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
782 CGGetActiveDisplayList(0,
NULL, &num_screens);
786 if (
ctx->list_devices) {
789 for (AVCaptureDevice *device in devices) {
790 const char *
name = [[device localizedName] UTF8String];
791 index = [devices indexOfObject:device];
794 for (AVCaptureDevice *device in devices_muxed) {
795 const char *
name = [[device localizedName] UTF8String];
796 index = [devices count] + [devices_muxed indexOfObject:device];
799 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
800 if (num_screens > 0) {
801 CGDirectDisplayID screens[num_screens];
802 CGGetActiveDisplayList(num_screens, screens, &num_screens);
803 for (
int i = 0;
i < num_screens;
i++) {
810 devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
811 for (AVCaptureDevice *device in devices) {
812 const char *
name = [[device localizedName] UTF8String];
813 int index = [devices indexOfObject:device];
825 if (
ctx->video_device_index == -1 &&
ctx->video_filename) {
826 sscanf(
ctx->video_filename,
"%d", &
ctx->video_device_index);
828 if (
ctx->audio_device_index == -1 &&
ctx->audio_filename) {
829 sscanf(
ctx->audio_filename,
"%d", &
ctx->audio_device_index);
832 if (
ctx->video_device_index >= 0) {
833 if (
ctx->video_device_index <
ctx->num_video_devices) {
834 if (
ctx->video_device_index < [devices count]) {
835 video_device = [devices objectAtIndex:ctx->video_device_index];
837 video_device = [devices_muxed objectAtIndex:(ctx->video_device_index - [devices count])];
838 ctx->video_is_muxed = 1;
840 }
else if (
ctx->video_device_index <
ctx->num_video_devices + num_screens) {
841 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
842 CGDirectDisplayID screens[num_screens];
843 CGGetActiveDisplayList(num_screens, screens, &num_screens);
844 AVCaptureScreenInput* capture_screen_input = [[[AVCaptureScreenInput alloc] initWithDisplayID:screens[ctx->video_device_index - ctx->
num_video_devices]] autorelease];
846 if (
ctx->framerate.num > 0) {
847 capture_screen_input.minFrameDuration = CMTimeMake(
ctx->framerate.den,
ctx->framerate.num);
850 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
851 if (
ctx->capture_cursor) {
852 capture_screen_input.capturesCursor = YES;
854 capture_screen_input.capturesCursor = NO;
858 if (
ctx->capture_mouse_clicks) {
859 capture_screen_input.capturesMouseClicks = YES;
861 capture_screen_input.capturesMouseClicks = NO;
864 video_device = (AVCaptureDevice*) capture_screen_input;
865 ctx->video_is_screen = 1;
871 }
else if (
ctx->video_filename &&
872 strncmp(
ctx->video_filename,
"none", 4)) {
873 if (!strncmp(
ctx->video_filename,
"default", 7)) {
874 video_device = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo];
877 for (AVCaptureDevice *device in devices) {
878 if (!strncmp(
ctx->video_filename, [[device localizedName] UTF8String], strlen(
ctx->video_filename))) {
879 video_device = device;
884 for (AVCaptureDevice *device in devices_muxed) {
885 if (!strncmp(
ctx->video_filename, [[device localizedName] UTF8String], strlen(
ctx->video_filename))) {
886 video_device = device;
887 ctx->video_is_muxed = 1;
892 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
896 if(sscanf(
ctx->video_filename,
"Capture screen %d", &idx) && idx < num_screens) {
897 CGDirectDisplayID screens[num_screens];
898 CGGetActiveDisplayList(num_screens, screens, &num_screens);
899 AVCaptureScreenInput* capture_screen_input = [[[AVCaptureScreenInput alloc] initWithDisplayID:screens[idx]] autorelease];
900 video_device = (AVCaptureDevice*) capture_screen_input;
901 ctx->video_device_index =
ctx->num_video_devices + idx;
902 ctx->video_is_screen = 1;
904 if (
ctx->framerate.num > 0) {
905 capture_screen_input.minFrameDuration = CMTimeMake(
ctx->framerate.den,
ctx->framerate.num);
908 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
909 if (
ctx->capture_cursor) {
910 capture_screen_input.capturesCursor = YES;
912 capture_screen_input.capturesCursor = NO;
916 if (
ctx->capture_mouse_clicks) {
917 capture_screen_input.capturesMouseClicks = YES;
919 capture_screen_input.capturesMouseClicks = NO;
933 if (
ctx->audio_device_index >= 0) {
934 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
936 if (
ctx->audio_device_index >= [devices count]) {
941 audio_device = [devices objectAtIndex:ctx->audio_device_index];
942 }
else if (
ctx->audio_filename &&
943 strncmp(
ctx->audio_filename,
"none", 4)) {
944 if (!strncmp(
ctx->audio_filename,
"default", 7)) {
945 audio_device = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeAudio];
947 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
949 for (AVCaptureDevice *device in devices) {
950 if (!strncmp(
ctx->audio_filename, [[device localizedName] UTF8String], strlen(
ctx->audio_filename))) {
951 audio_device = device;
964 if (!video_device && !audio_device) {
970 if (
ctx->video_device_index <
ctx->num_video_devices) {
977 av_log(
s,
AV_LOG_DEBUG,
"audio device '%s' opened\n", [[audio_device localizedName] UTF8String]);
981 ctx->capture_session = [[AVCaptureSession alloc] init];
989 [ctx->capture_session startRunning];
993 if (!
ctx->video_is_screen) {
994 [video_device unlockForConfiguration];
1018 CVPixelBufferRef image_buffer,
1022 int src_linesize[4];
1023 const uint8_t *src_data[4];
1024 int width = CVPixelBufferGetWidth(image_buffer);
1025 int height = CVPixelBufferGetHeight(image_buffer);
1028 memset(src_linesize, 0,
sizeof(src_linesize));
1029 memset(src_data, 0,
sizeof(src_data));
1031 status = CVPixelBufferLockBaseAddress(image_buffer, 0);
1032 if (
status != kCVReturnSuccess) {
1037 if (CVPixelBufferIsPlanar(image_buffer)) {
1038 size_t plane_count = CVPixelBufferGetPlaneCount(image_buffer);
1040 for(
i = 0;
i < plane_count;
i++){
1041 src_linesize[i] = CVPixelBufferGetBytesPerRowOfPlane(image_buffer,
i);
1042 src_data[i] = CVPixelBufferGetBaseAddressOfPlane(image_buffer,
i);
1045 src_linesize[0] = CVPixelBufferGetBytesPerRow(image_buffer);
1046 src_data[0] = CVPixelBufferGetBaseAddress(image_buffer);
1050 src_data, src_linesize,
1055 CVPixelBufferUnlockBaseAddress(image_buffer, 0);
1065 CVImageBufferRef image_buffer;
1066 CMBlockBufferRef block_buffer;
1069 if (
ctx->current_frame != nil) {
1073 image_buffer = CMSampleBufferGetImageBuffer(
ctx->current_frame);
1074 block_buffer = CMSampleBufferGetDataBuffer(
ctx->current_frame);
1076 if (image_buffer != nil) {
1077 length = (
int)CVPixelBufferGetDataSize(image_buffer);
1078 }
else if (block_buffer != nil) {
1079 length = (
int)CMBlockBufferGetDataLength(block_buffer);
1093 if (CMSampleBufferGetOutputSampleTimingInfoArray(
ctx->current_frame, 1, &
timing_info, &count) == noErr) {
1105 OSStatus
ret = CMBlockBufferCopyDataBytes(block_buffer, 0,
pkt->
size,
pkt->
data);
1106 if (
ret != kCMBlockBufferNoErr) {
1110 CFRelease(
ctx->current_frame);
1111 ctx->current_frame = nil;
1117 }
else if (
ctx->current_audio_frame != nil) {
1118 CMBlockBufferRef block_buffer = CMSampleBufferGetDataBuffer(
ctx->current_audio_frame);
1119 int block_buffer_size = CMBlockBufferGetDataLength(block_buffer);
1121 if (!block_buffer || !block_buffer_size) {
1126 if (
ctx->audio_non_interleaved && block_buffer_size >
ctx->audio_buffer_size) {
1139 if (CMSampleBufferGetOutputSampleTimingInfoArray(
ctx->current_audio_frame, 1, &
timing_info, &count) == noErr) {
1147 if (
ctx->audio_non_interleaved) {
1150 OSStatus
ret = CMBlockBufferCopyDataBytes(block_buffer, 0,
pkt->
size,
ctx->audio_buffer);
1151 if (
ret != kCMBlockBufferNoErr) {
1156 num_samples =
pkt->
size / (
ctx->audio_channels * (
ctx->audio_bits_per_sample >> 3));
1159 #define INTERLEAVE_OUTPUT(bps) \
1161 int##bps##_t **src; \
1162 int##bps##_t *dest; \
1163 src = av_malloc(ctx->audio_channels * sizeof(int##bps##_t*)); \
1165 unlock_frames(ctx); \
1166 return AVERROR(EIO); \
1169 for (c = 0; c < ctx->audio_channels; c++) { \
1170 src[c] = ((int##bps##_t*)ctx->audio_buffer) + c * num_samples; \
1172 dest = (int##bps##_t*)pkt->data; \
1173 shift = bps - ctx->audio_bits_per_sample; \
1174 for (sample = 0; sample < num_samples; sample++) \
1175 for (c = 0; c < ctx->audio_channels; c++) \
1176 *dest++ = src[c][sample] << shift; \
1180 if (
ctx->audio_bits_per_sample <= 16) {
1186 OSStatus
ret = CMBlockBufferCopyDataBytes(block_buffer, 0,
pkt->
size,
pkt->
data);
1187 if (
ret != kCMBlockBufferNoErr) {
1193 CFRelease(
ctx->current_audio_frame);
1194 ctx->current_audio_frame = nil;
1198 if (
ctx->observed_quit) {
1242 .
name =
"avfoundation",