// // OutputCoreAudio.m // Cog // // Created by Vincent Spader on 8/2/05. // Copyright 2005 Vincent Spader. All rights reserved. // #import "OutputCoreAudio.h" #import "OutputNode.h" #import "Logging.h" @implementation OutputCoreAudio - (id)initWithController:(OutputNode *)c { self = [super init]; if (self) { outputController = c; outputUnit = NULL; [[NSUserDefaultsController sharedUserDefaultsController] addObserver:self forKeyPath:@"values.outputDevice" options:0 context:NULL]; } return self; } static OSStatus Sound_Renderer(void *inRefCon, AudioUnitRenderActionFlags *ioActionFlags, const AudioTimeStamp *inTimeStamp, UInt32 inBusNumber, UInt32 inNumberFrames, AudioBufferList *ioData) { OutputCoreAudio *output = (__bridge OutputCoreAudio *)inRefCon; OSStatus err = noErr; void *readPointer = ioData->mBuffers[0].mData; int amountToRead, amountRead; if ([output->outputController shouldContinue] == NO) { AudioOutputUnitStop(output->outputUnit); // [output stop]; return err; } amountToRead = inNumberFrames*(output->deviceFormat.mBytesPerPacket); amountRead = [output->outputController readData:(readPointer) amount:amountToRead]; if ((amountRead < amountToRead) && [output->outputController endOfStream] == NO) //Try one more time! for track changes! { int amountRead2; //Use this since return type of readdata isnt known...may want to fix then can do a simple += to readdata amountRead2 = [output->outputController readData:(readPointer+amountRead) amount:amountToRead-amountRead]; amountRead += amountRead2; } ioData->mBuffers[0].mDataByteSize = amountRead; ioData->mBuffers[0].mNumberChannels = output->deviceFormat.mChannelsPerFrame; ioData->mNumberBuffers = 1; return err; } - (void)observeValueForKeyPath:(NSString *)keyPath ofObject:(id)object change:(NSDictionary *)change context:(void *)context { if ([keyPath isEqualToString:@"values.outputDevice"]) { NSDictionary *device = [[[NSUserDefaultsController sharedUserDefaultsController] defaults] objectForKey:@"outputDevice"]; NSNumber *deviceID = [device objectForKey:@"deviceID"]; [self setOutputDevice:[deviceID longValue]]; } } - (BOOL)setOutputDevice:(AudioDeviceID)outputDevice { // Set the output device AudioDeviceID deviceID = outputDevice; //XXX use default if null OSStatus err; if (outputDevice == -1) { DLog(@"DEVICE IS -1"); UInt32 size = sizeof(AudioDeviceID); AudioObjectPropertyAddress theAddress = { .mSelector = kAudioHardwarePropertyDefaultOutputDevice, .mScope = kAudioObjectPropertyScopeGlobal, .mElement = kAudioObjectPropertyElementMaster }; err = AudioObjectGetPropertyData(kAudioObjectSystemObject, &theAddress, 0, NULL, &size, &deviceID); if (err != noErr) { ALog(@"THERES NO DEFAULT OUTPUT DEVICE"); return NO; } } printf("DEVICE: %i\n", deviceID); outputDeviceID = deviceID; err = AudioUnitSetProperty(outputUnit, kAudioOutputUnitProperty_CurrentDevice, kAudioUnitScope_Output, 0, &deviceID, sizeof(AudioDeviceID)); if (err != noErr) { ALog(@"No output device could be found, your random error code is %d. Have a nice day!", err); return NO; } return YES; } - (BOOL)setup { if (outputUnit) [self stop]; AudioObjectPropertyAddress propertyAddress = { .mElement = kAudioObjectPropertyElementMaster }; UInt32 dataSize; ComponentDescription desc; OSStatus err; desc.componentType = kAudioUnitType_Output; desc.componentSubType = kAudioUnitSubType_DefaultOutput; desc.componentManufacturer = kAudioUnitManufacturer_Apple; desc.componentFlags = 0; desc.componentFlagsMask = 0; Component comp = FindNextComponent(NULL, &desc); //Finds an component that meets the desc spec's if (comp == NULL) return NO; err = OpenAComponent(comp, &outputUnit); //gains access to the services provided by the component if (err) return NO; // Initialize AudioUnit err = AudioUnitInitialize(outputUnit); if (err != noErr) return NO; // Setup the output device before mucking with settings NSDictionary *device = [[[NSUserDefaultsController sharedUserDefaultsController] defaults] objectForKey:@"outputDevice"]; if (device) { BOOL ok = [self setOutputDevice:[[device objectForKey:@"deviceID"] longValue]]; if (!ok) { //Ruh roh. [self setOutputDevice: -1]; [[[NSUserDefaultsController sharedUserDefaultsController] defaults] removeObjectForKey:@"outputDevice"]; } } else { [self setOutputDevice: -1]; } UInt32 size = sizeof (AudioStreamBasicDescription); Boolean outWritable; //Gets the size of the Stream Format Property and if it is writable AudioUnitGetPropertyInfo(outputUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Output, 0, &size, &outWritable); //Get the current stream format of the output err = AudioUnitGetProperty (outputUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Output, 0, &deviceFormat, &size); if (err != noErr) return NO; // change output format... // The default channel map is silence SInt32 deviceChannelMap [deviceFormat.mChannelsPerFrame]; for(UInt32 i = 0; i < deviceFormat.mChannelsPerFrame; ++i) deviceChannelMap[i] = -1; // Determine the device's preferred stereo channels for output mapping if(1 == deviceFormat.mChannelsPerFrame || 2 == deviceFormat.mChannelsPerFrame) { propertyAddress.mSelector = kAudioDevicePropertyPreferredChannelsForStereo; propertyAddress.mScope = kAudioDevicePropertyScopeOutput; UInt32 preferredStereoChannels [2] = { 1, 2 }; if(AudioObjectHasProperty(outputDeviceID, &propertyAddress)) { dataSize = sizeof(preferredStereoChannels); err = AudioObjectGetPropertyData(outputDeviceID, &propertyAddress, 0, nil, &dataSize, &preferredStereoChannels); } AudioChannelLayout stereoLayout; stereoLayout.mChannelLayoutTag = kAudioChannelLayoutTag_Stereo; const AudioChannelLayout *specifier [1] = { &stereoLayout }; SInt32 stereoChannelMap [2] = { 1, 2 }; dataSize = sizeof(stereoChannelMap); err = AudioFormatGetProperty(kAudioFormatProperty_ChannelMap, sizeof(specifier), specifier, &dataSize, stereoChannelMap); if(noErr == err) { deviceChannelMap[preferredStereoChannels[0] - 1] = stereoChannelMap[0]; deviceChannelMap[preferredStereoChannels[1] - 1] = stereoChannelMap[1]; } else { // Just use a channel map that makes sense deviceChannelMap[preferredStereoChannels[0] - 1] = 0; deviceChannelMap[preferredStereoChannels[1] - 1] = 1; } } // Determine the device's preferred multichannel layout else { propertyAddress.mSelector = kAudioDevicePropertyPreferredChannelLayout; propertyAddress.mScope = kAudioDevicePropertyScopeOutput; if(AudioObjectHasProperty(outputDeviceID, &propertyAddress)) { err = AudioObjectGetPropertyDataSize(outputDeviceID, &propertyAddress, 0, nil, &dataSize); AudioChannelLayout *preferredChannelLayout = (AudioChannelLayout *)(malloc(dataSize)); err = AudioObjectGetPropertyData(outputDeviceID, &propertyAddress, 0, nil, &dataSize, preferredChannelLayout); const AudioChannelLayout *specifier [1] = { preferredChannelLayout }; // Not all channel layouts can be mapped, so handle failure with a generic mapping dataSize = (UInt32)sizeof(deviceChannelMap); err = AudioFormatGetProperty(kAudioFormatProperty_ChannelMap, sizeof(specifier), specifier, &dataSize, deviceChannelMap); if(noErr != err) { // Just use a channel map that makes sense for(UInt32 i = 0; i < deviceFormat.mChannelsPerFrame; ++i) deviceChannelMap[i] = i; } free(preferredChannelLayout), preferredChannelLayout = nil; } else { // Just use a channel map that makes sense for(UInt32 i = 0; i < deviceFormat.mChannelsPerFrame; ++i) deviceChannelMap[i] = i; } } ///Seems some 3rd party devices return incorrect stuff...or I just don't like noninterleaved data. deviceFormat.mFormatFlags &= ~kLinearPCMFormatFlagIsNonInterleaved; // deviceFormat.mFormatFlags &= ~kLinearPCMFormatFlagIsFloat; // deviceFormat.mFormatFlags = kLinearPCMFormatFlagIsSignedInteger; deviceFormat.mBytesPerFrame = deviceFormat.mChannelsPerFrame*(deviceFormat.mBitsPerChannel/8); deviceFormat.mBytesPerPacket = deviceFormat.mBytesPerFrame * deviceFormat.mFramesPerPacket; err = AudioUnitSetProperty (outputUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Output, 0, &deviceFormat, size); //Set the stream format of the output to match the input err = AudioUnitSetProperty (outputUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Input, 0, &deviceFormat, size); size = sizeof(deviceChannelMap); err = AudioUnitSetProperty(outputUnit, kAudioOutputUnitProperty_ChannelMap, kAudioUnitScope_Output, 0, deviceChannelMap, size); //setup render callbacks renderCallback.inputProc = Sound_Renderer; renderCallback.inputProcRefCon = (__bridge void * _Nullable)(self); AudioUnitSetProperty(outputUnit, kAudioUnitProperty_SetRenderCallback, kAudioUnitScope_Input, 0, &renderCallback, sizeof(AURenderCallbackStruct)); [outputController setFormat:&deviceFormat]; return (err == noErr); } - (void)setVolume:(double)v { AudioUnitSetParameter (outputUnit, kHALOutputParam_Volume, kAudioUnitScope_Global, 0, v * 0.01f, 0); } - (void)start { AudioOutputUnitStart(outputUnit); } - (void)stop { if (outputUnit) { AudioOutputUnitStop(outputUnit); AudioUnitUninitialize (outputUnit); CloseComponent(outputUnit); outputUnit = NULL; } } - (void)dealloc { [self stop]; [[NSUserDefaultsController sharedUserDefaultsController] removeObserver:self forKeyPath:@"values.outputDevice"]; } - (void)pause { AudioOutputUnitStop(outputUnit); } - (void)resume { AudioOutputUnitStart(outputUnit); } @end