一、前言
上一篇文章,分享了視頻h264硬編碼,有了視頻,怎么能少了音頻呢!接下來分享音頻aac硬編碼。
二、音頻硬編碼
1.編碼器類DDHardwareAudioEncoder.h文件中,該類繼承自DDAudioEncoding(編碼器抽象接口類)其中,DDLiveAudioConfiguration是音頻配置文件,里面是音頻采樣率、碼率、聲道數(shù)目等音頻相關(guān)屬性,具體文件實(shí)現(xiàn)如下:
#import "DDAudioEncoding.h"
@interface DDHardwareAudioEncoder : NSObject<DDAudioEncoding>
#pragma mark - Initializer
///=============================================================================
/// @name Initializer
///=============================================================================
- (nullable instancetype)init UNAVAILABLE_ATTRIBUTE;
+ (nullable instancetype)new UNAVAILABLE_ATTRIBUTE;
@end
2.編碼器抽象接口類DDAudioEncoding.h文件實(shí)現(xiàn)如下:
其中DDAudioFrame類是編碼成功后數(shù)據(jù)處理類,里面有每幀編碼成功后的data、nSamplePerSec(采樣率)、nChannel(通道數(shù))、audioHeader(音頻頭)等屬性。
#import <Foundation/Foundation.h>
#import <AVFoundation/AVFoundation.h>
#import "DDAudioFrame.h"
#import "DDLiveAudioConfiguration.h"
@protocol DDAudioEncoding;
/// 編碼器編碼后回調(diào)
@protocol DDAudioEncodingDelegate <NSObject>
@required
- (void)audioEncoder:(nullable id<DDAudioEncoding>)encoder audioFrame:(nullable DDAudioFrame*)frame;
@end
/// 編碼器抽象的接口
@protocol DDAudioEncoding <NSObject>
@required
- (void)encodeAudioData:(AudioBufferList)inBufferList timeStamp:(uint64_t)timeStamp;
@optional
- (nullable instancetype)initWithAudioStreamConfiguration:(nullable DDLiveAudioConfiguration*)configuration;
- (void)setDelegate:(nullable id<DDAudioEncodingDelegate>)delegate;
- (nullable NSData*)adtsData:(NSInteger)channel rawDataLength:(NSInteger)rawDataLength;
@end
3.下面是具體實(shí)現(xiàn)的DDHardwareAudioEncoder.m類文件
#import "DDHardwareAudioEncoder.h"
@interface DDHardwareAudioEncoder (){
AudioConverterRef m_converter;
char *aacBuf;
}
@property (nonatomic, strong) DDLiveAudioConfiguration *configuration;
@property (nonatomic, weak) id<DDAudioEncodingDelegate> aacDeleage;
@end
@implementation DDHardwareAudioEncoder
- (instancetype)initWithAudioStreamConfiguration:(DDLiveAudioConfiguration *)configuration{
if(self = [super init]){
_configuration = configuration;
}
return self;
}
- (void)dealloc{
if(aacBuf) free(aacBuf);
}
#pragma mark -- DDAudioEncoder
- (void)setDelegate:(id<DDAudioEncodingDelegate>)delegate{
_aacDeleage = delegate;
}
- (void)encodeAudioData:(AudioBufferList)inBufferList timeStamp:(uint64_t)timeStamp{
if (![self createAudioConvert]){
return;
}
if(!aacBuf){
aacBuf = malloc(inBufferList.mBuffers[0].mDataByteSize);
}
// 初始化一個(gè)輸出緩沖列表
AudioBufferList outBufferList;
outBufferList.mNumberBuffers = 1;
outBufferList.mBuffers[0].mNumberChannels = inBufferList.mBuffers[0].mNumberChannels;
outBufferList.mBuffers[0].mDataByteSize = inBufferList.mBuffers[0].mDataByteSize; // 設(shè)置緩沖區(qū)大小
outBufferList.mBuffers[0].mData = aacBuf; // 設(shè)置AAC緩沖區(qū)
UInt32 outputDataPacketSize = 1;
if (AudioConverterFillComplexBuffer(m_converter, inputDataProc, &inBufferList, &outputDataPacketSize, &outBufferList, NULL) != noErr){
return;
}
DDAudioFrame *audioFrame = [[DDAudioFrame alloc] init];
audioFrame.timestamp = timeStamp;
audioFrame.nSamplePerSec = self.configuration.audioSampleRate;
audioFrame.nChannel = self.configuration.numberOfChannels;
NSData *rawAAC = [NSData dataWithBytes:outBufferList.mBuffers[0].mData length:outBufferList.mBuffers[0].mDataByteSize];
NSData *adtsHeader = [self adtsData:2 rawDataLength:rawAAC.length];
NSMutableData *fullData = [NSMutableData dataWithData:adtsHeader];
[fullData appendData:rawAAC];
audioFrame.data = fullData;
char exeData[2];
exeData[0] = _configuration.asc[0];
exeData[1] = _configuration.asc[1];
if(self.aacDeleage && [self.aacDeleage respondsToSelector:@selector(audioEncoder:audioFrame:)]){
[self.aacDeleage audioEncoder:self audioFrame:audioFrame]; // 數(shù)據(jù)傳出去之后,實(shí)現(xiàn)該代理方法,根據(jù)后臺數(shù)據(jù)格式進(jìn)行數(shù)據(jù)封裝,然后發(fā)送
}
}
#pragma mark -- CustomMethod
-(BOOL)createAudioConvert{ //根據(jù)輸入樣本初始化一個(gè)編碼轉(zhuǎn)換器
if (m_converter != nil){
return TRUE;
}
AudioStreamBasicDescription inputFormat = {0};
inputFormat.mSampleRate = _configuration.audioSampleRate;
inputFormat.mFormatID = kAudioFormatLinearPCM;
inputFormat.mFormatFlags = kAudioFormatFlagIsSignedInteger | kAudioFormatFlagsNativeEndian | kAudioFormatFlagIsPacked;
inputFormat.mChannelsPerFrame = (UInt32)_configuration.numberOfChannels;
inputFormat.mFramesPerPacket = 1;
inputFormat.mBitsPerChannel = 16;
inputFormat.mBytesPerFrame = inputFormat.mBitsPerChannel / 8 * inputFormat.mChannelsPerFrame;
inputFormat.mBytesPerPacket = inputFormat.mBytesPerFrame * inputFormat.mFramesPerPacket;
AudioStreamBasicDescription outputFormat; // 這里開始是輸出音頻格式
memset(&outputFormat, 0, sizeof(outputFormat));
outputFormat.mSampleRate = inputFormat.mSampleRate; // 采樣率保持一致
outputFormat.mFormatID = kAudioFormatMPEG4AAC; // AAC編碼 kAudioFormatMPEG4AAC kAudioFormatMPEG4AAC_HE_V2
outputFormat.mChannelsPerFrame = (UInt32)_configuration.numberOfChannels;;
outputFormat.mFramesPerPacket = 1024; // AAC一幀是1024個(gè)字節(jié)
const OSType subtype = kAudioFormatMPEG4AAC;
AudioClassDescription requestedCodecs[2] = {
{
kAudioEncoderComponentType,
subtype,
kAppleSoftwareAudioCodecManufacturer
},
{
kAudioEncoderComponentType,
subtype,
kAppleHardwareAudioCodecManufacturer
}
};
OSStatus result = AudioConverterNewSpecific(&inputFormat, &outputFormat, 2, requestedCodecs, &m_converter);
if(result != noErr) return NO;
return YES;
}
#pragma mark -- AudioCallBack
OSStatus inputDataProc(AudioConverterRef inConverter, UInt32 *ioNumberDataPackets, AudioBufferList *ioData,AudioStreamPacketDescription **outDataPacketDescription, void *inUserData) {
AudioBufferList bufferList = *(AudioBufferList*)inUserData;
ioData->mBuffers[0].mNumberChannels = 1;
ioData->mBuffers[0].mData = bufferList.mBuffers[0].mData;
ioData->mBuffers[0].mDataByteSize = bufferList.mBuffers[0].mDataByteSize;
return noErr;
}
/**
* Add ADTS header at the beginning of each and every AAC packet.
* This is needed as MediaCodec encoder generates a packet of raw
* AAC data.
*
* Note the packetLen must count in the ADTS header itself.
* See: http://wiki.multimedia.cx/index.php?title=ADTS
* Also: http://wiki.multimedia.cx/index.php?title=MPEG-4_Audio#Channel_Configurations
**/
- (NSData*)adtsData:(NSInteger)channel rawDataLength:(NSInteger)rawDataLength {
int adtsLength = 7;
char *packet = malloc(sizeof(char) * adtsLength);
// Variables Recycled by addADTStoPacket
int profile = 2; //AAC LC
//39=MediaCodecInfo.CodecProfileLevel.AACObjectELD;
int freqIdx = 4; //44.1KHz
int chanCfg = (int)channel; //MPEG-4 Audio Channel Configuration. 1 Channel front-center
NSUInteger fullLength = adtsLength + rawDataLength;
// fill in ADTS data
packet[0] = (char)0xFF; // 11111111 = syncword
packet[1] = (char)0xF9; // 1111 1 00 1 = syncword MPEG-2 Layer CRC
packet[2] = (char)(((profile-1)<<6) + (freqIdx<<2) +(chanCfg>>2));
packet[3] = (char)(((chanCfg&3)<<6) + (fullLength>>11));
packet[4] = (char)((fullLength&0x7FF) >> 3);
packet[5] = (char)(((fullLength&7)<<5) + 0x1F);
packet[6] = (char)0xFC;
NSData *data = [NSData dataWithBytesNoCopy:packet length:adtsLength freeWhenDone:YES];
return data;
}
@end
三、FFLiveKit
如果做視頻拉流端既采集音視頻、編碼、封裝、推流,推薦參考FFLiveKit這一框架,很詳細(xì),自己看源碼就行了,這里就不再贅述。