在 iOS 平台上,直播技术已经很成熟,有许多强大的第三方框架可以帮助开发者轻松实现直播功能。当前主流的直播第三方框架包括但不限于:
- LFLiveKit:一款开源的直播推流 SDK。
- PLMediaStreamingKit:由云天存提供的一站式音视频解决方案。
- AliyunPlayer:阿里云提供的音视频播放解决方案。
- Agora SDK:声网提供的大规模实时视频通讯解决方案。
以下将详细介绍 LFLiveKit 和 PLMediaStreamingKit 的使用,并给出相应的示例代码。
一、LFLiveKit
1. LFLiveKit 安装
要使用 LFLiveKit 首先需要通过 CocoaPods 添加到你的项目。
在你的 Podfile
文件中添加如下内容:
pod 'LFLiveKit'
然后运行 pod install
。
2. 配置和使用
Import LFLiveKit 的头文件:
#import <LFLiveKit/LFLiveKit.h>
创建直播会话
LFLiveKit 提供了多种配置选项。首先,你需要创建一个 LFLiveSession
,这是 LFLiveKit 的核心类,负责管理音视频捕获和推流处理。
- (LFLiveSession*)liveSession {
if (!_liveSession) {
// 自定义音频和视频配置
LFLiveAudioConfiguration *audioConfiguration = [LFLiveAudioConfiguration defaultConfiguration];
LFLiveVideoConfiguration *videoConfiguration = [LFLiveVideoConfiguration defaultConfiguration];
_liveSession = [[LFLiveSession alloc] initWithAudioConfiguration:audioConfiguration videoConfiguration:videoConfiguration];
_liveSession.delegate = self;
_liveSession.preView = self.view; // 设置预览视图
}
return _liveSession;
}
请求权限
在 iOS 开发中,需要请求相机和麦克风权限。以下是请求权限代码:
- (void)requestAccessForVideo {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeVideo];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeVideo completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.liveSession setRunning:YES];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.liveSession setRunning:YES];
}
}
- (void)requestAccessForAudio {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeAudio];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeAudio completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.liveSession setRunning:YES];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.liveSession setRunning:YES];
}
}
开始直播
配置完 LFLiveStreamInfo
对象,并调用 startLive:
方法开始直播。
- (void)startLive {
LFLiveStreamInfo *streamInfo = [LFLiveStreamInfo new];
streamInfo.url = @"rtmp://your_server/live_stream";
[self.liveSession startLive:streamInfo];
}
- (void)stopLive {
[self.liveSession stopLive];
}
处理直播状态变化
通过实现 LFLiveSessionDelegate
,可以监测广播的状态变更。
- (void)liveSession:(LFLiveSession *)session liveStateDidChange:(LFLiveState)state {
switch (state) {
// 在每个状态变化时的对应处理
case LFLiveReady:
NSLog(@"Ready to start live streaming");
break;
case LFLivePending:
NSLog(@"Connecting...");
break;
case LFLiveStart:
NSLog(@"Live streaming started");
break;
case LFLiveStop:
NSLog(@"Live streaming stopped");
break;
case LFLiveError:
NSLog(@"Live streaming error");
break;
case LFLiveRefresh:
NSLog(@"Live streaming refreshing");
break;
}
}
完整示例:
完整的 ViewController.m
看起来如下:
#import "ViewController.h"
#import <LFLiveKit/LFLiveKit.h>
@interface ViewController () <LFLiveSessionDelegate>
@property (nonatomic, strong) LFLiveSession *liveSession;
@end
@implementation ViewController
- (void)viewDidLoad {
[super viewDidLoad];
[self requestAccessForVideo];
[self requestAccessForAudio];
[self startLive];
}
- (LFLiveSession*)liveSession {
if (!_liveSession) {
LFLiveAudioConfiguration *audioConfiguration = [LFLiveAudioConfiguration defaultConfiguration];
LFLiveVideoConfiguration *videoConfiguration = [LFLiveVideoConfiguration defaultConfiguration];
_liveSession = [[LFLiveSession alloc] initWithAudioConfiguration:audioConfiguration videoConfiguration:videoConfiguration];
_liveSession.delegate = self;
_liveSession.preView = self.view;
}
return _liveSession;
}
- (void)requestAccessForVideo {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeVideo];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeVideo completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.liveSession setRunning:YES];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.liveSession setRunning:YES];
}
}
- (void)requestAccessForAudio {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeAudio];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeAudio completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.liveSession setRunning:YES];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.liveSession setRunning:YES];
}
}
- (void)startLive {
LFLiveStreamInfo *streamInfo = [LFLiveStreamInfo new];
streamInfo.url = @"rtmp://your_server/live_stream";
[self.liveSession startLive:streamInfo];
}
- (void)stopLive {
[self.liveSession stopLive];
}
- (void)liveSession:(LFLiveSession *)session liveStateDidChange:(LFLiveState)state {
switch (state) {
case LFLiveReady:
NSLog(@"Ready to start live streaming");
break;
case LFLivePending:
NSLog(@"Connecting...");
break;
case LFLiveStart:
NSLog(@"Live streaming started");
break;
case LFLiveStop:
NSLog(@"Live streaming stopped");
break;
case LFLiveError:
NSLog(@"Live streaming error");
break;
case LFLiveRefresh:
NSLog(@"Live streaming refreshing");
break;
}
}
@end
二、PLMediaStreamingKit
1. PLMediaStreamingKit 安装
使用 CocoaPods
安装:
pod 'PLMediaStreamingKit'
运行 pod install
之后,在项目的任意位置导入 PLMediaStreamingKit
:
#import <PLMediaStreamingKit/PLMediaStreamingKit.h>
2. 配置和使用
创建推流会话
PLMediaStreamingSession
是此框架的核心类,用于音视频捕获、编码和推流。
- (PLMediaStreamingSession *)streamingSession {
if (!_streamingSession) {
PLVideoCaptureConfiguration *videoConfiguration = [PLVideoCaptureConfiguration defaultConfiguration];
PLAudioCaptureConfiguration *audioConfiguration = [PLAudioCaptureConfiguration defaultConfiguration];
PLVideoStreamingConfiguration *videoStreamingConfiguration = [PLVideoStreamingConfiguration defaultConfiguration];
PLAudioStreamingConfiguration *audioStreamingConfiguration = [PLAudioStreamingConfiguration defaultConfiguration];
_streamingSession = [[PLMediaStreamingSession alloc] initWithVideoCaptureConfiguration:videoConfiguration
audioCaptureConfiguration:audioConfiguration
videoStreamingConfiguration:videoStreamingConfiguration
audioStreamingConfiguration:audioStreamingConfiguration];
_streamingSession.delegate = self;
_streamingSession.previewView = self.view;
}
return _streamingSession;
}
检查并请求权限
和 LFLiveKit 类似,我们需要请求相机和麦克风的权限:
- (void)requestAccessForVideo {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeVideo];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeVideo completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.streamingSession startCaptureSession];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.streamingSession startCaptureSession];
}
}
- (void)requestAccessForAudio {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeAudio];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeAudio completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.streamingSession startCaptureSession];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.streamingSession startCaptureSession];
}
}
开始直播
创建一个 PLStream
对象,包含推流的 URL 和其他配置信息,并启动推流。
- (void)startStreaming {
PLStream *stream = [PLStream new];
stream.url = @"rtmp://your_server/live_stream";
[self.streamingSession startWithStream:stream feedback:^(PLStreamStartStateFeedback *feedback) {
if (feedback.state == PLStreamStartStateSuccess) {
NSLog(@"Streaming Started Successfully");
} else {
NSLog(@"Failed to start streaming: %@", feedback.error.localizedDescription);
}
}];
}
- (void)stopStreaming {
[self.streamingSession stop];
}
处理推流状态变化
通过实现 PLMediaStreamingSessionDelegate
的相关方法,可以监测推流状态的变化。
- (void)mediaStreamingSession:(PLMediaStreamingSession *)session streamStatusDidUpdate:(PLStreamStatus *)status {
NSLog(@"Stream status: %@", status);
}
- (void)mediaStreamingSession:(PLMediaStreamingSession *)session didDisconnectWithError:(NSError *)error {
NSLog(@"Stream disconnected with error: %@", error.localizedDescription);
}
完整示例:
完整的 ViewController.m
可以如下:
#import "ViewController.h"
#import <PLMediaStreamingKit/PLMediaStreamingKit.h>
@interface ViewController () <PLMediaStreamingSessionDelegate>
@property (nonatomic, strong) PLMediaStreamingSession *streamingSession;
@end
@implementation ViewController
- (void)viewDidLoad {
[super viewDidLoad];
[self requestAccessForVideo];
[self requestAccessForAudio];
}
- (PLMediaStreamingSession *)streamingSession {
if (!_streamingSession) {
PLVideoCaptureConfiguration *videoConfiguration = [PLVideoCaptureConfiguration defaultConfiguration];
PLAudioCaptureConfiguration *audioConfiguration = [PLAudioCaptureConfiguration defaultConfiguration];
PLVideoStreamingConfiguration *videoStreamingConfiguration = [PLVideoStreamingConfiguration defaultConfiguration];
PLAudioStreamingConfiguration *audioStreamingConfiguration = [PLAudioStreamingConfiguration defaultConfiguration];
_streamingSession = [[PLMediaStreamingSession alloc] initWithVideoCaptureConfiguration:videoConfiguration
audioCaptureConfiguration:audioConfiguration
videoStreamingConfiguration:videoStreamingConfiguration
audioStreamingConfiguration:audioStreamingConfiguration
stream:nil];
_streamingSession.delegate = self;
_streamingSession.previewView = self.view;
}
return _streamingSession;
}
- (void)requestAccessForVideo {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeVideo];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeVideo completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.streamingSession startCaptureSession];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.streamingSession startCaptureSession];
}
}
- (void)requestAccessForAudio {
AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType:AVMediaTypeAudio];
if (status == AVAuthorizationStatusNotDetermined) {
[AVCaptureDevice requestAccessForMediaType:AVMediaTypeAudio completionHandler:^(BOOL granted) {
if (granted) {
dispatch_async(dispatch_get_main_queue(), ^{
[self.streamingSession startCaptureSession];
});
}
}];
} else if (status == AVAuthorizationStatusAuthorized) {
[self.streamingSession startCaptureSession];
}
}
- (void)startStreaming {
PLStream *stream = [PLStream new];
stream.url = @"rtmp://your_server/live_stream";
[self.streamingSession startWithStream:stream feedback:^(PLStreamStartStateFeedback *feedback) {
if (feedback.state == PLStreamStartStateSuccess) {
NSLog(@"Streaming Started Successfully");
} else {
NSLog(@"Failed to start streaming: %@", feedback.error.localizedDescription);
}
}];
}
- (void)stopStreaming {
[self.streamingSession stop];
}
- (void)mediaStreamingSession:(PLMediaStreamingSession *)session streamStatusDidUpdate:(PLStreamStatus *)status {
NSLog(@"Stream status: %@", status);
}
- (void)mediaStreamingSession:(PLMediaStreamingSession *)session didDisconnectWithError:(NSError *)error {
NSLog(@"Stream disconnected with error: %@", error.localizedDescription);
}
@end
标签:status,NSLog,self,iOS,直播,void,liveSession,streamingSession,107
From: https://www.cnblogs.com/chglog/p/18305877