神刀安全网

iOS10新特性之Speech新框架

1、前言
在这之前,我相信大家都知道CIDetector能够做到人脸检测这一功能吧,注意是检测并不是识别,如果要做到识别的话也即区别人脸的异同就要用到OpenCV了,说了这么多就是要引出苹果在今年新推出的Speech新框架,能够做到语音识别~~~额,其实只是能够做到把一段语音转化成文字。

2、框架主要类

#import <Foundation/Foundation.h>  #import <Speech/SFSpeechRecognitionResult.h> #import <Speech/SFSpeechRecognitionRequest.h> #import <Speech/SFSpeechRecognitionTask.h> #import <Speech/SFSpeechRecognitionTaskHint.h> #import <Speech/SFSpeechRecognizer.h> #import <Speech/SFTranscriptionSegment.h> #import <Speech/SFTranscription.h>

3、怎么用

#import "NFSpeechViewController.h" #import <Speech/Speech.h>  @interface NFSpeechViewController ()<SFSpeechRecognitionTaskDelegate>  @property (nonatomic ,strong) SFSpeechRecognitionTask *recognitionTask; @property (nonatomic ,strong) SFSpeechRecognizer      *speechRecognizer; @property (nonatomic ,strong) UILabel                 *recognizerLabel;  @end  @implementation NFSpeechViewController  - (void)dealloc {     [self.recognitionTask cancel];     self.recognitionTask = nil; }  - (void)viewDidLoad {     [super viewDidLoad];     // Do any additional setup after loading the view.     self.view.backgroundColor = [UIColor whiteColor];      //0.0获取权限     //0.1在info.plist里面配置     /*      typedef NS_ENUM(NSInteger, SFSpeechRecognizerAuthorizationStatus) {      SFSpeechRecognizerAuthorizationStatusNotDetermined,      SFSpeechRecognizerAuthorizationStatusDenied,      SFSpeechRecognizerAuthorizationStatusRestricted,      SFSpeechRecognizerAuthorizationStatusAuthorized,      };      */     [SFSpeechRecognizer requestAuthorization:^(SFSpeechRecognizerAuthorizationStatus status) {         switch (status) {             case SFSpeechRecognizerAuthorizationStatusNotDetermined:                 NSLog(@"NotDetermined");                 break;             case SFSpeechRecognizerAuthorizationStatusDenied:                 NSLog(@"Denied");                 break;             case SFSpeechRecognizerAuthorizationStatusRestricted:                 NSLog(@"Restricted");                 break;             case SFSpeechRecognizerAuthorizationStatusAuthorized:                 NSLog(@"Authorized");                 break;             default:                 break;         }     }];      //1.创建SFSpeechRecognizer识别实例     self.speechRecognizer = [[SFSpeechRecognizer alloc] initWithLocale:[[NSLocale alloc] initWithLocaleIdentifier:@"zh_CN"]];      //2.创建识别请求     SFSpeechURLRecognitionRequest *request = [[SFSpeechURLRecognitionRequest alloc] initWithURL:[NSURL fileURLWithPath:[[NSBundle mainBundle] pathForResource:@"游子吟.mp3" ofType:nil]]];      //3.开始识别任务     self.recognitionTask = [self recognitionTaskWithRequest1:request]; }  - (SFSpeechRecognitionTask *)recognitionTaskWithRequest0:(SFSpeechURLRecognitionRequest *)request{     return [self.speechRecognizer recognitionTaskWithRequest:request resultHandler:^(SFSpeechRecognitionResult * _Nullable result, NSError * _Nullable error) {         if (!error) {             NSLog(@"语音识别解析正确--%@", result.bestTranscription.formattedString);         }else {             NSLog(@"语音识别解析失败--%@", error);         }     }]; }  - (SFSpeechRecognitionTask *)recognitionTaskWithRequest1:(SFSpeechURLRecognitionRequest *)request{     return [self.speechRecognizer recognitionTaskWithRequest:request delegate:self]; }  - (void)didReceiveMemoryWarning {     [super didReceiveMemoryWarning];     // Dispose of any resources that can be recreated. }  #pragma mark- SFSpeechRecognitionTaskDelegate  // Called when the task first detects speech in the source audio - (void)speechRecognitionDidDetectSpeech:(SFSpeechRecognitionTask *)task {  }  // Called for all recognitions, including non-final hypothesis - (void)speechRecognitionTask:(SFSpeechRecognitionTask *)task didHypothesizeTranscription:(SFTranscription *)transcription {  }  // Called only for final recognitions of utterances. No more about the utterance will be reported - (void)speechRecognitionTask:(SFSpeechRecognitionTask *)task didFinishRecognition:(SFSpeechRecognitionResult *)recognitionResult {     NSDictionary *attributes = @{                                  NSFontAttributeName:[UIFont systemFontOfSize:18],                                  };      CGRect rect = [recognitionResult.bestTranscription.formattedString boundingRectWithSize:CGSizeMake(self.view.bounds.size.width - 100, CGFLOAT_MAX) options:NSStringDrawingUsesLineFragmentOrigin attributes:attributes context:nil];     self.recognizerLabel.text = recognitionResult.bestTranscription.formattedString;     self.recognizerLabel.frame = CGRectMake(50, 120, rect.size.width, rect.size.height); }  // Called when the task is no longer accepting new audio but may be finishing final processing - (void)speechRecognitionTaskFinishedReadingAudio:(SFSpeechRecognitionTask *)task {  }  // Called when the task has been cancelled, either by client app, the user, or the system - (void)speechRecognitionTaskWasCancelled:(SFSpeechRecognitionTask *)task {  }  // Called when recognition of all requested utterances is finished. // If successfully is false, the error property of the task will contain error information - (void)speechRecognitionTask:(SFSpeechRecognitionTask *)task didFinishSuccessfully:(BOOL)successfully {     if (successfully) {         NSLog(@"全部解析完毕");     } }  #pragma mark- getter  - (UILabel *)recognizerLabel {     if (!_recognizerLabel) {         _recognizerLabel = [[UILabel alloc] initWithFrame:CGRectMake(50, 120, self.view.bounds.size.width - 100, 100)];         _recognizerLabel.numberOfLines = 0;         _recognizerLabel.font = [UIFont preferredFontForTextStyle:UIFontTextStyleBody];         _recognizerLabel.adjustsFontForContentSizeCategory = YES;         _recognizerLabel.textColor = [UIColor orangeColor];         [self.view addSubview:_recognizerLabel];     }     return _recognizerLabel; }  @end

上面一坨代码就是新框架的一个简单的不能在简单的例子,你可以复制代码试一试,你可以创建一个和我同名的类,这样直接拷贝上面代码就可以了。

4、步骤解析
首先、我们要做的就是获取权限,苹果对权限安全这一面是做了很大的改进的,通过iOS10关于权限的新特性不难看出来。

iOS10新特性之Speech新框架
权限.png

通过上图,我们需要在info.plist配置文件中添加对应的权限,这次iOS10更新之后我们终于不需要像以前那样Copy、Paste了,因为有输入提示了~~~~开心

其次、接下来就是创建SFSpeechRecognizer实例了
这里需要提一下的就是本地化语言的坑,如果你不知道

[[NSLocale alloc] initWithLocaleIdentifier:@"zh_CN"]

@”zh_CN”字符串的话,你可以使用

[SFSpeechRecognizer supportedLocales];

进行查看,选取合适的使用,这里@”zh”在iOS9之后就不是简体中文了,而是TW繁体中文,简体中文就要使用@”zh_CN”了!!!!

for (NSLocale *lacal in [SFSpeechRecognizer supportedLocales].allObjects) {         NSLog(@"countryCode:%@  languageCode:%@ ", lacal.countryCode, lacal.languageCode); }

第2步 创建识别请求,就简单了直接把要识别的语音url传进去就行了(这里使用的是本地资源)

第3步中开始识别任务我用的是delegate的方式在代理方法中去做处理,苹果还提供了block方式。对应的代码:

- (SFSpeechRecognitionTask *)recognitionTaskWithRequest0:(SFSpeechURLRecognitionRequest *)request{     return [self.speechRecognizer recognitionTaskWithRequest:request resultHandler:^(SFSpeechRecognitionResult * _Nullable result, NSError * _Nullable error) {         if (!error) {             NSLog(@"语音识别解析正确--%@", result.bestTranscription.formattedString);         }else {             NSLog(@"语音识别解析失败--%@", error);         }     }]; }  - (SFSpeechRecognitionTask *)recognitionTaskWithRequest1:(SFSpeechURLRecognitionRequest *)request{     return [self.speechRecognizer recognitionTaskWithRequest:request delegate:self]; }

怎么识别我已经在代码新的很清楚了,按照步骤来就可以了,注释也很清楚。
好啦~~
赶紧试试去吧~~

转载本站任何文章请注明:转载至神刀安全网,谢谢神刀安全网 » iOS10新特性之Speech新框架

分享到:更多 ()

评论 抢沙发

  • 昵称 (必填)
  • 邮箱 (必填)
  • 网址