我正在尝试使用Swift在iOS项目中使用Azure speech SDK实现语音识别,我遇到了语音识别完成功能(stopContinuousRecognition()
(阻塞应用程序UI几秒钟的问题,但没有内存或处理器负载或泄漏。我试图将此函数移到DispatchQueue.main.async {}
,但没有结果。也许有人面临这样的问题?有必要把它放在一个单独的线程中吗?为什么这个函数需要这么长时间才能完成?
编辑:很难提供工作示例,但基本上我在按下按钮时调用这个函数:
private func startListenAzureRecognition(lang:String) {
let audioFormat = SPXAudioStreamFormat.init(usingPCMWithSampleRate: 8000, bitsPerSample: 16, channels: 1)
azurePushAudioStream = SPXPushAudioInputStream(audioFormat: audioFormat!)
let audioConfig = SPXAudioConfiguration(streamInput: azurePushAudioStream!)!
var speechConfig: SPXSpeechConfiguration?
do {
let sub = "enter your code here"
let region = "enter you region here"
try speechConfig = SPXSpeechConfiguration(subscription: sub, region: region)
speechConfig!.enableDictation();
speechConfig?.speechRecognitionLanguage = lang
} catch {
print("error (error) happened")
speechConfig = nil
}
self.azureRecognition = try! SPXSpeechRecognizer(speechConfiguration: speechConfig!, audioConfiguration: audioConfig)
self.azureRecognition!.addRecognizingEventHandler() {reco, evt in
if (evt.result.text != nil && evt.result.text != "") {
print(evt.result.text ?? "no result")
}
}
self.azureRecognition!.addRecognizedEventHandler() {reco, evt in
if (evt.result.text != nil && evt.result.text != "") {
print(evt.result.text ?? "no result")
}
}
do {
try! self.azureRecognition?.startContinuousRecognition()
} catch {
print("error (error) happened")
}
}
当我再次按下按钮停止识别时,我调用的是这个功能:
private func stopListenAzureRecognition(){
DispatchQueue.main.async {
print("start")
// app blocks here
try! self.azureRecognition?.stopContinuousRecognition()
self.azurePushAudioStream!.close()
self.azureRecognition = nil
self.azurePushAudioStream = nil
print("stop")
}
}
此外,我正在使用麦克风的原始音频数据(recognizeOnce
非常适合第一个短语,所以音频数据一切都很好(
尝试先关闭流,然后停止连续识别:
azurePushAudioStream!.close()
try! azureRecognition?.stopContinuousRecognition()
azureRecognition = nil
azurePushAudioStream = nil
您甚至不需要异步执行。
至少这对我有用。