Speech
https://developer.apple.com/forums/tags/speech
Recognize spoken words in recorded or live audio using Speech.How to use the SpeechDetector Module
https://developer.apple.com/forums/thread/797544
I am trying to use SpeechDetector Module in Speech framework along with SpeechTranscriber. and it is giving me an error
Cannot convert value of type 'SpeechDetector' to expected element type 'Array.ArrayLiteralElement' (aka 'any SpeechModule')
Below is how I am using it
let speechDetector = Speech.SpeechDetector()
let transcriber = SpeechTranscriber(locale: Locale.current,
transcriptionOptions: [],
reportingOptions: [.volatileResults],
...Wed, 20 Aug 2025 03:20:57 GMTdoomdaveSpeechAnalyzer speech to text wwdc sample app
https://developer.apple.com/forums/thread/790108
I am using the sample app from:
https://developer.apple.com/videos/play/wwdc2025/277/?time=763
I installed this on an Iphone 15 Pro with iOS 26 beta 1. I was able to get good transcription with it. The app did crash sometimes when transcribing and I was going to post here with the details. I then installed iOS beta 2 and uninstalled the sample app. Now every time I try to run the sample app on the 15 Pro I get this message:
SpeechAnalyzer: Input loop ending with error: Error Domain=SFSpeechErrorDomain Code=10 "Cannot use modules with unallocated locales [en_US (fixed en_US)]" UserInfo={NSLocal...Wed, 25 Jun 2025 10:11:42 GMTafsasdfSpeechAnalyzer error "asset not found after attempted download" for certain languages
https://developer.apple.com/forums/thread/797835
I am trying to use the new SpeechAnalyzer framework in my Mac app, and am running into an issue for some languages.
When I call AssetInstallationRequest.downloadAndInstall() for some languages, it throws an error:
Error Domain=SFSpeechErrorDomain Code=1 "transcription.ar asset not found after attempted download."
The ".ar" appears to be the language code, which in this case was Arabic.
When I call AssetInventory.status(forModules:) before attempting the download, it is giving me a status of "downloading" (perhaps from an earlier attempt?). If this language was completely unsupported, I would e...Fri, 22 Aug 2025 03:36:22 GMTcvasselliSpeechAnalyzer > AnalysisContext lack of documentation
https://developer.apple.com/forums/thread/811083
I'm using the new SpeechAnalyzer framework to detect certain commands and want to improve accuracy by giving context. Seems like AnalysisContext is the solution for this, but couldn't find any usage example. So I want to make sure that I'm doing it right or not.
let context = AnalysisContext()
context.contextualStrings = [
AnalysisContext.ContextualStringsTag("commands"): [
"set speed level",
"set jump level",
"increase speed",
"decrease speed",
...
],
AnalysisContext.Cont...Tue, 23 Dec 2025 09:03:26 GMTalpaycliCarPlay: Voice Conversational Entitlement Details
https://developer.apple.com/forums/thread/821632
With the Voice Conversational Entitlement, can a CarPlay app establish a turn-based audio interface that operates in two modes:
Speaking mode:
Audio Session configured for playback
Buffered audio
Listening mode:
Switch Audio Session to .record or .playAndRecord
Activate SFSpeechRecognizer
And continue toggling back and forth. The app should listen for responses to questions or other audio cues, and assuming those answers are correct (based on analysis of results from SFSpeechRecognizer), continue this pattern of mode 1 and 2 alternating.
This appears to be a valid use of this entitleme...Sun, 05 Apr 2026 03:18:01 GMTTimTheFoolManAXSpeech Thread Crash SEGV_ACCERR
https://developer.apple.com/forums/thread/790615
Hi everyone,
I've encountered a rare and strange crash in my app that I can't consistently reproduce. The crash seems to occur deep within Apple's internal frameworks, and I can't pinpoint which line of my own code is causing it. Here's the crash stack trace:
#44 AXSpeech
SIGSEGV
SEGV_ACCERR
0 CoreFoundation ___CFCheckCFInfoPACSignature + 4
1 CoreFoundation _CFRunLoopSourceSignal + 28
2 Foundation _performQueueDequeue + 492
3 Foundation ___NSThreadPerformPerform + 88
4 CoreFoundation ___CFRUNLOOP_IS_CALLING_OUT_TO_A_SOURCE0_PERFORM_FUNCTION__ + 28
5 CoreF...Fri, 27 Jun 2025 02:47:11 GMTlinhenSpeechTranscriber/SpeechAnalyzer being relatively slow compared to FoundationModel and TTS
https://developer.apple.com/forums/thread/794720
So,
I've been wondering how fast a an offline STT -> ML Prompt -> TTS roundtrip would be.
Interestingly, for many tests, the SpeechTranscriber (STT) takes the bulk of the time, compared to generating a FoundationModel response and creating the Audio using TTS.
E.g.
InteractionStatistics:
- listeningStarted: 21:24:23 4480 2423
- timeTillFirstAboveNoiseFloor: 01.794
- timeTillLastNoiseAboveFloor: 02.383
- timeTillFirstSpeechDetected: 02.399
- timeTillTranscriptFinalized: 04.510
- timeTillFirstMLModelResponse: 04.938
...Sat, 26 Jul 2025 19:36:31 GMTBersaelorStrange crash in iOS AudioToolboxCore when using AVSpeechSynthesizer in iOS 16
https://developer.apple.com/forums/thread/731400
I'm getting Crashlytics crashes from some my users, deep in the Apple code:
Crashed: AXSpeech
EXC_BAD_ACCESS KERN_INVALID_ADDRESS 0x00000007ec54b360
0 libobjc.A.dylib 0x3c9c objc_retain_x8 + 16
1 AudioToolboxCore 0x99580 auoop::RenderPipeUser::~RenderPipeUser() + 112
2 AudioToolboxCore 0xe6090 -[AUAudioUnit_XPC internalDeallocateRenderResources] + 92
3 AVFAudio 0x90a0 AUInterfaceBaseV3::Uninitialize() + 60
4 AVFAudio 0x4cbe0 AVAudioEngineGraph::PerformCommand(AUGraphNodeBaseV3&, AVAudioEngineGraph:...Fri, 09 Jun 2023 16:49:34 GMTKiwi1212SpeechTranscriber not supported
https://developer.apple.com/forums/thread/806765
I've tried SpeechTranscriber with a lot of my devices (from iPhone 12 series ~ iPhone 17 series) without issues. However, SpeechTranscriber.isAvailable value is false for my iPhone 11 Pro.
https://developer.apple.com/documentation/speech/speechtranscriber/isavailable
I'am curious why the iPhone 11 Pro device is not supported. Are all iPhone 11 series not supported intentionally? Or is there any problem with my specific device?
I've also checked the supportedLocales, and the value is an empty array.
https://developer.apple.com/documentation/speech/speechtranscriber/supportedlocalesMon, 10 Nov 2025 06:08:50 GMTMYMYVideo Audio + Speech To Text
https://developer.apple.com/forums/thread/805817
Hello,
I am wondering if it is possible to have audio from my AirPods be sent to my speech to text service and at the same time have the built in mic audio input be sent to recording a video?
I ask because I want my users to be able to say "CAPTURE" and I start recording a video (with audio from the built in mic) and then when the user says "STOP" I stop the recording.Fri, 31 Oct 2025 18:17:49 GMTcaptadohAVAudioEngine fails to start during FaceTime call (error 2003329396)
https://developer.apple.com/forums/thread/814269
Is it possible to perform speech-to-text using AVAudioEngine to capture microphone input while being on a FaceTime call at the same time?
I tried implementing this, but whenever I attempt to start the AVAudioEngine while a FaceTime call is active, I get the following error:
“The operation couldn’t be completed. (OSStatus error 2003329396)”
I assume this might be due to microphone resource restrictions during FaceTime, but I’d like to confirm whether this limitation is at the system level or if there’s any possible workaround or entitlement that allows concurrent microphone access.
Has anyone...Fri, 30 Jan 2026 15:52:29 GMTguido732Building Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
https://developer.apple.com/forums/thread/819525
We built an open-source macOS menu bar app that turns speech into text
and pastes it into the active app — using SpeechAnalyzer for on-device
transcription, ScreenCaptureKit + Vision for screen-aware context,
and FluidAudio for speaker diarization in meeting mode.
Here's what we learned shipping it on macOS 26.
GitHub: github.com/Marvinngg/ambient-voice
Architecture
The app has two modes: hotkey dictation (press to talk, release to inject)
and meeting recording (continuous transcription with a floating panel).
Dictation Mode
Audio capture uses AVCaptureSession (more on why below).
The captured...Fri, 20 Mar 2026 05:46:25 GMTMarvinFangBuilding Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
https://developer.apple.com/forums/thread/819555
We built an open-source macOS menu bar app that turns speech into text
and pastes it into the active app — using SpeechAnalyzer for on-device
transcription, ScreenCaptureKit + Vision for screen-aware context,
and FluidAudio for speaker diarization in meeting mode.
Here's what we learned shipping it on macOS 26.
GitHub: github.com/Marvinngg/ambient-voice
Architecture
The app has two modes: hotkey dictation (press to talk, release to inject)
and meeting recording (continuous transcription with a floating panel).
Dictation Mode
Audio capture uses AVCaptureSession (more on why below).
The captured...Fri, 20 Mar 2026 05:38:07 GMTMarvinFangSpeechAnalyzer.start(inputSequence:) fails with _GenericObjCError nilError, while the same WAV succeeds with start(inputAudioFile:)
https://developer.apple.com/forums/thread/818005
I'm trying to use the new Speech framework for streaming transcription on macOS 26.3, and I can reproduce a failure with SpeechAnalyzer.start(inputSequence:).
What is working:
SpeechAnalyzer + SpeechTranscriber
offline path using start(inputAudioFile:finishAfterFile:)
same Spanish WAV file transcribes successfully and returns a coherent final result
What is not working:
SpeechAnalyzer + SpeechTranscriber
stream path using start(inputSequence:)
same WAV, replayed as AnalyzerInput(buffer:bufferStartTime:)
fails once replay starts with:
_GenericObjCError domain=Foundation._GenericObjCEr...Fri, 06 Mar 2026 17:01:14 GMTnicleCrashed: AXSpeech EXC_BAD_ACCESS KERN_INVALID_ADDRESS 0x000056f023efbeb0
https://developer.apple.com/forums/thread/743995
Application is getting Crashed: AXSpeech
EXC_BAD_ACCESS KERN_INVALID_ADDRESS 0x000056f023efbeb0
Crashed: AXSpeech
0 libobjc.A.dylib 0x4820 objc_msgSend + 32
1 libsystem_trace.dylib 0x6c34 _os_log_fmt_flatten_object + 116
2 libsystem_trace.dylib 0x5344 _os_log_impl_flatten_and_send + 1884
3 libsystem_trace.dylib 0x4bd0 _os_log + 152
4 libsystem_trace.dylib 0x9c48 _os_log_error_impl + 24
5 TextToSpeech 0xd0a8c _pcre2_xclass_8
6 TextToSpeech 0x3bc04 TTSSpeechUnitTestingMode
7 TextToSpeech ...Thu, 04 Jan 2024 07:19:07 GMTRupendraAXSpeech Crash
https://developer.apple.com/forums/thread/116490
I have a very terrible crash problem in my App when I use AVSpeechSynthesizer and I can't repetition it.Here is my code, It's a singleton- (void)stopSpeech {
if ([self.synthesizer isPaused]) {
return;
}
if ([self.synthesizer isSpeaking]) {
BOOL isSpeech = [self.synthesizer stopSpeakingAtBoundary:AVSpeechBoundaryImmediate];
if (!isSpeech) {
[self.synthesizer stopSpeakingAtBoundary:AVSpeechBoundaryWord];
}
}
self.stopBlock ? self.stopBlock() : nil;
}
-(AVSpeechSynthesizer *)synthesizer {
if (!_synthesizer) {
_synthesi...Mon, 13 May 2019 03:09:09 GMTcc412Crashed: AXSpeech
https://developer.apple.com/forums/thread/103612
Hello, My app is crashing a lot with this issue. I can't reproduce the problem but I can see it occurs at the user's devices. The Crashlytics report shows the following lines:Crashed: AXSpeech
0 libsystem_pthread.dylib 0x1824386bc pthread_mutex_lock$VARIANT$mp + 278
1 CoreFoundation 0x1826d3a34 CFRunLoopSourceSignal + 68
2 Foundation 0x18319ec90 performQueueDequeue + 468
3 Foundation 0x18325a020 __NSThreadPerformPerform + 136
4 CoreFoundation 0x1827b7404 __CFRUNLOOP_IS_CALLING_OUT_TO_A_SOURCE0_PERFORM...Sat, 02 Jun 2018 21:33:47 GMTCGRGAVSpeechSynthesizer read Mandarin as Cantonese(iOS 26 beta 3))
https://developer.apple.com/forums/thread/792744
In iOS 26, AVSpeechSynthesizer read Mandarin into Cantonese pronunciation.
No matter how you set the language, and change the settings of my phone system, it doesn't work.
let utterance = AVSpeechUtterance(string: "你好啊")
//let voice = AVSpeechSynthesisVoice(language: "zh-CN") // not work
let voice = AVSpeechSynthesisVoice(language: "zh-Hans") // not work too
utterance.voice = voice
et synth = AVSpeechSynthesizer()
synth.speak(utterance)Sat, 12 Jul 2025 02:16:20 GMThaozesAVSpeechSynthesizer system voices (SLA clarification)
https://developer.apple.com/forums/thread/812285
Hello,
I am building an iOS-only, commercial app that uses AVSpeechSynthesizer with system voices, strictly using the APIs provided by Apple. Before distributing the app, I want to ensure that my current implementation does not conflict with the iOS Software License Agreement (SLA) and is aligned with Apple’s intended usage.
For a better playback experience (more accurate estimation of utterance duration and smoother skip forward/backward during playback), I currently synthesize speech using:
AVSpeechSynthesizer.write(_:toBufferCallback:)
Converting the received AVAudioPCMBuffer buffers into ...Fri, 09 Jan 2026 10:20:14 GMTcuriosityofmineIs Speech framework allowed?
https://developer.apple.com/forums/thread/813709
Hello,
I want to use the Speech framework in my app. However, I found that if I want it to work offline, it must be downloaded separately on the device. Do I understand correctly that it is not allowed to use it in a Swift Student Challenge submission if English (as the speech language) must be downloaded by the tester on their device using the internet beforehand?Sun, 25 Jan 2026 00:52:13 GMT[email protected]