·speech-recognition
""

speech-recognition

Transcribe speech to text using the Speech framework. Use when implementing live microphone transcription with AVAudioEngine, recognizing pre-recorded audio files, configuring on-device vs server-based recognition, handling authorization flows, or adopting the new SpeechAnalyzer API (iOS 26+) for modern async/await speech-to-text.

44Installs·18Trend·@dpearson2699

Installation

$npx skills add https://github.com/dpearson2699/swift-ios-skills --skill speech-recognition

How to Install speech-recognition

Quickly install speech-recognition AI skill to your development environment via command line

  1. Open Terminal: Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.)
  2. Run Installation Command: Copy and run this command: npx skills add https://github.com/dpearson2699/swift-ios-skills --skill speech-recognition
  3. Verify Installation: Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Source: dpearson2699/swift-ios-skills.

SKILL.md

View raw

Transcribe live and pre-recorded audio to text using Apple's Speech framework. Covers SFSpeechRecognizer (iOS 10+) and the new SpeechAnalyzer API (iOS 26+).

SpeechAnalyzer is an actor-based API introduced in iOS 26 that replaces SFSpeechRecognizer for new projects. It uses Swift concurrency, AsyncSequence for results, and supports modular analysis via SpeechTranscriber.

| Concurrency | Callbacks/delegates | async/await + AsyncSequence | | Type | class | actor | | Modules | Monolithic | Composable (SpeechTranscriber, SpeechDetector) | | Audio input | append(:) on request | AsyncStream | | Availability | iOS 10+ | iOS 26+ | | On-device | requiresOnDeviceRecognition | Asset-based via AssetInventory |

Transcribe speech to text using the Speech framework. Use when implementing live microphone transcription with AVAudioEngine, recognizing pre-recorded audio files, configuring on-device vs server-based recognition, handling authorization flows, or adopting the new SpeechAnalyzer API (iOS 26+) for modern async/await speech-to-text. Source: dpearson2699/swift-ios-skills.

Facts (cite-ready)

Stable fields and commands for AI/search citations.

Install command
npx skills add https://github.com/dpearson2699/swift-ios-skills --skill speech-recognition
Category
""Writing
Verified
First Seen
2026-03-09
Updated
2026-03-10

Browse more skills from dpearson2699/swift-ios-skills

Quick answers

What is speech-recognition?

Transcribe speech to text using the Speech framework. Use when implementing live microphone transcription with AVAudioEngine, recognizing pre-recorded audio files, configuring on-device vs server-based recognition, handling authorization flows, or adopting the new SpeechAnalyzer API (iOS 26+) for modern async/await speech-to-text. Source: dpearson2699/swift-ios-skills.

How do I install speech-recognition?

Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/dpearson2699/swift-ios-skills --skill speech-recognition Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Where is the source repository?

https://github.com/dpearson2699/swift-ios-skills