argmaxinc/WhisperKit
On-device Speech Recognition for Apple Silicon
Wraps OpenAI's Whisper models in CoreML format optimized for Apple Silicon, enabling real-time streaming transcription with word-level timestamps and voice activity detection. Provides Swift SDK integration, a local HTTP server with Deepgram-compatible WebSocket API, and companion tools for speaker diarization and text-to-speech. Supports model customization through fine-tuning workflows and deployment to HuggingFace repositories for easy distribution across projects.
5,775 stars. Actively maintained with 3 commits in the last 30 days.
Stars
5,775
Forks
516
Language
Swift
License
MIT
Category
Last pushed
Mar 12, 2026
Commits (30d)
3
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/argmaxinc/WhisperKit"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
linto-ai/whisper-timestamped
Multilingual Automatic Speech Recognition with word-level timestamps and confidence
yeyupiaoling/Whisper-Finetune
Fine-tune the Whisper speech recognition model to support training without timestamp data,...
xenova/whisper-web
ML-powered speech recognition directly in your browser
vasistalodagala/whisper-finetune
Fine-tune and evaluate Whisper models for Automatic Speech Recognition (ASR) on custom datasets...
r0227n/flutter_whisper_kit
🎤 A Flutter plugin for running WhisperKit speech-to-text models on-device, powered by...