huggingface/distil-whisper

Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.

45
/ 100
Emerging

Knowledge distillation from OpenAI's Whisper model enables this variant to achieve competitive accuracy on English speech through parameter reduction and architectural optimization. Fully compatible with Hugging Face Transformers' `AutoModelForSpeechSeq2Seq` API, it supports both short-form (<30s) and long-form transcription via sequential windowing, with optional features like beam search, timestamp generation, and prompt-based conditioning through the standard generation interface.

4,056 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 19 / 25

How are scores calculated?

Stars

4,056

Forks

349

Language

Python

License

MIT

Last pushed

Jan 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/huggingface/distil-whisper"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.