whisper-timestamped and Whisper-Finetune

The timestamped variant provides the inference capability that the fine-tuning tool enhances, making them complements—one extends Whisper's base transcription output with word-level timing while the other optimizes Whisper through custom training on domain-specific data.

whisper-timestamped
68
Established
Whisper-Finetune
56
Established
Maintenance 2/25
Adoption 22/25
Maturity 25/25
Community 19/25
Maintenance 6/25
Adoption 10/25
Maturity 16/25
Community 24/25
Stars: 2,778
Forks: 209
Downloads: 66,844
Commits (30d): 0
Language: Python
License: AGPL-3.0
Stars: 1,200
Forks: 213
Downloads:
Commits (30d): 0
Language: C
License: Apache-2.0
Stale 6m
No Package No Dependents

About whisper-timestamped

linto-ai/whisper-timestamped

Multilingual Automatic Speech Recognition with word-level timestamps and confidence

Builds on OpenAI's Whisper by applying Dynamic Time Warping to cross-attention weights for precise word-level timestamp alignment without additional inference steps. Includes optional Voice Activity Detection preprocessing to reduce hallucinations and provides per-word confidence scores alongside segment-level predictions. Designed as a drop-in extension compatible with any Whisper version, with memory-efficient processing for long audio files.

About Whisper-Finetune

yeyupiaoling/Whisper-Finetune

Fine-tune the Whisper speech recognition model to support training without timestamp data, training with timestamp data, and training without speech data. Accelerate inference and support Web deployment, Windows desktop deployment, and Android deployment

Implements parameter-efficient fine-tuning using LoRA adapters while maintaining compatibility with OpenAI's base Whisper models across all variants (tiny through large-v3-turbo). Provides dual inference acceleration paths through CTranslate2 and GGML quantization, enabling deployment across heterogeneous platforms without requiring model conversion—original Whisper checkpoints convert directly. Integrates with PyTorch/Transformers ecosystem and includes end-to-end tooling: training pipelines supporting mixed data conditions, evaluation harnesses, and turnkey deployment templates for web services, Windows desktop (native apps), and Android via JNI bindings.

Scores updated daily from GitHub, PyPI, and npm data. How scores work