faster-whisper and WhisperS2T

Faster-whisper is a lower-level inference optimization library that WhisperS2T likely depends on or compares against as one of several possible backend inference engines for its higher-level speech-to-text pipeline abstraction.

faster-whisper
65
Established
WhisperS2T
46
Emerging
Maintenance 6/25
Adoption 15/25
Maturity 25/25
Community 19/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 20/25
Stars: 21,444
Forks: 1,752
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 541
Forks: 73
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
No risk flags
Stale 6m No Package No Dependents

About faster-whisper

SYSTRAN/faster-whisper

Faster Whisper transcription with CTranslate2

Reimplements OpenAI's Whisper using CTranslate2, a fast inference engine optimized for Transformer models, achieving up to 4× speedup with identical accuracy while reducing memory footprint. Supports 8-bit quantization for both CPU and GPU, batched inference for throughput optimization, and integrates with Hugging Face's distil-whisper checkpoints. Uses PyAV for audio decoding, eliminating external FFmpeg dependencies.

About WhisperS2T

shashikg/WhisperS2T

An Optimized Speech-to-Text Pipeline for the Whisper Model Supporting Multiple Inference Engine

Supports multiple inference backends (Original OpenAI, HuggingFace with FlashAttention2, CTranslate2, TensorRT-LLM) with intelligent audio batching, VAD integration, and dynamic time-length processing to reduce computation overhead. Incorporates hallucination-reduction heuristics and asynchronous large-file loading while simultaneously transcribing batched segments, enabling multi-language and multi-task decoding in a single batch pass.

Related comparisons

Scores updated daily from GitHub, PyPI, and npm data. How scores work