shashikg/WhisperS2T
An Optimized Speech-to-Text Pipeline for the Whisper Model Supporting Multiple Inference Engine
Supports multiple inference backends (Original OpenAI, HuggingFace with FlashAttention2, CTranslate2, TensorRT-LLM) with intelligent audio batching, VAD integration, and dynamic time-length processing to reduce computation overhead. Incorporates hallucination-reduction heuristics and asynchronous large-file loading while simultaneously transcribing batched segments, enabling multi-language and multi-task decoding in a single batch pass.
541 stars. No commits in the last 6 months.
Stars
541
Forks
73
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Aug 27, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/shashikg/WhisperS2T"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
SYSTRAN/faster-whisper
Faster Whisper transcription with CTranslate2
oseiskar/autosubsync
Automatically synchronize subtitles with audio using machine learning
FL33TW00D/whisper-turbo
Cross-Platform, GPU Accelerated Whisper 🏎️
machinelearningZH/audio-transcription
Transcribe any audio or video file. Edit and view your transcripts in a standalone HTML editor.
saharmor/whisper-playground
Build real time speech2text web apps using OpenAI's Whisper https://openai.com/blog/whisper/