whisperX-FastAPI and tiny-openai-whisper-api
These are competitors—both provide FastAPI wrappers around Whisper-based speech-to-text models, differing primarily in that WhisperX adds speaker diarization and timestamp alignment capabilities on top of the base Whisper engine.
About whisperX-FastAPI
pavelzbornik/whisperX-FastAPI
FastAPI service on top of WhisperX
Provides modular speech processing services including transcription, speaker diarization, and transcript alignment via individual endpoints, with async SQLAlchemy task persistence supporting SQLite or PostgreSQL backends. Configurable Whisper model selection and compute precision (float16/int8) enables deployment across CUDA and CPU environments. Includes Kubernetes-ready health probes and Swagger UI documentation for integration into broader audio/video processing pipelines.
About tiny-openai-whisper-api
morioka/tiny-openai-whisper-api
OpenAI Whisper API-style local server, runnig on FastAPI
Wraps OpenAI's Whisper model and experimental GPT-4o audio APIs locally using FastAPI, enabling drop-in compatibility with OpenAI's audio endpoints for air-gapped or compliance-restricted environments. Supports speech-to-text transcription and multimodal chat completions with audio input, with configurable model selection via environment variables and Docker deployment. Includes experimental Dify integration and client examples across multiple platforms (curl, Python, PowerShell).
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work