whisper-asr-webservice and whisper.api

These two tools are competitors, both offering an API to access a Whisper ASR model for speech-to-text transcription, with "ahmetoner/whisper-asr-webservice" being a more popular and general-purpose implementation, and "innovatorved/whisper.api" emphasizing user-level access support and a finetuned model.

whisper-asr-webservice
56
Established
whisper.api
55
Established
Maintenance 6/25
Adoption 10/25
Maturity 16/25
Community 24/25
Maintenance 16/25
Adoption 10/25
Maturity 16/25
Community 13/25
Stars: 3,202
Forks: 568
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 914
Forks: 38
Downloads:
Commits (30d): 22
Language: Python
License: MIT
No Package No Dependents
No Package No Dependents

About whisper-asr-webservice

ahmetoner/whisper-asr-webservice

OpenAI Whisper ASR Webservice API

Provides pluggable ASR engine selection (OpenAI Whisper, Faster Whisper, WhisperX) with support for speaker diarization, word-level timestamps, and voice activity detection. Containerized with Docker for CPU/GPU deployment via environment variables, integrates FFmpeg for diverse audio/video codec handling, and exposes a REST API with Swagger documentation for transcription, translation, and language identification tasks.

About whisper.api

innovatorved/whisper.api

This project provides an API with user level access support to transcribe speech to text using a finetuned and processed Whisper ASR model.

Implements asynchronous transcription with built-in concurrency control and request queuing via a FastAPI-based HTTP API, supporting quantized model variants (tiny.en.q5, base.en.q5) for efficient inference. Includes ffmpeg audio processing, token-based authentication for user access management, and Docker containerization for self-hosted deployment. The architecture uses uvicorn as the ASGI server with configurable parallel job limits via environment variables.

Scores updated daily from GitHub, PyPI, and npm data. How scores work