whisper.cpp and whisper_android

The C/C++ port provides the core inference engine that the Android implementation wraps with TensorFlow Lite for mobile deployment, making them complements in a performance-optimized stack rather than alternatives.

whisper.cpp
72
Verified
whisper_android
64
Established
Maintenance 25/25
Adoption 10/25
Maturity 16/25
Community 21/25
Maintenance 16/25
Adoption 10/25
Maturity 16/25
Community 22/25
Stars: 47,665
Forks: 5,311
Downloads:
Commits (30d): 160
Language: C++
License: MIT
Stars: 630
Forks: 106
Downloads:
Commits (30d): 2
Language: C++
License: MIT
No Package No Dependents
No Package No Dependents

About whisper.cpp

ggml-org/whisper.cpp

Port of OpenAI's Whisper model in C/C++

Optimized for resource-constrained environments through integer quantization, mixed-precision inference (F16/F32), and zero runtime memory allocations, enabling on-device ASR on mobile and embedded platforms. Leverages the GGML inference library with multi-platform GPU acceleration via Metal, Vulkan, CUDA, and Core ML, alongside CPU-optimized SIMD paths for ARM NEON, AVX, and POWER VSX architectures. Provides a minimal C API and supports deployment across iOS, Android, WebAssembly, Raspberry Pi, and standard desktop/server platforms.

About whisper_android

vilassn/whisper_android

Offline Speech Recognition with OpenAI Whisper and TensorFlow Lite for Android

Provides dual implementation paths via TensorFlow Lite Java and Native APIs, allowing developers to choose between ease of integration and optimized performance. Includes a Python conversion pipeline to transform OpenAI Whisper models into TFLite format, plus support for live streaming transcription through buffer-based audio input alongside file-based batch processing. The architecture handles multilingual models with configurable vocabulary filters and manages audio preprocessing at 16kHz mono format for inference compatibility.

Scores updated daily from GitHub, PyPI, and npm data. How scores work