whisper.cpp and whisper_android
The C/C++ port provides the core inference engine that the Android implementation wraps with TensorFlow Lite for mobile deployment, making them complements in a performance-optimized stack rather than alternatives.
About whisper.cpp
ggml-org/whisper.cpp
Port of OpenAI's Whisper model in C/C++
Optimized for resource-constrained environments through integer quantization, mixed-precision inference (F16/F32), and zero runtime memory allocations, enabling on-device ASR on mobile and embedded platforms. Leverages the GGML inference library with multi-platform GPU acceleration via Metal, Vulkan, CUDA, and Core ML, alongside CPU-optimized SIMD paths for ARM NEON, AVX, and POWER VSX architectures. Provides a minimal C API and supports deployment across iOS, Android, WebAssembly, Raspberry Pi, and standard desktop/server platforms.
About whisper_android
vilassn/whisper_android
Offline Speech Recognition with OpenAI Whisper and TensorFlow Lite for Android
Provides dual implementation paths via TensorFlow Lite Java and Native APIs, allowing developers to choose between ease of integration and optimized performance. Includes a Python conversion pipeline to transform OpenAI Whisper models into TFLite format, plus support for live streaming transcription through buffer-based audio input alongside file-based batch processing. The architecture handles multilingual models with configurable vocabulary filters and manages audio preprocessing at 16kHz mono format for inference compatibility.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work