zai-org/GLM-ASR

GLM-ASR-Nano: A robust, open-source speech recognition model with 1.5B parameters

54
/ 100
Established

Built on a multimodal architecture compatible with Hugging Face Transformers and SGLang, GLM-ASR-Nano excels at low-volume and dialectal speech recognition—particularly Cantonese and regional Chinese variants—achieving 4.10% average error rate across challenging real-world benchmarks like meeting transcription. The model supports 17 languages through a unified chat-based interface, enabling flexible audio-to-text inference via standard APIs or direct model loading.

759 stars. Actively maintained with 1 commit in the last 30 days.

No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 13 / 25
Community 18 / 25

How are scores calculated?

Stars

759

Forks

70

Language

Python

License

Apache-2.0

Last pushed

Mar 06, 2026

Commits (30d)

1

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/zai-org/GLM-ASR"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.