zai-org/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
Based on the README, here's a technical summary: Built on the GLM base architecture with Multi-Query Attention for efficient inference, ChatGLM2-6B expands context length to 32K tokens (8K in conversation) using FlashAttention, achieving 42% faster inference and reducing INT4 quantization memory from 6GB to support 8K token conversations. Trained on 1.4T bilingual tokens with hybrid objectives and human preference alignment, it integrates seamlessly with HuggingFace's transformers library and supports INT4/INT8 quantization for deployment on resource-constrained hardware.
15,645 stars. No commits in the last 6 months.
Stars
15,645
Forks
1,820
Language
Python
License
—
Category
Last pushed
Jun 27, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/zai-org/ChatGLM2-6B"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
hoochanlon/hamuleite
🏔️国立台湾大学、新加坡国立大学、早稻田大学、东京大学,中央研究院(台湾)以及中国重点高校及科研机构,社科、经济、数学、博弈论、哲学、系统工程类学术论文等知识库。
yuanjie-ai/ChatLLM
轻松玩转LLM兼容openai&langchain,支持文心一言、讯飞星火、腾讯混元、智谱ChatGLM等
JiauZhang/chatchat
Large Language Models Python API
cambrian-mllm/cambrian
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
li-plus/chatglm.cpp
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)