zai-org/ChatGLM2-6B

ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型

47
/ 100
Emerging

Based on the README, here's a technical summary: Built on the GLM base architecture with Multi-Query Attention for efficient inference, ChatGLM2-6B expands context length to 32K tokens (8K in conversation) using FlashAttention, achieving 42% faster inference and reducing INT4 quantization memory from 6GB to support 8K token conversations. Trained on 1.4T bilingual tokens with hybrid objectives and human preference alignment, it integrates seamlessly with HuggingFace's transformers library and supports INT4/INT8 quantization for deployment on resource-constrained hardware.

15,645 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 21 / 25

How are scores calculated?

Stars

15,645

Forks

1,820

Language

Python

License

Last pushed

Jun 27, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/zai-org/ChatGLM2-6B"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.