huggingface/alignment-handbook

Robust recipes to align language models with human and AI preferences

61
/ 100
Established

Implements a full post-training pipeline spanning continued pretraining, supervised fine-tuning, and preference alignment techniques including DPO, ORPO, and Constitutional AI. Training scripts support distributed training via DeepSpeed ZeRO-3 and parameter-efficient approaches (LoRA/QLoRA), with reproducible YAML-based recipes for models like Zephyr and SmolLM. Integrates with Hugging Face Hub for dataset and model management, supporting both human feedback and AI preference signals.

5,523 stars and 151 monthly downloads. No commits in the last 6 months. Available on PyPI.

Stale 6m
Maintenance 2 / 25
Adoption 15 / 25
Maturity 25 / 25
Community 19 / 25

How are scores calculated?

Stars

5,523

Forks

474

Language

Python

License

Apache-2.0

Last pushed

Sep 08, 2025

Monthly downloads

151

Commits (30d)

0

Dependencies

21

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/huggingface/alignment-handbook"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.