huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
Implements a full post-training pipeline spanning continued pretraining, supervised fine-tuning, and preference alignment techniques including DPO, ORPO, and Constitutional AI. Training scripts support distributed training via DeepSpeed ZeRO-3 and parameter-efficient approaches (LoRA/QLoRA), with reproducible YAML-based recipes for models like Zephyr and SmolLM. Integrates with Hugging Face Hub for dataset and model management, supporting both human feedback and AI preference signals.
5,523 stars and 151 monthly downloads. No commits in the last 6 months. Available on PyPI.
Stars
5,523
Forks
474
Language
Python
License
Apache-2.0
Category
Last pushed
Sep 08, 2025
Monthly downloads
151
Commits (30d)
0
Dependencies
21
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/huggingface/alignment-handbook"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
agentscope-ai/Trinity-RFT
Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement...
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO &...
zjunlp/EasyEdit
[ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
hyunwoongko/nanoRLHF
nanoRLHF: from-scratch journey into how LLMs and RLHF really work.
PKU-Alignment/align-anything
Align Anything: Training All-modality Model with Feedback