hiyouga/EasyR1
EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL
Leverages HybridEngine architecture and vLLM's SPMD mode for efficient distributed training across multiple algorithms (GRPO, DAPO, Reinforce++, ReMax, RLOO, GSPO, CISPO). Supports both language models (Llama3, Qwen2/2.5/3) and vision-language models (Qwen2-VL/2.5-VL/3-VL, DeepSeek-R1 distill) with padding-free training, LoRA fine-tuning, and multi-node Ray-based orchestration. Integrates with vLLM for inference and standard ML experiment trackers (Weights & Biases, SwanLab, MLflow, TensorBoard).
4,721 stars. Actively maintained with 4 commits in the last 30 days.
Stars
4,721
Forks
362
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 10, 2026
Commits (30d)
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/hiyouga/EasyR1"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
hud-evals/hud-python
OSS RL environment + evals toolkit
OpenRL-Lab/openrl
Unified Reinforcement Learning Framework
sail-sg/oat
🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning,...
opendilab/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
NVlabs/GDPO
Official implementation of GDPO: Group reward-Decoupled Normalization Policy Optimization for...