alexiglad/EBT
PyTorch Code for Energy-Based Transformers paper -- generalizable reasoning and scalable learning
Energy-Based Transformers reformulate transformer inference as an energy minimization problem, enabling iterative refinement ("System 2 thinking") at every token prediction rather than just feed-forward generation. The approach scales favorably across multiple axes—data, model depth, parameters, and FLOPs—while improving generalization, demonstrating this through multi-modal experiments (NLP, vision, video). Built on PyTorch Lightning, it provides modular training/inference pipelines with support for distributed training, HuggingFace dataset integration, and W&B logging, along with minimal examples for quick experimentation.
613 stars.
Stars
613
Forks
85
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 01, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/alexiglad/EBT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
NVIDIA-NeMo/NeMo
A scalable generative AI framework built for researchers and developers working on Large...
vlm-run/vlmrun-hub
A hub for various industry-specific schemas to be used with VLMs.
HyperGAI/HPT
HPT - Open Multimodal LLMs from HyperGAI
yash9439/Falcon-Local-AI-Model
Explore this GitHub repository housing 3 versions of Falcon code for text generation. Each...
DorsaRoh/transformer-from-scratch
Complete transformer from scratch, using only numpy