kyegomez/Sophia

Effortless plugin and play Optimizer to cut model training costs by 50%. New optimizer that is 2x faster than Adam on LLMs.

39
/ 100
Emerging

Implements a second-order optimization algorithm using inexpensive stochastic Hessian diagonal estimation (via Hutchinson or Gauss-Newton-Bartlett methods) with gradient clipping to precondition parameter updates. Supports both SophiaG and DecoupledSophia variants that integrate directly into PyTorch training loops without architectural changes, requiring minimal hyperparameter tuning beyond learning rate adjustment (typically half of AdamW's value) and rho clipping bounds.

381 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

381

Forks

26

Language

Python

License

Apache-2.0

Last pushed

Jun 04, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/kyegomez/Sophia"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.