flash-linear-attention and flame
Flash-linear-attention provides the optimized model implementations that flame is purpose-built to train at scale, making them complements that work together in a stack rather than alternatives.
About flash-linear-attention
fla-org/flash-linear-attention
🚀 Efficient implementations of state-of-the-art linear attention models
Provides PyTorch and Triton kernels for linear attention variants (RetNet, GLA, Mamba, RWKV, DeltaNet, and 20+ emerging architectures), optimized for CPU and GPU across NVIDIA, AMD, and Intel platforms. Includes fused operators, hybrid model support, and variable-length sequence handling to reduce memory overhead during training. Integrates with Hugging Face model hub and the companion `flame` training framework for distributed model development.
About flame
fla-org/flame
🔥 A minimal training framework for scaling FLA models
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work