lyj20071013/Sparse-MoE-Language-Model-v1
This repository contains an implementation of a Sparse Mixture of Experts (MoE) Language Model using PyTorch. The model is designed to handle large-scale text generation tasks efficiently by leveraging multiple expert networks and a routing mechanism to dynamically select the most relevant experts for each input.
No commits in the last 6 months.
Stars
2
Forks
1
Language
Python
License
—
Category
Last pushed
Mar 10, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/lyj20071013/Sparse-MoE-Language-Model-v1"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
yaserkl/RLSeq2Seq
Deep Reinforcement Learning For Sequence to Sequence Models
kefirski/pytorch_RVAE
Recurrent Variational Autoencoder that generates sequential data implemented with pytorch
ctr4si/A-Hierarchical-Latent-Structure-for-Variational-Conversation-Modeling
PyTorch Implementation of "A Hierarchical Latent Structure for Variational Conversation...
georgian-io/Multimodal-Toolkit
Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
nurpeiis/LeakGAN-PyTorch
A simple implementation of LeakGAN in PyTorch