yaserkl/RLSeq2Seq
Deep Reinforcement Learning For Sequence to Sequence Models
Implements policy-gradient with self-critic learning and actor-critic methods (DDQN, dueling networks) to address exposure bias and train/test measurement inconsistency in seq2seq models. Built on TensorFlow 1.10.1, it supports scheduled sampling variants and intra-decoder attention mechanisms, with pre-processed CNN/Daily Mail and Newsroom datasets for abstractive text summarization tasks.
768 stars. No commits in the last 6 months.
Stars
768
Forks
161
Language
Python
License
MIT
Category
Last pushed
Mar 24, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/yaserkl/RLSeq2Seq"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
kefirski/pytorch_RVAE
Recurrent Variational Autoencoder that generates sequential data implemented with pytorch
ctr4si/A-Hierarchical-Latent-Structure-for-Variational-Conversation-Modeling
PyTorch Implementation of "A Hierarchical Latent Structure for Variational Conversation...
georgian-io/Multimodal-Toolkit
Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
nurpeiis/LeakGAN-PyTorch
A simple implementation of LeakGAN in PyTorch
facebookresearch/large_concept_model
Large Concept Models: Language modeling in a sentence representation space