Research paper code for sequence-to-sequence models using deep reinforcement learning
Top 46.4% on sourcepulse
This repository provides a framework for applying Deep Reinforcement Learning (RL) techniques to sequence-to-sequence (seq2seq) models, primarily for abstractive text summarization. It addresses common seq2seq challenges like exposure bias and train/test inconsistency by integrating RL methods. The target audience includes researchers and practitioners in NLP and deep learning looking to leverage RL for improved seq2seq performance.
How It Works
The framework implements several RL approaches for seq2seq tasks, including Scheduled Sampling (with hard/soft argmax), End-to-End Backpropagation, Policy-Gradient with Self-Critic, and Actor-Critic methods using DDQN and Dueling Networks. These RL techniques aim to optimize seq2seq models directly for task-specific metrics (like ROUGE scores) rather than relying solely on maximum likelihood estimation, thereby mitigating exposure bias and improving generation quality.
Quick Start & Requirements
pip install -r python_requirements.txt
Highlighted Details
Maintenance & Community
The project is marked as "no longer actively maintained." Contributions are welcome via pull requests.
Licensing & Compatibility
LICENSE.txt
file is not directly linked).Limitations & Caveats
The project explicitly states it is "no longer actively maintained." The reliance on outdated TensorFlow (1.10.1) and Python (2.7) versions presents significant adoption hurdles and potential compatibility issues with current hardware and software ecosystems.
2 years ago
Inactive