Code for generative pre-training research paper
Top 20.8% on sourcepulse
This repository provides the code and models for the paper "Improving Language Understanding by Generative Pre-Training." It is intended for researchers and practitioners interested in generative pre-training for language understanding tasks, specifically demonstrating results on the ROCStories dataset.
How It Works
The project implements a generative pre-training approach for language models. The core idea is to leverage a transformer architecture and train it on a large corpus to learn general language understanding capabilities, which can then be fine-tuned for specific downstream tasks. The provided code focuses on reproducing the ROCStories Cloze Test results.
Quick Start & Requirements
python train.py --dataset rocstories --desc rocstories --submit --analysis --data_dir [path to data here]
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is archived and will not receive updates. The code's non-deterministic nature due to GPU operations may affect reproducibility.
6 years ago
Inactive