Language modeling and sentiment classification in PyTorch (deprecated, see Megatron-LM)
Top 36.1% on sourcepulse
This repository provides code for unsupervised language modeling and sentiment classification, targeting researchers and practitioners in NLP. It enables training state-of-the-art classification models on custom datasets and reproducing results from NVIDIA's large-scale pretraining and transfer learning papers.
How It Works
The project leverages PyTorch for implementing Transformer and mLSTM language models. It supports unsupervised pretraining on large text corpora, followed by transfer learning or end-to-end finetuning for classification tasks. Key advantages include mixed-precision (FP16) training and distributed, multi-GPU, multi-node training capabilities, building upon the NVIDIA APEx project for scalability and efficiency.
Quick Start & Requirements
python3 setup.py install
.Highlighted Details
Maintenance & Community
The project acknowledges contributions from Neel Kant, @csarofeen, and @Michael Carilli. It references the APEx GitHub page for utilities. The project is built using the Amazon review dataset collected by J. McAuley.
Licensing & Compatibility
The repository does not explicitly state a license in the provided README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
This repository is DEPRECATED. Users are directed to Megatron-LM for up-to-date code. To use this codebase, one must rely on tagged releases and ensure compatibility with the software and dependencies available at that specific date.
5 years ago
1 day