sentiment-discovery  by NVIDIA

Language modeling and sentiment classification in PyTorch (deprecated, see Megatron-LM)

created 7 years ago
1,064 stars

Top 36.1% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides code for unsupervised language modeling and sentiment classification, targeting researchers and practitioners in NLP. It enables training state-of-the-art classification models on custom datasets and reproducing results from NVIDIA's large-scale pretraining and transfer learning papers.

How It Works

The project leverages PyTorch for implementing Transformer and mLSTM language models. It supports unsupervised pretraining on large text corpora, followed by transfer learning or end-to-end finetuning for classification tasks. Key advantages include mixed-precision (FP16) training and distributed, multi-GPU, multi-node training capabilities, building upon the NVIDIA APEx project for scalability and efficiency.

Quick Start & Requirements

Highlighted Details

  • Supports training and inference with FP16 for improved performance and reduced memory footprint.
  • Offers distributed training across multiple nodes and GPUs for large-scale model training.
  • Includes pretrained models for sentiment classification (SST, IMDB) and emotion classification (SemEval).
  • Provides scripts for sentiment classification, language modeling, transfer learning, and text generation.

Maintenance & Community

The project acknowledges contributions from Neel Kant, @csarofeen, and @Michael Carilli. It references the APEx GitHub page for utilities. The project is built using the Amazon review dataset collected by J. McAuley.

Licensing & Compatibility

The repository does not explicitly state a license in the provided README. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

This repository is DEPRECATED. Users are directed to Megatron-LM for up-to-date code. To use this codebase, one must rely on tagged releases and ensure compatibility with the software and dependencies available at that specific date.

Health Check
Last commit

5 years ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
4 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Alex Cheema Alex Cheema(Cofounder of EXO Labs), and
1 more.

recurrent-pretraining by seal-rg

0.1%
806
Pretraining code for depth-recurrent language model research
created 5 months ago
updated 2 weeks ago
Feedback? Help us improve.