Discover and explore top open-source AI tools and projects—updated daily.
Faster RNN Language Modeling Toolkit
Top 57.1% on SourcePulse
This toolkit provides a highly optimized implementation of Recurrent Neural Network Language Models (RNNLMs) designed for training on massive datasets with very large vocabularies. It targets researchers and practitioners in Automatic Speech Recognition (ASR) and Machine Translation (MT) who need to achieve state-of-the-art performance and scalability. The primary benefit is significantly faster training and inference speeds compared to other implementations, enabling practical application on real-world, large-scale problems.
How It Works
The core of faster-rnnlm leverages efficient implementations of various RNN architectures, including GRU variants, and supports advanced training techniques like Noise Contrastive Estimation (NCE) and Hierarchical Softmax (HS). NCE is particularly highlighted for its ability to scale training speed independently of vocabulary size, offering comparable or better results than HS, which can become computationally infeasible for large vocabularies. The toolkit also incorporates optimizations such as ReLU activation, diagonal initialization, RMSProp, and gradient clipping for improved training dynamics and performance.
Quick Start & Requirements
./build.sh
to download Eigen and compile the toolkit../rnnlm -rnnlm model_name -train train.txt -valid validation.txt -hidden 128 -hidden-type gru -nce 20 -alpha 0.01
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
3 years ago
Inactive