Small, efficient language models distilled from ChatGPT for research
Top 44.1% on sourcepulse
LaMini-LM offers a diverse collection of small, efficient language models distilled from ChatGPT and trained on a 2.58M instruction dataset. Targeting researchers and developers seeking performant, compact LLMs, it provides a range of architectures and sizes for various NLP tasks.
How It Works
LaMini-LM employs offline distillation from GPT-3.5-turbo, generating 2.58M instruction-response pairs using prompts from existing resources like Self-Instruct, P3, Flan, and Alpaca. This approach allows for the creation of smaller, more manageable models that retain significant instruction-following capabilities, making them suitable for resource-constrained environments.
Quick Start & Requirements
pip install -q transformers
pipeline()
.transformers
library.Highlighted Details
lm-evaluation-harness
.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The CC BY NC 4.0 license prohibits commercial use. The README notes that reported LLaMA results are not directly comparable due to insufficient detail for reproducible evaluation.
2 years ago
1 week