Python tool for text-based AI training and generation
Top 24.0% on sourcepulse
This Python package provides a robust tool for text-based AI training and generation using GPT-2 and EleutherAI's GPT Neo/GPT-3 architectures. It's designed for researchers and developers looking to fine-tune or train custom language models efficiently, offering enhanced speed and memory usage compared to previous tools.
How It Works
Leveraging PyTorch, Hugging Face Transformers, and PyTorch Lightning, aitextgen
enables training on CPUs, multiple GPUs, and eventually TPUs. It supports OpenAI's GPT-2 models (124M to 774M parameters) and EleutherAI's GPT Neo models (125M to 350M parameters), or allows training from scratch with custom tokenizers and configurations. Its dataset handling includes caching, compression, and merging capabilities for efficient data management.
Quick Start & Requirements
pip3 install aitextgen
from aitextgen import aitextgen; ai = aitextgen(); ai.generate()
aitextgen generate
input.txt
) and optionally a custom tokenizer.Highlighted Details
gpt-2-simple
.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 years ago
Inactive