Fine-tuned LLaMA weights, recreated from Stanford Alpaca
Top 72.8% on sourcepulse
This repository provides the fully fine-tuned weights for the Point-Alpaca language model, a recreation of Stanford's Alpaca experiment. It targets researchers and developers looking to leverage a powerful instruction-following LLM, offering a significant improvement over the original Alpaca through extensive fine-tuning on a synthetic dataset.
How It Works
Point-Alpaca is a full fine-tune of the LLaMA model, trained for three epochs on an 8x A100 80GB setup. This extensive training process reduced the loss from approximately 2 to 0.5, aiming for enhanced instruction-following capabilities. The project distributes the fine-tuned weights as XOR-encrypted diffs to circumvent LLaMA's licensing restrictions, requiring users to possess the original LLaMA weights for decryption.
Quick Start & Requirements
pip3 install -r requirements.txt
python3 chat.py
original/7B/consolidated.00.pth
), Python 3, wget
or equivalent for downloading diffs.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The "encryption" is a simple XOR, not intended for security. Users must legally obtain and possess the original LLaMA weights to reconstruct the fine-tuned model. The README mentions 13B models are planned, but availability is uncertain.
2 years ago
1 day