Discover and explore top open-source AI tools and projects—updated daily.
pointnetworkFine-tuned LLaMA weights, recreated from Stanford Alpaca
Top 71.9% on SourcePulse
This repository provides the fully fine-tuned weights for the Point-Alpaca language model, a recreation of Stanford's Alpaca experiment. It targets researchers and developers looking to leverage a powerful instruction-following LLM, offering a significant improvement over the original Alpaca through extensive fine-tuning on a synthetic dataset.
How It Works
Point-Alpaca is a full fine-tune of the LLaMA model, trained for three epochs on an 8x A100 80GB setup. This extensive training process reduced the loss from approximately 2 to 0.5, aiming for enhanced instruction-following capabilities. The project distributes the fine-tuned weights as XOR-encrypted diffs to circumvent LLaMA's licensing restrictions, requiring users to possess the original LLaMA weights for decryption.
Quick Start & Requirements
pip3 install -r requirements.txtpython3 chat.pyoriginal/7B/consolidated.00.pth), Python 3, wget or equivalent for downloading diffs.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The "encryption" is a simple XOR, not intended for security. Users must legally obtain and possess the original LLaMA weights to reconstruct the fine-tuned model. The README mentions 13B models are planned, but availability is uncertain.
2 years ago
1 day
young-geng
openlm-research
jzhang38
tloen
tatsu-lab
meta-llama