CLI tool for LLM finetuning, ablation, and unit testing
Top 42.9% on sourcepulse
This toolkit provides a config-based CLI for fine-tuning, ablating, and unit-testing open-source LLMs. It targets researchers and developers needing to manage complex LLM experimentation pipelines efficiently, offering a unified YAML configuration for prompts, models, optimization, and testing.
How It Works
The toolkit uses a YAML configuration file to define and execute LLM fine-tuning experiments. It supports various data sources (Hugging Face datasets, JSON, CSV) and allows custom prompt templating. Model configuration includes Hugging Face checkpoints, quantization (4-bit via bitsandbytes), and LoRA parameters. Flash Attention 2 can be enabled for performance gains. The system automatically manages experiment artifacts and allows for resuming interrupted runs.
Quick Start & Requirements
pipx install llm-toolkit
or pip install llm-toolkit
.pipx inject llm-toolkit flash-attn --pip-args=--no-build-isolation
or pip install flash-attn --no-build-isolation
.llmtune generate config
followed by llmtune run ./config.yml
.Highlighted Details
Maintenance & Community
Contributions are welcome. Further community and roadmap details are not explicitly provided in the README.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial use or closed-source linking is not detailed.
Limitations & Caveats
The README does not specify a license, which may impact commercial adoption. Detailed documentation on specific model compatibility beyond examples (Llama2, Mistral, Falcon) is not provided.
9 months ago
Inactive