Discover and explore top open-source AI tools and projects—updated daily.
arcee-aiLLM instruction enhancement framework
Top 99.6% on SourcePulse
<2-3 sentences summarising what the project addresses and solves, the target audience, and the benefit.> EvolKit is a framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Models (LLMs). It targets researchers and practitioners aiming to improve LLM instruction-following capabilities by leveraging open-source LLMs and an automated evolution process, offering an alternative to closed-source solutions.
How It Works
EvolKit employs a recurrent evolution strategy for instruction enhancement. It integrates with open-source LLMs via generators (like VLLM or OpenRouter) to create and refine instructions. The process involves an Analyzer for trajectory analysis, an Evaluator (using a reward model or a failure detector inspired by WizardLM) to assess instruction quality, and an Optimizer to refine the evolution methods for subsequent rounds, streamlining the fine-tuning data preparation pipeline.
Quick Start & Requirements
git clone https://github.com/arcee-ai/EvolKit.git), cd EvolKit, and install dependencies (pip install -r requirements.txt).export VLLM_BACKEND=http://your-vllm-backend-url:port/v1.python run_evol.py --dataset <dataset_name> --model <model_name> --generator <generator_type> --batch_size <int> --num_methods <int> --max_concurrent_batches <int> --evolve_epoch <int> --output_file <filename>.https://github.com/arcee-ai/EvolKit.git.Highlighted Details
Maintenance & Community
No specific details on contributors, sponsorships, community channels (Discord/Slack), or roadmaps are provided in the README.
Licensing & Compatibility
The README does not specify a license type or any compatibility notes for commercial use.
Limitations & Caveats
The framework requires LLMs that are proficient at generating structured content for parsing operations. The README advises against using a development set (--dev_set_size) for performance reasons, suggesting it significantly increases processing time.
1 year ago
Inactive
Ziems
ContextualAI
allenai