Prompt optimizer for aligning LLMs without training
Top 85.2% on sourcepulse
Black-Box Prompt Optimization (BPO) addresses the challenge of aligning Large Language Models (LLMs) with human preferences without requiring model retraining. It offers a novel approach for users seeking to improve LLM output quality and safety by optimizing prompts.
How It Works
BPO operates by treating prompt engineering as an optimization problem. It leverages a separate preference model to iteratively refine prompts based on pairwise comparisons of LLM responses. This black-box approach avoids direct model fine-tuning, making it applicable to proprietary LLMs and reducing computational overhead.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
llm_finetuning
, DeepSpeed-Chat
, and LLaMA-Factory
.Licensing & Compatibility
Limitations & Caveats
The project is presented as research code, and the README includes #TODO
comments indicating areas requiring user modification before execution, suggesting it may not be fully production-ready.
1 year ago
1 week