Python library for reinforcement learning with human feedback (RLHF)
Top 72.2% on sourcepulse
pykoi is a Python library designed to streamline the process of improving Large Language Models (LLMs) through Reinforcement Learning from Human Feedback (RLHF) and Retrieval-Augmented Generation (RAG). It offers a unified interface for collecting user feedback, fine-tuning models with RL, reward modeling, and comparing LLM performance, targeting researchers and developers seeking to enhance LLM capabilities with real-time user input.
How It Works
pykoi provides a modular approach to LLM improvement. For RAG, it allows users to upload documents to create context-aware chatbots, enabling source selection and saving modified responses for RLHF data collection. For RLHF, it facilitates fine-tuning LLMs using collected datasets, integrating human evaluative feedback to guide model learning. The library also includes features for easily comparing multiple LLMs interactively or on specific prompts.
Quick Start & Requirements
pip3 install "pykoi[rag]"
or pip3 install "pykoi[rlhf]"
. For GPU RAG with HuggingFace, use pip3 install "pykoi[rag, huggingface]"
.cu121
). RAG can run on CPU with OpenAI or Anthropic Claude2 APIs. RLHF requires a GPU.Highlighted Details
Maintenance & Community
The project is maintained by CambioML. Links to community channels or roadmaps are not explicitly provided in the README.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial use or closed-source linking is not detailed.
Limitations & Caveats
The provided demo links in the README appear to be broken or incorrectly labeled. The library is actively under development, with some features noted as "building now."
1 year ago
1 day