Reasoning teacher via LLM fine-tuning (ACL 2023 paper)
Top 82.2% on sourcepulse
This repository provides the official code for "Large Language Models Are Reasoning Teachers," an ACL 2023 paper. It enables users to run Chain-of-Thought (CoT) reasoning on OpenAI models and fine-tune student models (like T5, Flan-T5, GPT-2) using the Fine-tune-CoT method on custom datasets and hardware.
How It Works
The project implements two main functionalities: OpenAI API experiments and custom experiments on user-provided hardware. For OpenAI, it leverages their API for data collection and fine-tuning. For custom experiments, it utilizes PyTorch Lightning and Hugging Face Transformers to fine-tune open-source models. The core methodology, Fine-tune-CoT, trains student models using reasoning data generated by larger "teacher" models, aiming to transfer complex reasoning capabilities.
Quick Start & Requirements
pip install -r requirements.txt
and python setup.py develop
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project has specific Python and PyTorch version requirements. While it supports custom open-source models, the primary focus and extensive data are geared towards OpenAI models. The README mentions "Needs update" for data organization patterns, suggesting potential ongoing development or changes.
4 months ago
1 day