Open-source dataset for finetuning LLMs with reasoning
Top 71.3% on sourcepulse
Raspberry aims to create an open-source toy dataset for finetuning Large Language Models (LLMs) with enhanced reasoning abilities. Targeting researchers and developers focused on improving LLM reasoning, it offers a structured approach to generating complex queries and corresponding Chain-of-Thought (CoT) and self-critique data.
How It Works
The project synthesizes 500 distinct, complex user queries across various domains requiring math, coding, logic, and planning skills. These queries are then used to generate CoT and self-critique data via automated prompting strategies, leveraging LLMs' inherent reasoning capabilities. The generated samples undergo cleaning and rectification using rubrics and grading techniques to ensure coherence and suitability for single-shot reasoning datasets.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
The project is initiated by daveshap. Further community engagement and scaling plans are mentioned, including seeking funding via Manifund.
Licensing & Compatibility
Limitations & Caveats
The project is described as a "toy dataset" and a "pilot" for proof of concept. Achieving near-SOTA performance is an ambitious goal for a small, toy dataset. The initial dataset size is 500 queries, which may be insufficient for robust finetuning across all targeted reasoning abilities.
9 months ago
1 day