LLM combining instruction depth with multi-turn conversation
Top 49.0% on sourcepulse
WizardVicunaLM is an experimental large language model that combines the dataset expansion techniques of WizardLM with the multi-turn conversation tuning methods of VicunaLM. It aims to improve upon VicunaLM by creating richer, more in-depth conversational data, targeting researchers and power users interested in exploring advanced LLM training methodologies.
How It Works
The project leverages WizardLM's approach of in-depth instruction expansion but reformats it into a continuous conversational format. This expanded conversational data, generated using ChatGPT 3.5, is then fine-tuned using Vicuna's v1.1 training methodology. This dual approach aims to create a model that excels in nuanced, multi-turn dialogues by building upon a foundation of deeply explored topics.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
This project is explicitly described as highly experimental and a proof of concept, not intended for actual usage. The benchmark data is based on informal GPT-4 scoring, not rigorous testing.
2 years ago
1 day