Multi-round dialogue dataset and models for chat language model training
Top 18.2% on sourcepulse
UltraChat provides a large-scale, diverse, and informative multi-round dialogue dataset and associated models (UltraLM) for training conversational AI. It is designed for researchers and developers aiming to build powerful chat language models with general conversational capabilities, offering a significant resource for advancing open-source LLM development.
How It Works
The dataset is constructed using LLMs to generate conversational turns, simulating user interactions across three sectors: "Questions about the World," "Writing and Creation," and "Assistance on Existent Materials." This approach leverages Turbo APIs and carefully designed prompts to create multi-turn dialogues, ensuring diversity in topics and complexity. The models are trained using BMTrain for acceleration.
Quick Start & Requirements
/UltraLM/recover.sh
to obtain final weights./UltraLM/chat_cli.sh
with your recovered model path..src/
(e.g., train_bm.py
for LLaMA with BMTrain, train.py
for GPT-J with OpenPrompt). Requires data downloaded to ./data
.Highlighted Details
text-davinci-003
on AlpacaEval with UltraRM.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day