Pre-trained NLU models for task-oriented dialogue
Top 91.2% on sourcepulse
This repository provides pre-trained models and code for ToD-BERT, a natural language understanding model specifically designed for task-oriented dialogue systems. It addresses the linguistic differences between general text and dialogues, offering improved performance on downstream tasks like intent recognition, dialogue state tracking, and response selection, particularly in low-data scenarios. The target audience includes NLP researchers and developers working on conversational AI and task-oriented dialogue systems.
How It Works
ToD-BERT is pre-trained on a unified corpus of nine task-oriented dialogue datasets. It incorporates user and system tokens into the masked language modeling objective and introduces a contrastive objective function to simulate response selection. This dual-objective approach allows ToD-BERT to better capture dialogue-specific linguistic patterns and conversational dynamics compared to general-purpose language models.
Quick Start & Requirements
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("TODBERT/TOD-BERT-JNT-V1")
tod_bert = AutoModel.from_pretrained("TODBERT/TOD-BERT-JNT-V1")
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project was last updated in October 2020, indicating potential staleness. The license is not explicitly mentioned, which may pose a barrier for commercial adoption.
2 years ago
Inactive