Research paper on few-shot fine-tuning of language models
Top 48.5% on sourcepulse
LM-BFF provides a framework for improving few-shot fine-tuning of pre-trained language models. It targets researchers and practitioners working with limited labeled data, offering techniques to enhance model performance on downstream NLP tasks. The core benefit is achieving better few-shot learning capabilities through structured fine-tuning strategies.
How It Works
LM-BFF combines prompt-based fine-tuning with a novel pipeline for automating prompt generation and a refined strategy for incorporating demonstrations into the model's context. This approach aims to guide the language model more effectively with limited examples, leveraging the power of prompts and relevant demonstrations to improve generalization.
Quick Start & Requirements
pip install -r requirements.txt
./data/original
and run python tools/generate_k_shot_data.py
.python run.py --task_name SST-2 --data_dir data/k-shot/SST-2/16-42 --model_name_or_path roberta-large --few_shot_type prompt-demo ...
transformers
(version 3.4.0 recommended), pytorch
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
pytorch
, transformers
) may lead to results deviating from the paper, though trends should persist.2 years ago
1 week