JGLUE is a comprehensive benchmark for evaluating Japanese Natural Language Understanding (NLU) capabilities, designed to foster research in the Japanese language domain. It comprises six diverse tasks—text classification, sentence pair classification, and question answering—with multiple datasets for each, making it suitable for researchers and developers working with Japanese NLP models.
How It Works
JGLUE was constructed from scratch, avoiding translation from English benchmarks to ensure linguistic authenticity. It utilizes Yahoo! Crowdsourcing for data annotation and includes datasets like MARC-ja (text classification), JSTS (semantic textual similarity), JNLI (natural language inference), JSQuAD (reading comprehension), and JCommonsenseQA (commonsense reasoning). The benchmark provides detailed dataset descriptions and baseline performance scores using various Japanese BERT and RoBERTa models.
Quick Start & Requirements
transformers
library. Detailed instructions are available in fine-tuning/README.md
.preprocess/requirements.txt
. Fine-tuning requires significant computational resources typical for large language models.Highlighted Details
Maintenance & Community
Developed through a joint research project between Yahoo Japan Corporation and Kawahara Lab at Waseda University. A leaderboard was planned but the test set has been released.
Licensing & Compatibility
Limitations & Caveats
4 months ago
Inactive