Discover and explore top open-source AI tools and projects—updated daily.
oleg-yaroshevskiyAdvanced Q&A understanding for complex content
Top 100.0% on SourcePulse
Summary
This project offers a reproducible, 1st-place solution for the Google QUEST Q&A Labeling competition, enhancing automated understanding of complex question-answer content. It targets NLP researchers and practitioners seeking high-performance QA system methodologies. The primary benefit is a proven, top-scoring approach to tackling this specific NLP task.
How It Works
The approach involves finetuning language models on StackExchange data, followed by pseudo-label generation. An ensemble of BERT-base-cased, RoBERTa-base, and BART-large models is trained using both datasets. A 5-fold cross-validation strategy is applied per model type, with final predictions derived from averaging checkpoints and blending diverse model outputs. This ensemble and pseudo-labeling methodology is central to achieving top performance.
Quick Start & Requirements
Setup requires a Conda environment (conda create -n qa_quest_env python=3.6.6, conda activate qa_quest_env) and dependency installation via pip install -r requirements_full.txt or requirements_minimal.txt. Custom installations for mag and a modified fairseq are handled by bash/setup.sh.
bash/download_comp_data.sh) and ~18 GB of model checkpoints for inference (bash/download_all_model_ckpts_for_inference.sh).
The README references a Kaggle Notebook for inference reproduction.Highlighted Details
Maintenance & Community
Project maintained by oleg-yaroshevskiy; contact yury.kashnitsky@gmail.com for questions. No explicit community channels or roadmap links are provided.
Licensing & Compatibility
License type is not explicitly stated in the provided README content.
Limitations & Caveats
Requires specific, older versions of Python (3.6.6), CUDA (10.0.130), and cuDNN (7.5.0), posing potential compatibility challenges. Significant GPU hardware is necessary for training and efficient inference. Setup involves custom library installations and handling large model checkpoints (~18 GB).
4 years ago
Inactive
danqi
langchain-ai
paperswithcode
facebookresearch
llmware-ai