Prodigy recipes for zero/few-shot learning via OpenAI GPT-3
Top 85.3% on sourcepulse
This repository provides recipes for integrating OpenAI's GPT-3 with Prodigy for efficient data annotation. It targets NLP practitioners and researchers looking to bootstrap annotation workflows using zero- and few-shot learning, enabling faster creation of high-quality datasets for training custom models.
How It Works
The core approach leverages OpenAI's LLMs to generate initial predictions for tasks like Named Entity Recognition (NER) and text classification. These predictions are then presented within the Prodigy annotation interface, allowing users to quickly curate them. Users can refine prompts and provide examples interactively, with corrections feeding back into the LLM's context for improved future predictions.
Quick Start & Requirements
python -m pip install prodigy -f https://XXXX-XXXX-XXXX-XXXX@download.prodi.gy
python -m pip install -r requirements.txt
OPENAI_ORG
and OPENAI_KEY
environment variables.Highlighted Details
Maintenance & Community
This repository is marked as archival, with its functionality moved to Prodigy and spaCy-llm for continued maintenance and upgrades.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility with commercial or closed-source projects would depend on the specific license chosen for the new, maintained versions in Prodigy.
Limitations & Caveats
The archival notice indicates that this repository is no longer actively maintained, with all features migrated to Prodigy. Users should refer to the Prodigy documentation for the latest implementations and support. OpenAI's prompt size limit (4079 tokens) restricts the complexity and length of prompts.
2 years ago
1 week