Discover and explore top open-source AI tools and projects—updated daily.
IndicoDataSolutionsNLP finetuning library with scikit-learn style API
Top 47.9% on SourcePulse
This library provides a scikit-learn-like interface for fine-tuning state-of-the-art NLP models, targeting researchers and developers who need to adapt pre-trained models to specific downstream tasks. It simplifies the process of leveraging models like BERT, RoBERTa, and GPT for classification, regression, and sequence labeling.
How It Works
Finetune abstracts the complexity of transformer architectures and training loops behind a familiar Classifier.fit() API. Users select a base model (e.g., BERT, RoBERTa) and configure training parameters (learning rate, length, regularization) to fine-tune on custom datasets. It supports multi-stage fine-tuning, first on unlabeled data and then on labeled data, to maximize performance.
Quick Start & Requirements
pip3 install finetunepython3 -m spacy download enHighlighted Details
Classifier, Regressor, SequenceLabeler, Comparison, etc.DeploymentModel for optimizing serialized models for production.Maintenance & Community
The project is maintained by IndicoDataSolutions. Links to community channels or roadmaps are not explicitly provided in the README.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial use or closed-source linking is not detailed.
Limitations & Caveats
The library requires TensorFlow 1.x, which is deprecated. The README mentions tensorflow-gpu >= 1.14.0, indicating it is not compatible with TensorFlow 2.x.
2 months ago
1 day
facebookresearch