quaterion  by qdrant

Framework for fine-tuning similarity learning models

created 3 years ago
656 stars

Top 51.9% on sourcepulse

GitHubView on GitHub
Project Summary

Quaterion is a Python framework designed to fine-tune similarity learning models for tasks like semantic search, recommendations, and anomaly detection. It targets users needing to adapt pre-trained models to custom tasks efficiently, especially with limited data or computational resources, by avoiding slow, costly full model retraining.

How It Works

Quaterion leverages a caching mechanism to enable training with large batch sizes on modest hardware, significantly speeding up the fine-tuning process. It incorporates specially designed head layers that allow pre-trained models to specialize effectively even with small, custom datasets, mitigating the risk of catastrophic forgetting. The framework is built on PyTorch Lightning, inheriting its scalability, cost-efficiency, and reliability for complex training pipelines.

Quick Start & Requirements

  • Training: pip install quaterion
  • Inference Service: pip install quaterion-models
  • Prerequisites: PyTorch, PyTorch Lightning. GPU recommended for training.
  • Docs: https://github.com/qdrant-tech/quaterion
  • Tutorials: Fine-tuning NLP models, Fine-tuning CV models.

Highlighted Details

  • Enables training thousands of epochs with large batch sizes on laptop GPUs via built-in caching.
  • Compatible with small datasets, allowing effective fine-tuning with custom head layers.
  • Highly customizable architecture for sophisticated training pipelines.
  • Built on PyTorch Lightning for scalability and reliability.

Maintenance & Community

  • Community support via Discord.
  • Active development indicated by recent commits and releases.
  • Contact: info@qdrant.tech

Licensing & Compatibility

  • Licensed under Apache License, Version 2.0.
  • Permissive license suitable for commercial use and integration into closed-source projects.

Limitations & Caveats

The quaterion-models package is separate to minimize inference dependencies, requiring users to install the appropriate package based on their needs. Specific hardware requirements beyond a GPU for optimal training performance are not detailed.

Health Check
Last commit

3 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
2 stars in the last 90 days

Explore Similar Projects

Starred by Jeremy Howard Jeremy Howard(Cofounder of fast.ai) and Stas Bekman Stas Bekman(Author of Machine Learning Engineering Open Book; Research Engineer at Snowflake).

SwissArmyTransformer by THUDM

0.3%
1k
Transformer library for flexible model development
created 3 years ago
updated 7 months ago
Starred by Lewis Tunstall Lewis Tunstall(Researcher at Hugging Face), Lysandre Debut Lysandre Debut(Chief Open-Source Officer at Hugging Face), and
3 more.

FARM by deepset-ai

0%
2k
NLP framework for transfer learning with BERT & Co
created 6 years ago
updated 1 year ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Stefan van der Walt Stefan van der Walt(Core Contributor to scientific Python ecosystem), and
8 more.

litgpt by Lightning-AI

0.2%
13k
LLM SDK for pretraining, finetuning, and deploying 20+ high-performance LLMs
created 2 years ago
updated 1 week ago
Feedback? Help us improve.