Qwen3-Embedding  by QwenLM

Text embedding and reranking model

created 1 month ago
1,138 stars

Top 34.4% on sourcepulse

GitHubView on GitHub
Project Summary

The Qwen3 Embedding model series offers a suite of proprietary text embedding and reranking models designed for diverse NLP tasks like retrieval, classification, and clustering. Targeting developers and researchers, it provides state-of-the-art performance across multiple benchmarks, leveraging the advanced multilingual and long-text understanding capabilities of the Qwen3 foundational models.

How It Works

This series builds upon dense foundational models, offering embedding and reranking capabilities in sizes ranging from 0.6B to 8B parameters. A key feature is Matryoshka Representation Learning (MRL) support, allowing flexible vector dimension definition. The models are also "instruction aware," enabling task-specific prompt engineering for performance boosts, with English instructions recommended for optimal multilingual results.

Quick Start & Requirements

  • Transformers: pip install transformers>=4.51.0
  • Sentence Transformers: pip install sentence-transformers>=2.7.0
  • vLLM: pip install vllm>=0.8.5
  • Dependencies: PyTorch, CUDA (recommended for acceleration).
  • Usage: Examples provided for Transformers, vLLM, and Sentence Transformers libraries.
  • Docs: Huggingface, ModelScope, Blog, Arxiv

Highlighted Details

  • The 8B embedding model achieved the #1 rank on the MTEB multilingual leaderboard (70.58 score as of June 5, 2025).
  • Supports over 100 languages, including programming languages, for robust multilingual and cross-lingual retrieval.
  • Offers models in 0.6B, 4B, and 8B parameter sizes for both embedding and reranking, balancing efficiency and effectiveness.
  • Instruction-aware design allows customization for specific tasks, languages, or scenarios, potentially improving performance by 1-5%.

Maintenance & Community

  • Developed by Alibaba Cloud.
  • Community support available via Discord.

Licensing & Compatibility

  • Proprietary model. Specific license details are not explicitly stated in the README but are typically governed by Alibaba Cloud's terms for their proprietary models. Commercial use should be verified.

Limitations & Caveats

  • Requires transformers>=4.51.0 to avoid a KeyError.
  • Flash Attention 2 is recommended for performance but requires compatible hardware.
  • The README mentions "proprietary model" without detailing licensing terms for commercial use.
Health Check
Last commit

2 weeks ago

Responsiveness

Inactive

Pull Requests (30d)
1
Issues (30d)
39
Star History
1,159 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Didier Lopes Didier Lopes(Founder of OpenBB), and
11 more.

sentence-transformers by UKPLab

0.2%
17k
Framework for text embeddings, retrieval, and reranking
created 6 years ago
updated 3 days ago
Feedback? Help us improve.