Chat & pretrained LLM by Alibaba Cloud
Top 2.4% on sourcepulse
Qwen provides a suite of large language models (LLMs) and chat models, including base models (Qwen, Qwen-1.8B, Qwen-7B, Qwen-14B, Qwen-72B) and their chat-tuned variants. Developed by Alibaba Cloud, these models are designed for a wide range of natural language processing tasks, from content creation and summarization to tool usage and agentic behavior, targeting researchers and developers.
How It Works
The Qwen models are pretrained on extensive multilingual datasets (up to 3 trillion tokens), focusing on Chinese and English across various domains. They employ techniques to support long context windows (up to 32K tokens) and offer various quantization methods (Int4, Int8, KV cache quantization) for improved efficiency. The chat models are further aligned with human preferences using SFT and RLHF, enabling conversational capabilities and tool integration.
Quick Start & Requirements
pip install -r requirements.txt
flash-attention
for performance.Highlighted Details
Maintenance & Community
QwenLM/Qwen
is noted as no longer actively maintained due to codebase differences with newer versions.Licensing & Compatibility
Limitations & Caveats
QwenLM/Qwen
is not actively maintained; users should refer to QwenLM/Qwen2
.auto-gptq
) may have version compatibility issues with transformers
and optimum
..cpp
, .cu
) might be necessary for specific functionalities.1 week ago
1 day