Yi-1.5: upgraded open-source language model series
Top 58.7% on sourcepulse
Yi-1.5 is a suite of large language models (LLMs) offering enhanced capabilities in coding, math, reasoning, and instruction following. Targeting developers and researchers, it provides improved performance over its predecessor, Yi, with models available in 34B, 9B, and 6B parameter sizes.
How It Works
Yi-1.5 is built upon a foundation of continuous pre-training on a 500 billion token corpus, followed by fine-tuning on 3 million diverse samples. This extensive training regimen aims to bolster its proficiency in complex cognitive tasks while retaining strong language understanding and commonsense reasoning abilities.
Quick Start & Requirements
pip install -r requirements.txt
transformers
library. Models can be downloaded from Hugging Face, ModelScope, or WiseModel.transformers
for local execution on CUDA-enabled GPUs.ollama run yi:v1.5
.python demo/web_demo.py -c <your-model-path>
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The README does not specify hardware requirements for each model size or provide explicit benchmarks comparing Yi-1.5 against other leading LLMs.
8 months ago
1 week