Open-sourced LLM reaching LLaMA2 performance with limited resources
Top 38.2% on sourcepulse
JetMoE offers an open-source Mixture-of-Experts (MoE) large language model, JetMoE-8B, designed to achieve performance comparable to larger models like LLaMA2-7B with significantly reduced training costs (under $0.1M) and inference computation. It targets researchers and developers seeking efficient, high-performing LLMs that can be fine-tuned on consumer-grade hardware.
How It Works
JetMoE-8B utilizes a Mixture-of-Experts architecture, activating only 2.2 billion parameters during inference. This sparse activation drastically lowers computational requirements compared to dense models of similar capabilities, enabling faster inference and more accessible fine-tuning. The model is trained on publicly available datasets, making it suitable for academic and open-source applications.
Quick Start & Requirements
pip install -e .
transformers
:from transformers import AutoTokenizer, AutoModelForCausalLM
from jetmoe import JetMoEForCausalLM, JetMoEConfig
AutoConfig.register("jetmoe", JetMoEConfig)
AutoModelForCausalLM.register(JetMoEConfig, JetMoEForCausalLM)
tokenizer = AutoTokenizer.from_pretrained('jetmoe/jetmoe-8b')
model = AutoModelForCausalLM.from_pretrained('jetmoe/jetmoe-8b')
transformers
library.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1+ week