LLM for Traditional Mandarin, tailored for Taiwanese culture
Top 30.3% on sourcepulse
This repository provides TAME (TAiwan Mixture of Experts) LLMs, specifically fine-tuned for Traditional Mandarin and Taiwanese culture. It targets researchers and developers needing robust Mandarin language capabilities, offering state-of-the-art performance on local benchmarks and supporting diverse applications like chatbots, RAG, and structured data generation.
How It Works
The project leverages the Llama-3 architecture, fine-tuning it on a large corpus of Traditional Mandarin and English data. This includes specialized knowledge from legal, manufacturing, medical, and electronics domains. The models are trained using NVIDIA NeMo and Megatron on DGX H100 systems, with inference optimized via NVIDIA TensorRT-LLM, enabling efficient deployment and high performance.
Quick Start & Requirements
docker run --gpus '"all"' --rm -it winglian/axolotl:main-latest
) or direct execution (accelerate launch -m axolotl.cli.train example_training_config_for_finetuning_twllm.yaml
).pipeline("text-generation", model="yentinglin/Llama-3-Taiwan-70B-Instruct")
docker run ... vllm/vllm-openai:v0.4.0.post1 --model "yentinglin/Llama-3-Taiwan-70B-Instruct"
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
3 months ago
1 week