TaiPhone LLM
Collection
TaiPhone: A Phone-Scale LLM Rooted in Taiwanese Knowledge
•
4 items
•
Updated
TaiPhone is a low-cost, lightweight language model built for Traditional Chinese, with a strong focus on Taiwanese language, culture, and context. Trained on just 0.7 billion carefully curated tokens and enhanced with chat vector techniques, TaiPhone delivers superior performance compared to similarly sized open-source LLaMA-tuned 1B or 3B-scale LLMs. TaiPhone shows that with the right data, effective and culturally-aware models can be built at a fraction of the cost.
Model | TW-MCQ | MMLU-Redux |
---|---|---|
LLaMA3.2-1B-Instruct | 0.305 | 0.403 |
LLaMA3.2-1B-it-chinese-kyara | 0.360 | 0.405 |
LLaMA3.2-TaiPhone-1B-Instruct-v0.1 (Ours) | 0.375 | 0.421 |
Model | TW-MCQ | MMLU-Redux |
---|---|---|
LLaMA3.2-3B-Instruct | 0.442 | 0.569 |
LLaMA3.2-3B-it-chinese-kyara | 0.462 | 0.405 |
Llama-3.2-3B-F1-Instruct | 0.458 | 0.548 |
LLaMA3.2-TaiPhone-3B-Instruct-v0.1 (Ours) | 0.502 | 0.578 |
Model | writing | roleplay | reasoning | math | coding | extraction | stem | humanities |
---|---|---|---|---|---|---|---|---|
Llama-3.2-3B-Instruct | 4.2 | 3.9 | 4.1 | 4.3 | 4.9 | 3.8 | 4.0 | 4.3 |
Llama-3.2-3B-F1-Instruct | 5.5 | 6.9 | 4.2 | 3.9 | 3.8 | 4.7 | 5.2 | 7.6 |
Llama-3.2-Kyara-3B-it | 5.7 | 7.2 | 4.8 | 6.3 | 5.2 | 5.3 | 5.9 | 7.5 |
Llama-3.2-TaiPhone-3B-Instruct-v0.1 (Ours) | 5.5 | 5.8 | 4.9 | 5.0 | 5.0 | 3.8 | 4.5 | 7.3 |
Base model
meta-llama/Llama-3.2-1B