Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ base_model:
|
|
19 |
|
20 |
## Introduction
|
21 |
|
22 |
-
FairyR1-14B-Preview, a highly efficient large-language-model (LLM) that matches or exceeds larger models on select tasks
|
23 |
|
24 |
As a member of the FairyR1 series, FairyR1-14B-Preview shares the same training data and process as FairyR1-32B. We strongly recommend using the FairyR1-32B, which achieves comparable performance in math and coding to deepseek-R1-671B with only 5% of the parameters.
|
25 |
For more details, please view the page of [**FairyR1-32B**](https://huggingface.co/PKU-DS-LAB/FairyR1-32B).
|
|
|
19 |
|
20 |
## Introduction
|
21 |
|
22 |
+
FairyR1-14B-Preview, a highly efficient large-language-model (LLM) that matches or exceeds larger models on select tasks. Built atop the DeepSeek-R1-Distill-Qwen-14B base, this model continues to utilize the 'distill-and-merge' pipeline from [TinyR1-32B-Preview](https://huggingface.co/qihoo360/TinyR1-32B-Preview) and [Fairy-32B](https://huggingface.co/PKU-DS-LAB/FairyR1-32B), combining task-focused fine-tuning with model-merging techniques—to deliver competitive performance with drastically reduced size and inference cost. This project was funded by NSFC, Grant 624B2005.
|
23 |
|
24 |
As a member of the FairyR1 series, FairyR1-14B-Preview shares the same training data and process as FairyR1-32B. We strongly recommend using the FairyR1-32B, which achieves comparable performance in math and coding to deepseek-R1-671B with only 5% of the parameters.
|
25 |
For more details, please view the page of [**FairyR1-32B**](https://huggingface.co/PKU-DS-LAB/FairyR1-32B).
|