YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

This is the distilled educational Qwen2.5-7B-Instruct based on EduBench.

Model Details

Model Name: EDU-Qwen2.5-7B

Model Type: Distilled instruction-tuned language model (7B parameters)

Base Model: Qwen2.5-7B-Instruct

Training Data

To fully leverage the strengths of different response generation models across various scenarios, we adopt a multi-source distillation pipeline. For each task, we select the best-performing model on the test set as the response generator, using it to answer educational domain questions and construct the training dataset for the distillation model. Through the distillation pipeline, we obtain a training set of 17,000 samples covering various subtasks across all 9 educational scenarios.

More details are provided in Appendix K of our paper

Performance

Framework

🫣Citation

If you find our benchmark, evaluation pipeline or models useful or interesting, please cite our paper.

@misc{xu2025edubenchcomprehensivebenchmarkingdataset,
      title={EduBench: A Comprehensive Benchmarking Dataset for Evaluating Large Language Models in Diverse Educational Scenarios}, 
      author={Bin Xu and Yu Bai and Huashan Sun and Yiguan Lin and Siming Liu and Xinyue Liang and Yaolin Li and Yang Gao and Heyan Huang},
      year={2025},
      eprint={2505.16160},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.16160}, 
}
Downloads last month
10
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DirectionAI/EDU-Qwen2.5-7B

Quantizations
2 models