General-purpose Large Language Models (LLMs) are frequently fine-tuned through supervised fine-tuning (SFT) to enhance performance in specific domains. Better results can be achieved by distilling the chain-of-thought of a larger model at the cost of numerous expensive calls and a much greater amount of data.
We propose a novel blueprint for efficient fine-tuning that uses reasoning only for complex data identified by entropy. Specifically, across two small open models (
Note: This is an ongoing research. If you want to reproduce the results from the EMNLP 2025 version, check out this tag.
- Download CoT entropy data for MMLU to
data/out/cot_entropy
- Download reasoning data for MMLU to
data/out/reasoning_entropy
Other datasets are included in the repo and also published on Huggingface:
uv run src/experiments/REPLACE_ME.py
@misc{goncharov2025complexityawarefinetuning,
title={Complexity-aware fine-tuning},
author={Andrey Goncharov and Daniil Vyazhev and Petr Sychev and Edvard Khalafyan and Alexey Zaytsev},
year={2025},
eprint={2506.21220},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2506.21220},
}