General-purpose Large Language Models (LLMs) are frequently fine-tuned through supervised fine-tuning (SFT) to enhance performance in specific domains. Better results can be achieved by distilling the chain-of-thought of a larger model at the cost of numerous expensive calls and a much greater amount of data.
We propose a novel blueprint for efficient fine-tuning that uses reasoning only for complex data identified by entropy. Specifically, across two small open models (
Note: This is an ongoing research. If you want to reproduce the results from the EMNLP 2025 version, check out this tag.
- Download CoT entropy data for MMLU to 
data/out/cot_entropy - Download reasoning data for MMLU to 
data/out/reasoning_entropy 
Other datasets are included in the repo and also published on Huggingface:
uv run src/experiments/REPLACE_ME.py
@misc{goncharov2025complexityawarefinetuning,
      title={Complexity-aware fine-tuning}, 
      author={Andrey Goncharov and Daniil Vyazhev and Petr Sychev and Edvard Khalafyan and Alexey Zaytsev},
      year={2025},
      eprint={2506.21220},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2506.21220}, 
}