File size: 1,367 Bytes
d95b6a3 00df465 d95b6a3 22ab097 d95b6a3 00df465 d95b6a3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
license: mit
pipeline_tag: text-generation
---
<h1 align="center">
On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models
</h1>
<div align="center">
<a href="https://chenlong-clock.github.io">Charlie Zhang</a>, <a href="https://www.phontron.com">Graham Neubig</a>,
<a href="https://xiangyue9607.github.io">Xiang Yue</a>
Carnegie Mellon University, Language Technologies Institute
</div>
<div align="center">
[](https://arxiv.org/abs/2512.07783)
[](LICENSE)

</div>
This repository contains post-training related checkpoints in extrapolation tasks.
**Code:** [GitHub Repository](https://github.com/Interplay-LM-Reasoning/Interplay-LM-Reasoning)
## 📚 Citation
If you find this work or code useful, please consider citing:
```bibtex
@misc{zhang2025interplaypretrainingmidtrainingrl,
title={On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models},
author={Charlie Zhang and Graham Neubig and Xiang Yue},
year={2025},
eprint={2512.07783},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2512.07783},
}
``` |