Next-Token Reasoning in LLM Pretraining | Boost AI Performance

The new wave of approaches to pretraining large language models (LLMs) is attracting increasing attention, especially thanks to a recent paper by researchers from Microsoft and Beijing University. In the classical setting, pretraining involves the model learning to predict the next token as accurately as possible, based on the context — essentially, it’s just trying to guess what comes next and then comparing its predictions to real data.

However, in this work, the authors propose a fundamentally different perspective. They introduce the concept called “next-token reasoning” — almost adding reinforcement learning directly during the early stages of training. The idea is simple: from the very beginning, even before the model becomes the final version, it learns not only to predict tokens but to think step-by-step and reason.

Specifically, during training, the model first forms a chain of logical thoughts or reasoning (coT), and only afterward produces the next token. The model is rewarded for accurately matching the correct sequence — effectively, internal reinforcement learning that encourages more meaningful reasoning.

These approaches deliver comparable performance outcomes: a model with, say, 14 billion parameters, achieves results close to what a twice as large model could produce. In terms of next-token prediction accuracy, such models outperform standard variants like R1-Distill-Qwen-14B and approach the results of more substantial counterparts such as R1-Distill-Qwen-32B.

Of course, training with this method requires significantly more computational resources and time. But an additional advantage is that, since the model has been trained to reason from early on, it is much better prepared for further reinforcement learning (RL) fine-tuning. As a result, models trained with reinforcement learning post-supervised training are many times more efficient compared to traditional models lacking this “reasoning” experience.

Overall, there is much to ponder, especially for those interested in the latest breakthroughs in neural network training. More details can be found on the arXiv link.

Page view 29.04 20:35 Page view /ai-blog/grok-4-20-test-model-real-trading-breakthrough-alphaarena/ 29.04 20:31 Page view 29.04 20:19 Page view /ai-blog/international-diplomatic-talks-china-ukraine-iran-more/ 29.04 20:19 Page view /ai-blog/cling-3-0-video-omni-updates-enhanced-content-creation/ 29.04 20:18 Page view /ai-blog/maximize-your-ai-investment-effective-chatgpt-requests-insights/ 29.04 20:17 Page view /ai-blog/nvidia-ai-open-source-leadership-surpassing-tech-giants-in-2025/ 29.04 20:13 Page view 29.04 20:10 Page view /ai-blog/free-6-month-photoshop-extension-easy-online-image-editing/ 29.04 20:10 Page view /ai-blog/hunyuan-gamecraft-neural-engine-ultra-responsive-ai-for-games/ 29.04 19:50