Date: Thursday, November 20, 2025
Time: 11:45 a.m. to 12:45 p.m.
Location: G01 Gates Hall
Title: LLM Reasoning Beyond Scaling
Abstract: Large reasoning models have demonstrated capabilities to solve competition-level math problems, answer “deep research” questions, and address complex coding needs. Much of this progress has been enabled by scaling of data: pre-training data to learn vast knowledge, fine-tuning data to learn natural language reasoning, and RL environments to refine that reasoning. In this talk, I will describe the current LLM reasoning paradigm, its boundaries, and the future of LLM reasoning beyond scaling. First, I will describe the state of reasoning models and where I think scaling can lead to some additional successes. I will then shift to discussing more fundamental issues with models that scale will not resolve in the next few years. I will touch on current limitations including generator-validator gaps, poor compositional generalization, limited creativity, and outdated knowledge. In all cases, fundamental limitations of LLMs or of supervised learning in general make these problems challenging, inviting future study and novel solutions beyond scaling.
