LLMs Can’t Learn Maths & Reasoning: What Recent Research Reveals
The debate surrounding the ability of Large Language Models (LLMs) to truly learn mathematics and reasoning tasks has been a cornerstone of AI research. A recent paper has made significant strides in unpacking this mystery by employing causal analysis to examine how LLMs tackle arithmetic reasoning tasks. By identifying specific circuits responsible for arithmetic logic, researchers have provided a glimpse into how these models operate under the hood.
Let’s dive into this groundbreaking research and uncover its implications.
Defining Reasoning
In his seminal 2019 paper, “On the Measure of Intelligence,” François Chollet defines intelligence as "skill-acquisition efficiency," emphasizing adaptability and generalization over performance on specific tasks. Reasoning, within this context, involves deriving conclusions from principles or evidence, requiring both logical consistency and flexibility.
For LLMs, reasoning has always been a contentious subject. While these models can generate coherent text and solve structured problems, the mechanisms behind their reasoning remain elusive.
Types of Reasoning
Reasoning can be broadly classified into:
In mathematical contexts, LLMs appear to excel at deductive tasks within their training scope but struggle to generalize beyond.
Understanding Heuristics
A key takeaway from the research is that LLMs often rely on heuristics rather than true reasoning. Heuristics are mental shortcuts or rules of thumb that guide problem-solving but may not always yield accurate results.
For instance:
Breaking Down Black-Box AI Internals
The biggest challenge with LLMs is their “black-box” nature. Researchers have now begun using causal analysis to break this box open. By isolating subsets of the model responsible for specific behaviors—called circuits—they’ve started mapping how these models “reason” through tasks.
Recommended by LinkedIn
For arithmetic reasoning:
Mathematical Circuits
Circuits in LLMs are analogous to pathways in human brains. For arithmetic tasks:
An interesting finding was how these circuits handle carry-over logic in addition. Instead of understanding it conceptually, the circuits depend on patterns in training data to approximate results.
Understanding Circuits in More Detail
Using causal intervention techniques, researchers:
This revelation underscores the limitations of LLMs in acquiring true reasoning capabilities. They are exceptional imitators but lack the innate ability to reason abstractly like humans.
Conclusion
The recent research confirms that while LLMs excel at simulating reasoning and performing arithmetic tasks, their abilities are bounded by their reliance on heuristics and pre-defined circuits. This challenges the notion of “intelligence” in AI and calls for further exploration into building models that genuinely understand and generalize beyond their training.
Key Takeaways: