
Welcome back AI prodigies!
In todayâs Sunday Special:
đThe Prelude
đWhat Is Reasoning?
đŹCan LLMs Reason?
đ€Can LRMs Reason?
đKey Takeaway
Read Time: 7 minutes
đKey Terms
Large Language Models (LLMs): AI Models pre-trained on vast amounts of data to generate human-like text.
Large Reasoning Models (LRMs): AI Models designed to mimic a humanâs decision-making abilities to solve complex, multi-step problems.
đ©ș PULSE CHECK
Can AI reason at all?
đTHE PRELUDE
Consider this simple logic puzzle: âJared has two brothers and two sisters. How many siblings does his sister Jenny have?â
If you said âfour,â youâre right! Most of us solve this type of question instantly without a second thought. But conversational chatbots struggle with logic puzzles like this one.
Whatâs causing them to struggle? The problem lies in their limited ability to reason. While conversational chatbots are great at generating human-like text, they donât truly understand the logic behind what theyâre generating.
So, what exactly is reasoning? How do LLMs work? How well do they reason? And can LRMs do any better?
đWHAT IS REASONING?
Philosophers divide Reasoning into three categories: Deductive, Inductive, and Abductive.
During the 4th century BC, ancient Greek philosopher Aristotle conceived of Deductive Reasoning and Inductive Reasoning in the Organon, a collection of six works on logical analysis.
During the late 19th century, American mathematician Charles Sanders Peirce defined a new logical process known as Abductive Reasoning.
Hereâs what makes each type of Reasoning distinct:
Deductive Reasoning: The process of deriving specific conclusions from general premises. If all the general premises are true, then the specific conclusion must also be true. For example, âAll mammals are warm-blooded; all whales are mammals; therefore, all whales are warm-blooded.â
Inductive Reasoning: The process of forming probable conclusions based on repeated observations. For example, the sun rising every day is something weâve always observed, so we expect it to rise again tomorrow. But technically, we canât be 100% sure because itâs based on repeated observations, not absolute proof.
Abductive Reasoning: The process of starting with an observation and seeking the most plausible explanation. For example, if you notice your lawn is wet, you might conclude that it rained last night. In other words, Abductive Reasoning pinpoints the most likely causes of what you observe.
We often combine different forms of Reasoning to solve everyday problems. For example, scientists use Abductive Reasoning to generate hypotheses that explain observations. Then, they employ Deductive Reasoning to derive testable experiments from those hypotheses. Next, they rely on Inductive Reasoning to generalize results from those testable experiments into broader theories.
So, where do LLMs fail in the landscape of Reasoning?
đŹCAN LLMs REASON?
⊿ 1ïžâŁ đŠŸHow Do LLMs Work?
An LLM is a sophisticated autocomplete machine trained on the entire Internet.
To train an LLM, developers essentially show it millions of sentences with the last word covered up (i.e., âThe fat cat sat on the {BLANK}.â) and have it guess what comes next.
Each time the LLM guesses wrong, it adjusts thousands of Weights, which are numerical values that help it decide which words or patterns are most important for making better guesses in the future.
In simple terms, Weights control how tens of thousands of words relate to each other within an LLM. These relationships help form the Neural Network (NN): a highly interdependent framework that processes all the words using two methods:
Attention Mechanisms calculate how much each word in a sentence should âpay attentionâ to every other word. Consider the following sentence: âMiami, coined the âMagic City,â has beautiful white-sand beaches.â In this case, the words âMiamiâ and âbeachesâ would pay more attention to each other because theyâre closely related.
Transformer Layers help further clarify the meaning of each word within a sentence. This process helps the LLM develop a deeper understanding of the context. Consider the following sentence: âThe cat chased the mouse.â In this case, it looks at the word âchasedâ and determines that âcatâ is important because itâs doing the chasing. It also determines that âmouseâ is important because itâs being chased. So, it understands that âchasedâ is connected to âcatâ and âmouse.â
⊿ 2ïžâŁ đ§ Reasoning Capabilities?
While LLMs excel at generating human-like text, their ability to Reason is fundamentally different from ours.
Hereâs how they perform across the three distinct types of Reasoning:
â Deductive Reasoning {Simulated}: When high-quality training datasets contain explicit logical structures (e.g., if PâQ and QâZ, then PâZ), LLMs can appear to perform Deductive Reasoning. But this is Mimicry, not a genuine logical deduction.
â Inductive Reasoning {Primary Mode}: LLMs are inherently incredible at Inductive Reasoning because theyâre designed to recognize patterns. For example, when processing âThe cat sat on the {BLANK},â it knows to focus heavily on âcatâ and âsatâ to predict âmatâ rather than âfatâ because it draws on patterns itâs seen from similar phrases to identify likely word pairings.
â Abductive Reasoning {Severely Limited}: LLMs struggle with Abductive Reasoning because they lack a true understanding of how the world works beyond patterns of words. Imagine an LLM walks into a room and sees a window open, a puddle of water on the floor, and a wet cat. The LLM might say: âMaybe someone spilled water on the floor then gave the cat a bath.â This explanation is grammatically correct and logically sound, but the LLM overlooks the most plausible explanation because it lacks a true understanding of how cats behave and how that behavior triggers cause-and-effect outcomes.
⊿ 3ïžâŁ đ§© Failure in Deductive Reasoning?
Letâs revisit the simple logic puzzle: âJared has two brothers and two sisters. How many siblings does his sister Jenny have?â
When OpenAIâs GPT-4o (âoâ for âomniâ) is asked to solve this logic puzzle, it falls short.
Since Jared is one of the brothers, the other brother and the two sisters are his siblings. So, Jared has 4 siblings.
Now, for Jenny, who is one of the sisters.
That means her total siblings are:
đJared
đThe Other Brother
đThe Other Sister
That makes 3 siblings for Jenny.
â Final Answer: Jenny has 3 siblings.
GPT-4o (âoâ for âomniâ) mistakenly included Jared as one of the two brothers. We intuitively know that Jaredâs two brothers exclude him.
Fortunately, developers recently introduced LRMs to overcome this limitation, and theyâre powering all of todayâs most advanced AI models.
đ€CAN LRMs REASON?
LRMs are built to plan and reason, not just generate human-like text. They achieve this by utilizing Test-Time Compute (TTC), which allocates more computing power during AI Inference: everything that happens after you enter your prompt.
TTC leverages CoT and RL:
Chain-of-Thought (CoT) to break down complex problems into manageable sub-problems. Then, to solve each manageable sub-problem and combine them into a complete solution.
Reinforcement Learning (RL) to mimic the âtrial-and-errorâ process humans use to learn, where decisions that lead to desired outcomes are reinforced.
Together, CoT and RL enable LRMs to mimic certain components of Abductive Reasoning by deconstructing complex problems into manageable sub-problems and reinforcing multiple potential solutions that are logically sound.
This process is what allows OpenAIâs OpenAI o4-mini, Anthropicâs Claude Opus 4, and Googleâs Gemini 2.5 Pro to mimic a humanâs decision-making capabilities and problem-solving capacity.
When you ask OpenAI o4-mini to solve the same simple logic puzzle, it passes with flying colors.
Jenny has four siblings. In addition to herself, the family consists of:
đJared
đTwo Brothers
đOne Other Sister
So, Jennyâs brothers {2}, her sister {1}, and Jared {1} make {4} siblings in total.
đKEY TAKEAWAY
LLMs are great at recognizing patterns, but they often fall short of understanding the logic behind those patterns. LRMs, which leverage TTC to deploy CoT and RL, allow advanced AI models to reason more like humans.
This matters because it brings us closer to confidently using advanced AI models in critical fields like law, finance, or medicine, where hallucinations can have serious consequences for peopleâs fundamental rights, safety, or health.
đFINAL NOTE
FEEDBACK
How would you rate todayâs email?
â€ïžTAIP Review of The Week
âI understood all of it on the first read!!â
REFER & EARN
đYour Friends Learn, You Earn!
{{rp_personalized_text}}
Share your unique referral link: {{rp_refer_url}}
