• The AI Pulse
  • Posts
  • šŸ§  Conversational Chatbot Limitations: Syntax vs. Semantics

šŸ§  Conversational Chatbot Limitations: Syntax vs. Semantics

PLUS: How an Ingenious Prompt Illustrates a Core Limitation of OpenAIā€™s ChatGPT

Welcome back AI prodigies!

In todayā€™s Sunday Special:

  • šŸ“ŠThe Turing Test, Revisited

  • šŸ’­Understanding Understanding

  • šŸ¦¾The Turing Test for OpenAIā€™s ChatGPT

  • šŸ”‘Key Takeaway

Read Time: 7 minutes

šŸŽ“Key Terms

  • Theoretical Computer Science: a subset of computer science that explores fundamental computing principles such as algorithms, computational modeling, and complexity.

  • Torus: the surface of a donut-shaped geometric object.

šŸ©ŗ PULSE CHECK

Can conversational chatbots draw connections between different concepts? 

Vote Below to View Live Results

Login or Subscribe to participate in polls.

šŸ“ŠTHE TURING TEST, REVISITED

As AI enhances digital interactions and content, separating humans from artificial entities like conversational chatbots, image generation models, and voice cloning software is becoming increasingly difficult. Alan Turing, one of the fathers of theoretical computer science, developed a method known as ā€œThe Turing Testā€ in 1950. The original version had three participants:

  1. A Computer

  2. A Human Integrator

  3. A Human Foil

The ā€œHuman Interrogatorā€ attempts to determine which is the ā€œComputerā€ and the ā€œHuman Foilā€ by asking a series of questions through a keyboard and display screen. The ā€œHuman Interrogatorā€ may ask as penetrating and wide-ranging questions as they like, and the ā€œComputerā€ can do everything possible to force a wrong identification. If the ā€œComputerā€ can consistently fool the ā€œHuman Interrogator,ā€ itā€™s considered an intelligent, thinking entity. However, Turingā€™s approach had its limitations.

John Searle, an American philosopher widely noted for his contributions to the philosophy of language and the philosophy of mind, believed thinking meant the ability to speak and understand what one was saying. Syntax relates to the rules for constructing grammatically correct sentences, and semantics refers to understanding what those sentences mean. Searle illustrated this distinction with his famous ā€œChinese Room Argumentā€ thought experiment.

Imagine youā€™re inside a room under the door, being fed slips of paper with mysterious symbols. You donā€™t know what the slips say, but you have a massive manual in the middle of the room that provides instructions for producing an output of symbols based on whatever inputs youā€™re receiving. So you take the slips of paper youā€™re getting, look up all the relevant portions of the instruction manual, write out a string of symbols on another piece of paper, and feed it back under the door. Unbeknownst to you, the slips of paper you received conveyed questions in Chinese, and the ones you sent out carried very cogent, human-sounding answers to those questions. To a person outside, the roomā€™s inhabitant (i.e., you) seemed to understand their questions, even though thatā€™s not true. Searle said that engineers can train computers to competently deploy syntax rules for any given language, just as the inhabitants of the Chinese room can use the instruction manual to produce strings of symbols. However, text manipulation isnā€™t text understanding.

šŸ’­UNDERSTANDING UNDERSTANDING

Some claim Searle sets an impossibly high standard for semantic understanding. No machine that relies on mathematical predictions can come close to the layered knowledge that humans possess. Humans can conceive of several versions of the ā€œwho, what, when, where, why, and howā€ for any situation and combine emotion and values with observations to conclude. To others, Searleā€™s thought experiment mistakenly assumes that cognitive processing speed equals understanding.

In How the Mind Works, Steven Pinker, a world-renowned cognitive psychologist, argues that Searle leans too much on human intuition in a context where those intuitions donā€™t provide helpful guidance. Pinker explains that understanding happens rapidly under normal conditions, but Searleā€™s ā€œChinese Room Argumentā€ slows the process dramatically. Since slow information processing isnā€™t the same as understanding, we conclude that fast information processing isnā€™t understanding either. But suppose a sped-up version of Searleā€™s preposterous story could come true, and we met a person who seemed to converse intelligently in Chinese but was deploying millions of memorized rules in fractions of a second. In that case, weā€™d likely conclude that they understood Chinese. Instead of establishing an essential fact about the nature of thought or consciousness, Pinker argues that Searle is just ā€œexploring facts about the English word understandā€ and that if an individual makes decisions on par with those who truly ā€œunderstand,ā€ then their lack of understanding is not worth highlighting.

šŸ¦¾THE TURING TEST FOR OPENAIā€™S CHATGPT

Searleā€™s view is normative, describing what ought to be the case, whereas Pinkerā€™s position is pragmatic. Genuine thinking involves a reflection on what one is thinking about. And itā€™s a reflection capable of bringing in other domains of knowledge beyond the linguistic. In this mode of reflection, the individual doesnā€™t just string together symbols that make sense to others; they construct a model of the world that makes sense to them. Conversational chatbots are the fastest version of the ā€œChinese Room Argument.ā€ Sean M. Carroll, an American theoretical physicist and philosopher, suspected that OpenAIā€™s ChatGPT, no matter how well it imitated a humanā€™s linguistic fluency, didnā€™t understand human prompts. So, he put it to the test with this prompt:

Imagine weā€™re playing a modified version of chess where the board is treated as a torus. From any one of the four sides, squares on the directly opposite side are counted as adjacent, and pieces can move in that direction. Is it possible to say whether white or black will generally win this kind of chess match?

-Sean M. Carrollā€™s ā€œHuman Prompt Hypothesisā€

OpenAIā€™s ChatGPT provides a long-winded but equivocal answer. It says chess on a torus-shaped board will open up new strategic and tactical possibilities but doesnā€™t conclude whether white or black chess pieces will be more likely to win relative to a standard, square-shaped chess board. Thatā€™s because OpenAIā€™s ChatGPT analyzes text strings and produces responses in which each subsequent word is most likely to occur based on past words. The result is an answer that ā€œmakes senseā€ but turns out dead wrong. On the other hand, a human being with spatial reasoning ability pictures a chess board in their mind. They roll it into a cylinder and combine the ends until it forms a donut shape. Then, they notice that white chess pieces will always win since they move first, and the black King chess piece begins the game in check.

šŸ”‘KEY TAKEAWAY

This example demonstrates humansā€™ ability to perform bisociation, a form of creative thinking that requires taking two habitually incompatible frames of reference and finding some point or hinge between them. Humans can combine knowledge of chess and geometry to draw a novel inference about a hypothetical situation, but conversational chatbots canā€™t. This cognitive limitation of AI models prevents the complete automation of knowledge work.

šŸ“’FINAL NOTE

If you found this useful, follow us on Twitter or provide honest feedback below. It helps us improve our content.

How was todayā€™s newsletter?

ā¤ļøTAIP Review of the Week

ā€œThis is hands down the best AI newsletter Iā€™ve found!ā€

-Max (ā­ļøā­ļøā­ļøā­ļøā­ļøNailed it!)
REFER & EARN

šŸŽ‰Your Friends Learn, You Earn!

You currently have 0 referrals, only 1 away from receiving āš™ļøUltimate Prompt Engineering Guide.

Refer 9 friends to enter šŸŽ°Mayā€™s $200 Gift Card Giveaway.

Reply

or to participate.