
Welcome back AI prodigies!
In todayâs Sunday Special:
đŠHow Does It Work?
đThe Winner Doesnât Matter
âïžHow and Why Was It Created?
đKey Takeaway
Read Time: 7 minutes
đKey Terms
Binary Digit (i.e., âBitâ): The smallest unit of data that a computer can process.
Generative AI (GenAI): Uses AI models trained on text, image, audio, video, or code to generate new content.
Large Language Models (LLMs): AI models pre-trained on vast amounts of data to generate human-like text.
Anthropomorphize: To give human traits, emotions, or behaviors to non-human things, like animals, objects, or machines.
đ©ș PULSE CHECK
Could you tell the difference between a human and a chatbot?
đŠHOW DOES IT WORK?
âThe Turing Test,â Explained.
This summer, OpenAIâs ChatGPT passed âThe Turing Test.â British mathematician Alan Turing developed âThe Turing Testâ in 1950 to measure a computerâs ability to exhibit human-like intelligence.
âThe Turing Testâ had three participants:
A Computer
A Human Foil
A Human Interrogator
The Human Interrogator attempts to determine which is the Computer and the Human Foil by asking a series of questions through a keyboard. If the Computer can consistently fool the Human Interrogator, itâs considered an intelligent, thinking entity.
âWe werenât even close to passing it in 2021. Then, OpenAIâs ChatGPT passed it,â said former PayPal CEO Peter Thiel. âThat was the Holy Grail of AI research for the previous 60 years.â
How Do You Pass It?
But whoâs the human interrogator? Whatâs required to pass it? Turing never answered these questions because âThe Turing Testâ wasnât meant to be a benchmark for comparing the performance of different AI models.
However, in his landmark study in 1950 titled âComputing Machinery and Intelligence,â he predicted that by 2000, machines with a capacity of 1 billion Bits of memory would beat a Human Interrogator 30% of the time after five minutes of questioning.
Remarkably, part of his prediction came true. a PC bought in 2000 contained a little more than a billion Bits. However, during The Loebner Prize in 2000, an annual competition where Human Interrogators interact with Human Foils and Computers, the Human Interrogators could easily distinguish between them. Even though Turingâs timeline was slightly off, he got the trends exactly right. Chatbots gradually became more sophisticated over time.
đTHE WINNER DOESNâT MATTER
ELIZA Computer Program, Explained.
Examples of machines deceiving humans stretch back to the 1960s, starting with a program vaguely resembling a modern chatbot. Launched in 1966 by Massachusetts Institute of Technology (MIT) Professor Joseph Weizenbaum, the ELIZA Computer Program simulated a conversation between a patient and their psychotherapist. Weizenbaum deliberately selected the psychotherapy context to avoid the challenge of equipping the computer program with extensive real-world knowledge. By reflecting on the patientâs statements, ELIZA could sustain a dialogue without requiring a deep understanding of factual information. Although this design aimed to highlight the superficial nature of communication between humans and machines, ELIZA often appeared intelligent enough to convince some patients it was human. In an interview, Weizenbaum recounted an instance where his secretary requested that he leave the room so she could have a genuine conversation with ELIZA. Reflecting on this, Weizenbaum expressed his surprise: âI hadnât realized that even brief encounters with a relatively simple computer program could evoke strong delusional perceptions in otherwise normal individuals.â
ELIZA > GPT-3.5?
Not only did ELIZAâs conversational ability turn heads in the 1960s, but it also impressed AI researchers today. A group of cognitive scientists from the University of California San Diego (UCSD) recently evaluated GPT-3.5, GPT-4, and ELIZA in a version of âThe Turing Test.â
After completing this version of âThe Turing Testâ 1,405 times:
GPT-3.5 tricked the Human Interrogator 20% of the time.
GPT-4 tricked the Human Interrogator 50% of the time.
ELIZA tricked the Human Interrogator 22% of the time.
So, how did a 60-year-old ELIZA beat GPT-3.5? Thatâs because ELIZA doesnât exhibit the behaviors we associate with todayâs LLMs. For instance, GPT-3.5 was fine-tuned to have a formal tone and not express opinions, which makes it appear less human.
Also, as chatbots become more advanced, so do our abilities to detect them. For this reason, âThe Turing Testâ canât serve as a benchmark for comparing the performance of different AI models. However, by examining how and why Turing created it, we can ask better questions about the capabilities of GenAI.
âïžHOW AND WHY WAS IT CREATED?
The Origin Story?
Despite popular belief, âThe Turing Testâ originally involved three participants:
A Man
A Woman
A Human Judge
The Human Judge attempts to determine who the Man and the Woman are based on written communication. Then, Turing replaced one of the participants with a Computer, shifting the question to whether the Human Judge could distinguish between a human and a Computer.
So, what exactly was Turing testing? Was the gender-guessing version of âThe Turing Testâ probing identity, deception, or performance? How do these objectives translate to the Computer version? Turing himself offered a clue:
âThe original question âCan machines think?â I believe it to be too meaningless to deserve discussion. Nevertheless, I believe that at the end of the century, the use of words and general educated opinions will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted.â
For Turing, it wasnât meant to settle the debate over whether machines can âthink.â The question is so vague and meaningless that itâs not worth debating. Instead, he predicted that people would eventually accept the idea of machines âthinkingâ without hesitation as societal norms evolve to accommodate this belief. Whether weâve already reached that point is up for debate. We anthropomorphize chatbots all the time now. For example, we might say ChatGPT is smart, but does this mean ChatGPT can think? Turing might say it doesnât matter because we treat chatbots like thinking entities.
Debates Show Historical Context?
To fully appreciate how âThe Turing Testâ came to be, it helps to view it through the lens of Turingâs debates with critics. In his historical analysis, âThe Turing Test Argument,â Bernardo Gonçalves frames Turingâs proposal in its social, cultural, and historical context, reconstructing a debate in the 1940s between Turing and three critics: mathematician Douglas Hatreee, neurosurgeon Geoffrey Jefferson, and polymath Michael Polanyi.
Hartree argued that computers were merely calculation engines, incapable of creativity or spontaneity. Turing countered by emphasizing the potential of learning machines, or âunorganized machines,â to adapt and grow in ways that surpassed simple programming.
Jefferson was perhaps Turingâs most formidable critic, insisting that machines could only be considered intelligent if they could demonstrate creative expression. He famously declared that intelligence required not just producing a poem or sonnet but understanding that one had written it.
Polanyi argued that human intelligence relies on tacit knowledge, which refers to an intuitive, unformalized understanding that machines canât replicate. Turingâs decision to focus on conversational abilities rather than rule-based activities like chess was likely influenced by this critique.
đKEY TAKEAWAY
Todayâs modern adaptations of âThe Turing Testâ obscure its original purpose. Turing reframed the debate, shifting focus from whether machines can think to whether it matters if they do.
If chatbots canât reliably pass âThe Turing Test,â it means weâre learning faster than them. If chatbots can, itâs worth remembering that the real measure of their intelligence isnât found in their ability to fool us. In the short term, itâs found in the value they provide. And in the long term, itâs found in the questions their success inspires us to ask about the nature of our creativity, intelligence, and consciousness.
đFINAL NOTE
FEEDBACK
How would you rate todayâs email?
â€ïžTAIP Review of The Week
âCan you explain Alan Turingâs contributions to AI? Iâm a big fan of the newsletter!â
REFER & EARN
đYour Friends Learn, You Earn!
{{rp_personalized_text}}
Refer 3 friends to learn how to đ·ââïžBuild Custom Versions of OpenAIâs ChatGPT.
Copy and paste this link to friends: {{rp_refer_url}}
