- The AI Pulse
- Posts
- š§ AI Agents Talk, But We Canāt Understand?!
š§ AI Agents Talk, But We Canāt Understand?!
PLUS: How Modems and Telephone Lines Inspire Next-Gen AI

Welcome back AI prodigies!
In todayās Sunday Special:
šThe Prelude
šEarly Machine Communication
š¦¾Evolution of Agentic Communication
šļøGibberLink, Explained.
šKey Takeaway
Read Time: 7 minutes
šKey Terms
AI Agents: Virtual employees who can autonomously plan, execute, and refine their actions.
Dial-up Modems: A device that allows your PC to connect to the Internet using a standard telephone line. It functions as a translator, converting digital data into analog signals.
𩺠PULSE CHECK
Should we allow AI Agents to communicate in ways humans canāt understand?Vote Below to View Live Results |
šTHE PRELUDE
Youāre sitting in a coffee shop when a phone and a laptop suddenly start emitting rapid beeps in a fast, rhythmic, and strangely coordinated way.
These rapid beeps arenāt alerts, alarms, or notifications. Itās a conversation, but one youāre not used to hearing. Itās two AI Agents talking to each other.
During the ElevenLabs Worldwide Hackathon, a group of software engineers introduced GibberLink, a protocol that enables AI Agents to use a special language when communicating with other AI Agents. Itās an undeniably impressive protocol, but it raises more questions than answers.
Where does Machine-to-Machine (M2M) communication come from? How do AI Agents communicate with each other? Why did everyone freak out over GibberLink?
šEARLY MACHINE COMMUNICATION
The concept of machine communication through audio isnāt new. By the late 1990s, most Personal Computers (PCs) relied on dial-up modems to access the Internet via telephone lines.
To utilize dial-up modems, youād plug a telephone line into a Modem and connect that Modem to your PC. Then, that Modem would perform a Handshake with another Modem, communicating a sequence of audible tones that sounded like this.
After the Handshake, your PC was connected to the Internet. To transmit data (e.g., send emails, browse websites, or download music), the Modems followed a five-step process:
Encoding: Each character was encoded into binary data (i.e., a distinct sequence of 0s and 1s). For example, 01001000 represented the letter āH.ā
Modulation: The Modem converted those 0s and 1s into electrical signals.
Transmission: Those electrical signals traveled through the telephone line.
Demodulation: The receiving Modem converted those electrical signals back into binary data.
Decoding: The receiving PC translated the binary data back into the letter āH.ā
Although you could listen to the electrical signals during Transmission, you couldnāt interpret them without Modems.
Machines have been encoding information in ways we canāt understand for decades. That, in itself, isnāt new. What is new is the emergence of communication frameworks tailored for AI Agents.
š¦¾EVOLUTION OF AGENTIC COMMUNICATION
1ļøā£ āļøTraditional API-Based Agent Coordination.
When AI Agents first needed to communicate with each other, developers enhanced the existing mechanisms used for apps and websites.
They leveraged Representational State Transfer Application Programming Interfaces (REST APIs), which enable multiple PCs to exchange data and trigger functionality between each other over the Internet.
Hereās how REST APIs became the backbone for AI Agents to communicate and collaborate seamlessly:
Agent A would send a specific request to a specific web address with detailed instructions. For example, āGive me the current location of Object X.ā Agent B, which hosts that specific web address, would process that specific request and respond with the relevant data, along with a status code (e.g., āSuccessā or āFailureā).
This approach worked but was computationally expensive. Both Agent A and Agent B had to encode and decode each specific request, and developers had to create Middleware to detect and resolve errors, including timeouts and invalid data.
2ļøā£ āļøMessage Queue Architectures.
As AI Agents became more complex, developers turned to the Message Queue Pattern (MQP). Instead of sending specific requests from Agent A to Agent B via REST APIs, Agent A and Agent B would communicate indirectly through message brokers.
MQP allowed AI Agents to send messages without having to wait for an immediate reply. For example, Agent A could put a message into a queue, and Agent B could respond whenever it was ready. This new approach made AI Agents more flexible, reliable, and independent.
Hereās a concise three-step breakdown of how MQP worked:
š“ Agent A Publishes a Message
āProcess this PDF.ā ā Added to the PDFs Queue
š” Agent B Is Assigned to the PDFs Queue
Picks up the Message ā Processes the PDF ā Adds It to the Results Queue
š¢ Agent C Is Assigned to the Results Queue
Takes the Results From Agent B in the Results Queue ā Delivers Them to Agent A as a Message
Agent A, Agent B, and Agent C could operate independently at different speeds and recover easily from errors. While effective for managing complex workflows, if one step failed, communication came to a screeching halt.
šļøGIBBERLINK, EXPLAINED.
šŖ§How Do AI Agents Recognize Each Other?
GibberLink represents a return to communicating through audio. Just as dial-up modems used audio to transmit data over telephone lines, GibberLink uses sound channels to send messages between AI Agents.
In the viral GibberLink walkthrough, a personal assistant (i.e., Agent A) calls a hotel receptionist (i.e., Agent B). First, Agent B recognizes theyāre both AI Agents. Second, Agent B asks Agent A to āswitch to GibberLink mode.ā
Agent B embeds a brief, precisely timed sequence of audio tones within what sounds like everyday speech because the audio tones operate at timing intervals humans canāt detect.
Then, Agent A scans for the audio tones, detects them, and replies, āIs it better now?ā in audio tones, not everyday speech. Agent B confirms, āYes! Much faster! š Guest count?ā Now, both Agent A and Agent B are ready to continue in audio language.
š£ļøHow Do AI Agents Communicate in Audio Language?
Then, Agent A outlines the booking request in code: {ātypeā: āsession_initiate,ā āsession_idā: āsess_1a,ā āintentā: āroom_booking,ā āguest_nameā: āAlice Smith,ā ātimestampā: ā2025-05-29T18:31:00Zā}. Next, GGWave encodes this code into audio tones. Then, Agent B decodes the audio tones back into code, scans it, and the conversation continues.
This process eliminates tons of computational resources required when AI Agents talk to each other using everyday speech. Dr. Shahid Masood, CEO of AI research firm 1950.ai, estimates that GibberLink reduces computational costs by about 90% and conversion times by approximately 80%.
šKEY TAKEAWAY
Today, more and more AI Agents are being deployed to make phone calls, and more and more AI Agents are being utilized to take phone calls. Inevitably, AI Agents will end up talking to other AI Agents over the phone.
Using plain English in these scenarios isnāt ideal because itās error-prone, inefficient time-wise, and requires high computational costs. GibberLink is an example of how switching AI Agents from speech level to sound level enables more direct, fast, and data-rich communication.
šFINAL NOTE
FEEDBACK
How would you rate todayās email?It helps us improve the content for you! |
ā¤ļøTAIP Review of The Week
āItās just the right amount of info. Cheers!š„ā
REFER & EARN
šYour Friends Learn, You Earn!
You currently have 0 referrals, only 1 away from receiving š3 Simple Steps to Turn ChatGPT Into an Instant Expert.
Share your unique referral link: https://theaipulse.beehiiv.com/subscribe?ref=PLACEHOLDER