Welcome back, AI prodigies!

In today’s Sunday Special:

  • 📜 The Prelude

  • 🐣 A Brief History of Birding

  • đŸ„ AI Instantly Identifies Bird Songs

  • đŸ•ïž The Rise of Environmental Intelligence

  • 🔑 Key Takeaway

Read Time: 7 minutes

🎓Key Terms

  • Bioacoustics: The science of how animals produce sounds to track species and assess habitat health without disturbing wildlife.

  • Convolutional Neural Networks (CNNs): A network of specialized filters that detect complex visual patterns like edges and corners across images and videos.

đŸ©ș PULSE CHECK

Would you use a mobile app to help identify birds?

Vote Below To View Live Results

Login or Subscribe to participate

📜 THE PRELUDE

The glowing sun peaks over the horizon, setting treetops ablaze with gold and dew-drenched spiderwebs radiant with sparkle as the forest wakes from the night’s slumber.

A brisk breeze pierces your cheeks as you hike along the twisting trail. Beneath your boots, each step lands with a soft crunch
crunch
crunch as dry leaves crumble.

Suddenly, a zzzzrrrRREEEEKK tears through the canopy. You tense up your ears, trying to place it, but the forest doesn’t isolate sound for you: twigs snap, insects hum, and frogs croak.

A single minute of forest audio can contain thousands of overlapping sounds. Even expert birders, who possess highly trained identification skills to recognize bird species by sight and sound, can have varying interpretations of each bird’s vast soundscape.

That’s why “BirdNET” exists: it’s like Shazam, but for bird species. Instead of telling you what song is playing, it tells you which bird is singing. So, how exactly does AI identify bird calls?

🐣 A BRIEF HISTORY OF BIRDING

⊿ 1ïžâƒŁ Memory, Not Measurement?

Before 1889, bird identification was an exercise in trained memory, not calculated measurement. In other words, ornithologists learned bird sounds the way musicians learned melodies, by repetition and association. For example, bird calls were encoded into onomatopoeic phrases like: “Drink-your-teeeeeeea!” for the Eastern Towhee’s cheerful trilling call.

Early ornithological manuals relied on qualitative descriptors such as: “chirping,” “whistling,” “rising,” “falling,” and “buzzy.” There wasn’t any way to compare recordings across regions or develop shareable audio archives. If multiple ornithologists didn’t collectively agree on what they heard, there wasn’t an objective reference to resolve it.

⊿ 2ïžâƒŁ Recordings Make Sounds Shareable?

In 1889, British sound recordist Ludwig Koch leveraged a “phonograph cylinder” to record the melody of an Indian songbird called the White-Rumped Shama. The Indian songbird’s melody entered a large horn, vibrating a thin metal diaphragm attached to a small stylus needle. This small stylus needle traced grooves around a wax cylinder, with louder sounds producing deeper grooves and softer sounds producing shallower grooves.

To replay the recording, the grooved wax cylinder was placed back onto the “phonograph cylinder” and spun at a steady speed. As it spun, the small stylus needle followed the original grooves, reproducing the original vibrations etched during the original recording.

In 1929, the first American professor of birdwatching, Arthur A. Allen, established a systematic method for recording bird sounds to ensure all recordings were made at consistent speeds, with standardized equipment settings and precise field documentation protocols, so that avian acoustic data could be accurately studied. This effort led to the development of the first scientific archive of bird sounds, which eventually became housed at the CornellLab of Ornithology.

In 1932, Peter Paul Kellogg, a professor of biological acoustics at Cornell University, adopted new microphones called “parabolic reflectors,” which relied on curved dishes to amplify bird sounds from specific directions while suppressing background noise.

⊿ 3ïžâƒŁ Seeing Sounds at Scale?

In the 1940s, Bell Labs invented the “sonograph,” which converted audio into a spectrogram: a visual representation of the spectrum of frequencies of a sound signal over time. It’s essentially a heat map that plots time on the x-axis, frequency on the y-axis, and amplitude as colors. This technical breakthrough enabled ornithologists to extract Features, including the pitches, pauses, and phases between notes in a bird’s song, which are essential to studying mating rituals and territorial signals.

đŸ„ AI INSTANTLY IDENTIFIES BIRD SONGS

⊿ 4ïžâƒŁ How AI Analyzes Raw Bird Audio?

The K. Lisa Yang Center for Conservation Bioacoustics at CornellLab developed “BirdNET,” which recognizes bird vocalizations from audio recordings.

  1. 🔮 Data Capture:

    • Raw bird audio is captured at 48 kHz and divided into 3-second sound segments, optimizing the balance between “BirdNET’s” input size and the natural duration of bird vocalizations.

    • These 3-second sound segments are processed into 2 log-scaled Mel-spectrograms, which visualize how the different frequencies in the raw bird audio change over time, adjusted to complement human hearing since humans don’t perceive frequency linearly. This adjustment prevents “BirdNET” from wasting time on high-frequency details that humans can’t distinguish.

  2. 🟠 Data Patterns:

    • The 2 log-scaled Mel-spectrograms are passed through a CNN, which scans the visualized frequencies through hundreds of specialized filters or “layers.” The earlier “layers” identify simple visual patterns like:

      1. 📍 Sharp Edges: Quick chirps create diagonal ridges.

      2. 📌 Vertical Lines: Trilled calls create upright striations.

  3. 🟡 Data Discovery:

    • The CNN outputs an initial probability on a scale from 0 to 1, with 1 indicating complete confidence that the raw bird audio matches a specific bird species. For example, it might assign an initial probability like American Robin = 0.82.

    • The initial probability is cross-referenced with local metadata, including the date and location of when and where the raw bird audio was recorded, to produce high-confidence bird species predictions.

đŸ•ïž THE RISE OF ENVIRONMENTAL INTELLIGENCE

⊿ 5ïžâƒŁ Any Real-World Impact?

It’s actively reshaping wildfire management. In California’s Sierra Nevada region, bird experts deployed 1,600 ARUs, or weatherproof microphones, across 6 million acres of forestland. Then, they utilized “BirdNET” to analyze over 700,000 hours of forest audio to track how bird populations respond to wildfires.

Out of 42 bird species, 28 had higher population density after low to moderate-severity wildfires. Surprisingly, these types of wildfires helped reduce canopy density by up to 50%, giving bird species more space to nest and forage. Additionally, the dead wood attracted more “insect larvae,” while the regrowing vegetation produced more seeds and berries.

🔑 KEY TAKEAWAY

Until 2018, identifying bird calls from forest recordings was a tedious and technical undertaking. Now, “BirdNET” powers mobile apps like “Merlin Bird ID,” which allows anyone to identify thousands of bird species from sight or sound for free. This birding breakthrough underscores how advances in AI are making premium-level technical tools more accessible for everyone around the world.

📒 FINAL NOTE

FEEDBACK

How would you rate today’s email?

It helps us improve the content for you!

Login or Subscribe to participate

❀TAIP Review of The Week

❝

“useful, interesting, positive.”

-Don (1ïžâƒŁ 👍Nailed it!)
REFER & EARN

🎉Your Friends Learn, You Earn!

{{rp_personalized_text}}

Share your unique referral link: {{rp_refer_url}}