A team of scientists from MIT, Harvard, and UC Berkeley says they have cracked the basic structure of sperm whale communication — and they plan to start talking back this year.

Project CETI (Cetacean Translation Initiative), a nonprofit research consortium, has used machine learning to identify what it calls a "phonetic alphabet" of 156 distinct clicking patterns, along with vowel-like spectral shifts that suggest whales encode far more meaning in their clicks than scientists ever suspected.

Key Facts
  • 156 codas identified as sperm whale "phonemes" — rhythmic click patterns that form the building blocks of communication
  • Vowel-like sounds discovered in November 2025, published in Open Mind
  • Two-way communication attempt planned for 2026 using AI-generated synthetic codas
  • $33 million in funding from the TED Audacious Project
  • 8,725 audio snippets analyzed in the landmark 2024 study

From Morse Code to Grammar

For decades, marine biologists treated sperm whale clicks like Morse code — simple, binary signals for echolocation and basic identification. Project CETI has upended that assumption entirely.

Working off the coast of Dominica in the Eastern Caribbean since 2020, the team deployed underwater listening stations, non-invasive drone tags developed at Harvard, and Google's DolphinGemma — a 400-million parameter AI model that runs on smartphones in the field for real-time acoustic analysis.

The result: clicks aren't just rhythmic patterns. They carry rhythm, tempo, rubato, and ornamentation — musical qualities that combine into what the team calls codas. And within those codas, whales modulate the frequency of individual clicks to produce something eerily close to vowels.

"What used to be conceived of as this alien-looking Morse-code like system just became much more human-like. It's a case of underwater vowels." — Dr. Gašper Beguš, Linguistics Lead, UC Berkeley

The Vowel Discovery That Changed Everything

In November 2025, a UC Berkeley-led team published findings that rocked the field. By applying unsupervised machine learning — the same approach used to translate between human languages without a shared dictionary — Dr. Beguš identified two distinct vowel patterns in sperm whale clicks:

  • "ɑ-vowels" — open, resonant frequency shifts similar to the 'a' in "father"
  • "i-vowels" — higher-frequency shifts resembling the 'ee' in "meet"
  • Diphthong-like patterns — frequency trajectories that glide between two positions, much like human diphthongs

This means a single coda isn't just a rhythmic "word." The whale can modify how it clicks within that pattern to layer additional meaning — analogous to how humans can change the vowel in a consonant frame ("bat" vs. "bit" vs. "but") to create entirely different words.

ℹ️
The implications extend beyond biology. NYU's MOTH (More Than Human Life) Project is already examining whether proof of linguistic complexity in whales could trigger legal challenges to their classification — shifting the framework from animal "welfare" to something closer to "personhood."

The Skeptics Push Back

Not everyone is convinced. The discovery has split the marine biology community.

Dr. Luke Rendell of the University of St. Andrews called the vowel comparison "completely nonsense," arguing the spectral shifts could be recording artifacts from varying distances between the whale and the hydrophone.

Dr. Stephanie King of the University of Bristol offered a more measured critique: the patterns might reflect the whale's arousal or alertness state rather than intentional linguistic encoding. A whale that's excited might simply click differently than one that's calm — not because it's choosing different "vowels," but because its physiology changes.

The CETI team acknowledges these concerns but points to their dataset's scale and controls. Their click detection system runs at 99.5% accuracy, and dialect recognition hits 95.3% — high enough, they argue, to distinguish genuine patterns from noise.

2017
David Gruber and Shafi Goldwasser first discuss applying cryptography to whale clicks at the Radcliffe Institute
March 2020
Project CETI officially founded with $33M from TED Audacious Project
July 2023
Team films the first-ever recorded sperm whale birth, providing crucial behavioral context
May 2024
Phonetic alphabet of 156 codas published in *Nature Communications*
October 2024
Related Whale-SETI team conducts 20-minute "conversation" with humpback whale Twain
November 2025
Vowel discovery published in *Open Mind*
December 2025
WhAM (Whale Acoustics Model) presented at NeurIPS
2026
First two-way communication attempt planned using synthetic codas

The Team Behind the Breakthrough

Project CETI isn't a single lab — it's a consortium that reads like a who's who of AI and marine science:

Role Researcher Affiliation
Founder & Lead Dr. David Gruber City University of New York
Biology Lead Dr. Shane Gero Carleton University
Linguistics Lead Dr. Gašper Beguš UC Berkeley
Robotics Lead Dr. Robert Wood Harvard SEAS
Cryptography Lead Shafi Goldwasser UC Berkeley (Turing Award winner)
WhAM Creator Orr Paradise UC Berkeley

Google contributed DolphinGemma for field analysis. MIT's CSAIL built the machine learning pipeline. And a Turing Award-winning cryptographer is treating whale codas like an encrypted language waiting to be broken.

What Talking to Whales Actually Looks Like

The 2026 goal isn't a conversation about philosophy. It's far more modest — and scientifically rigorous.

The team plans to use the WhAM model to generate synthetic codas that mimic known patterns — location markers, identity calls, group coordination signals. These would be played through underwater speakers near a studied whale clan off Dominica, and researchers would observe whether the whales respond appropriately.

Click Detection
99.5
Dialect Recognition
95.3
Coda Classification
91
Individual ID
87.2

Success wouldn't mean "translating" whale language in any human sense. It would mean proving that the communication system is interactive — that whales don't just broadcast, but respond to specific inputs with predictable outputs. That alone would be unprecedented for any non-primate species.

The 4-Billion-Click Problem

The team estimates it needs 4 billion vocalizations to achieve anything approaching full translation — orders of magnitude beyond the current dataset. New permanent listening stations are being deployed around Dominica to capture this volume autonomously.

Meanwhile, the legal and ethical questions are already outpacing the science. The MOTH project has drafted 12 ethical principles for interspecies communication, specifically designed to prevent military or commercial exploitation of any future translation capability.

If we can prove whales have grammar, dialects, and names — as signature codas suggest — the legal definition of "personhood" may need to expand beyond our species for the first time.

Whether Project CETI's critics are right about artifacts, or the team is right about underwater vowels, one thing is clear: AI has forced the question of animal language out of philosophy and into the lab. The whales have been clicking for 15 million years. We're just now learning to listen.


Project CETI's research is ongoing in Dominica. The team's full dataset and DolphinGemma model are expected to be open-sourced later in 2026.