At this point, I am well aware of the ubiquity of Artificial Intelligence and Machine Learning in our lifes today, and that living without them would set us back decades. One of the great things it allows us to access is a huge diversity of languages we do not necessarily speak, thanks to a variety of online dictionaries and translators that can convert our spoken or written language into another one, even if these translations sometimes turn out to be rather… interesting.
However, one thing I had never heard about before I recently stumbled across an article about it was the use of machine learning for the translation of animal language. Honestly, to me, this sounded more like the plot of a Sci-Fi movie, but there are indeed several research projects running that are attempting to do that exact thing. And while the achievement of their final goal still seems to lie in the more distant future, they have had some success that I would like to tell you about. As stated before, there are several groups actively working on the idea of translating animal language, but the most ambitious one is the nonprofit Earth Species Project (ESP). This California-based group aims to decode communication of all non-human species and subsequently make these discoveries available to the public to help us understand them, in the hopes of deepening the bond between humans and nature and thus helping protect them. According to them, a lot of our maltreatment of animals results from issues in communication, and an attempt at translation might deepen our respect for them and relieve us of our human preconceptions. The biggest obstacle in this project is the mere fact that, as opposed to other researchers, they wish to include every species, especially because their ways of non-verbal communication differ immensely from each other.
This huge project is supposed to work thanks to the development of an algorithm that would geometrically represent the words in a physical space, in which the direction and distance between different words stand for their semantic relationship to each other. The example given here is the following: king has to man the same distance and direction that woman has to queen. The mapping of these points is made merely by observing how often they appear near one another, without necessarily knowing their meaning. The shapes emerging from this process are comparable for different languages and would thus allow a translation through their alignment.


So, how likely is it that they will succeed? To be fair, my understanding of this topic is very limited, but so far, they have reached two milestones that pushed them a few steps toward their target.
Imagine being in a really busy bar, talking to a friend. The music is quite loud, and everyone else in the room is engaged in their own conversation, raising their voices to fight against the background noise. To be able to keep up your conversation you will need to pick out that specific sound that is of interest to you, in this case, your friend’s voice, and ignore all the other sounds.

Press play to be transported to a noisy cocktail party. At first, you can only hear a melange of sounds. Then, you run into an old friend, who starts talking to you. As you focus on what your friend is saying, you are eventually able to filter out the other sounds of the party, effectively turning them into background noise.
This phenomenon is called the ‘cocktail party problem’, and ESP managed to solve it in their research. They created a code that can make out which animal out of a big group of noisy animals is vocalizing. Another thing they have figured out is how to ‘talk’ to animals directly. The group came up with an algorithm that enables the AI to generate mimicked calls, the meaning of which are still a mystery to us.
In a project that is currently still in progress, ESP is working on the development of an algorithm that can determine the number of call types one species uses. In order to do so, they are applying self-monitored ML to learn patterns. Another project is concerned with the operative meanings of vocalizations and their automatic understanding. For this, they are studying the behavior of wild sea mammals underwater with the use of small biologging devices in one of the largest tagging programs in the world in which they record the animals’ location, motion, and visual field. The aim here is to automatically measure their activities and add audio data to understand whether the sounds tied to their behavior have a specific functional meaning.
I am always a little critical when it comes to projects like these, and of course, I am not the only one. While the main reason for the group to carry out this lengthy research is that of the creation of kinship with nature and damage control, it is always extremely important to consider the way that these advancements will be used. Critics of projects like ESP generally highlight the risk of the abuse of our newfound powers to dominate the natural world of animals and plants, exploiting them, and carrying on the military history of using animals.
What do you think of this project?
ESP is certainly not the only group actively working on decoding animal language, but their project is the most extensive one. Other groups that might be of interest to you, if you wish to further read up on this topic would be the Wild Dolphin Project, the Interspecies internet, DeepSqueak, and Project CETI (Cetacean Translation Initiative).
References
Neel Dhanesha, “How Tech is helping us talk to Animals,” Vox.com, 30 October 2022.
City Fairclough, “Have you heard about the Cocktail Party Problem?,” COMSOL Blog, 31 July 2021.
Emily Anthes, “The Animal Translators,” The New York Times, 30 August 2022.
I really really liked your blog. Talking to animals sounds more like a thing that you only see in movies or a Disney princess is doing it.
I do think that one of the hardests part of this concept is the fact that animals have many different species so also many different languages/dialects. This makes it very difficult to reach that goal.
This is the same for computers talking like a real human. The cannot understand human language that perfect (yet?), because of sound changes and dialect.
And what would you think what would we do with this technology when we are able to talk or understand animals? I find this something really nice to think about.
Oh wow this is an interesting development! Reminds me of the dogs on tiktok (and now cats too) that are trained to communicate through buttons that have pre-recorded words on them.
I surely agree that it is important to reconnect with animals and I hope that this will lower animal abuse. Although I do wonder whether this need to understand animal language is perhaps a little anthropocentric and I wonder if it could eventually be used against animals. For example the cows wearing VR headsets that show images of sunny fields in order to make them produce more milk.
This was a super interesting post. The concept of talking to animals like you said seems more like fiction and possible fact to me but this is looks like a very serious approach.
The motives behind it makes complete sense to me and just imagining the possibility of speaking with my cat makes me think that we could only become closer. However, the ambitious nature of the project makes me think that still take a really really long time, and perhaps time that we don’t have. Forgive me for taking a pessimistic outlook but by the time that we mange to communicate with animals, there may be very few left.
Interesting development. Didn’t know this type of research existed. I am also a bit critical about how this technology might be used. It might be used to scare animals. This is already done with orchards that use audio systems as a kind of scarecrow (often of predatory birds).