Using AI to Understand Wolf Communication
50+ sites across 500k acres
50,000 hours of audio recordings
240 hours of wolves (3,147 events)
1,500 wolf observation hours
182 other species detected
Footage Courtesy of Bob Landis
Yellowstone National Park and Wild Livelihoods Business Coalition, in collaboration with Grizzly Systems and several academic institutions have partnered to deploy low-power, AI-infused monitoring devices that capture acoustic and visual data for behavioral research and for monitoring the presence and distribution of wolves across the Greater Yellowstone Ecosystem. Accurate population and occupancy estimates play a vital role in shaping state and federal management policies. Through the use of various artificial intelligence algorithms, scientists can efficiently analyze large data sets of audio and video/stills to find and then study wolf communication behavior.
Because wolf vocalizations carry for relatively long distances, AI-infused autonomous recording units (ARUs) that are integrated into camera traps can serve as low-cost tools to enhance existing census efforts. Furthermore, better approaches to livestock-conflict deterrence may be possible with playbacks of scientifically validated wolf or guardian dog vocalizations, triggered by the AI devices. Finally, audio and video educational tools can be created via an end-to-end software platform that responsibly showcases the biodiversity in some regions in order to encourage other regions to follow suite.
Passive Acoustic Monitoring (PAM) has emerged as a cost-effective and noninvasive technique for wolf surveys, providing detection probabilities exceeding those attained through camera trapping. We are building ARU's with classifiers for real-time detection, as well as ML models for post-processing analysis of the behavioral functions of wolf vocalizations. While bioacoustic monitoring is not a novel concept, the advent of advanced AI algorithms has opened up new possibilities to reduce costs and enhance researcher productivity in telemetry monitoring (for more information see Using machine learning to decode animal communication). The Greater Yellowstone region holds realistic, lower-cost potential for bioacoustic research, due to the long-term knowledge already gained from radio collaring, flight surveys, camera traps, and field surveys. As such, this collaborative research project aims to collect 24x7x365 bioacoustics data at pre-determined locations in the GYE which can be set aside, similar to genetic data, and used later for research of any species that vocalizes below 12khz.
Questions Being Asked
Some of the fundamental questions driving the research objectives include:
-
When and how often do wolves vocalize?
-
Can we identify individuals or packs via idiolects and dialects?
-
Are there different functions for different types of wolf howls or chorus howls?
-
Can we count the number of wolves in a chorus howl?
-
Can low-cost and low-touch acoustic recorders inform population estimates?
-
Can acoustic recorders be practically used for mitigation of livestock conflict?
Objectives
The aim of this collaborative research project is to explore and evaluate bioacoustics parameters of wolf vocalizations that will:
-
build a 24x7x365 bioacoustics and observations dataset in the GYE for any species or soundscape below 12khz
-
test systems that automate real-time and non real-time collection and classification (by species and individuals) of bioacoustics and imagery data in the cloud (see t.ly/2dQ0q and t.ly/o0_xO)
-
model wolf occupancy, distribution, and abundance from acoustic data
-
evaluate behavioral and ecological questions about the purpose and flexibility of communication in wolves ("come here", "where are you", "this is me/us") - (see t.ly/JeZBD)
-
create GAN AI models to identify sound elements, patterns, and groupings in wolf howls that will facilitate identity of ecological and behavioral correlates and thereby the sounds’ potential communicative significance (see t.ly/o9ke1)
-
provide opportunities for education and outreach on this aspect of animal communication and its applications for conservation and stewardship
-
test non-lethal use cases for livestock-conflict scenarios
Management & Financial Benefits
-
Reduce labor costs associated with manually gathering population data.
-
Reduce redundancy of acoustic data collection across species
-
Reduce the costs and safety issues associated with the use of flying craft to gather population information.
-
Aid in the growing list of tools for predator-livestock conflict mitigation.
Conservation Benefits
-
Develop monitoring methods that are less invasive, risky, and logistically challenging compared to radio-collaring wolves with helicopters or traps.
-
Monitor for habitat degradation or restoration based on biodiversity bioacoustic indexes.
-
Provide fundamental research towards a better understanding of animal communication functions via large-scale, long-term bioacoustic data, time-synced with other sensor and observer data.
-
Increase the accuracy of wolf population, movement and presence estimates which are used to inform state and federal conservation practices. Knowledge of individual or group identity can be used to help calculate precise population estimates (which are essential to provide robust estimates of population sustainability), to collect data about survival (which can be used to estimate trends and identify causes of mortality), and to quantify individual movement (which represents ranging and resource use).
-
Educational curriculum and exhibit content
Collaboration Partners
What Have We Learned So Far
-
wolves predominantly vocalize during night time hours
-
wolves increase daytime vocalizations during the winter breeding season
-
wolves rapidly modulate their howls during "stressful" situations (e.g. when interacting with a rival pack)
-
wolves respond to coyote vocalizations, but do not silence the coyotes
-
wolf individuals can be identified by the pitch of their howl
-
wolves almost always initial a chorus howl with one or more individuals howling, and often the chorus howl (when multiple wolves join in) is initiated by a higher pitched howl
-
during breeding season, individual wolves will howl in a way that the pitch rises, plateaus and then falls
Gabe, a highschool intern annotating a wolf chorus howl for our machine learning algorithm
A Little about the Technology
Supervised Wolf Bioacoustic Detection
There is extensive precedent for applying ML for supervised bioacoustic detection tasks; examples include a sperm whale click detector, a humpback detector, and a model that detects and classifies birdsong, among many others. Employing similar methods, we can train a convolutional neural network (CNN) either from scratch or using pretrained weights to classify an acoustic window as non-signal or wolf signal depending on the absence or presence of a wolf vocalization in the given acoustic segment.
Further, we take advantage of collaborative work done by other teams, such as PNW_Cnet, Synature, and BirdNet (D Sassover) regarding AI-based wolf detection. We encourage academic institutions to combine efforts with our public and private institutions to iterate more quickly on the best general classifier, focusing on a common pipeline and growing the dataset (and its relevant ambient correlates) across canids and eventually all large carnivore species.
Supervised Wolf Chorus Counting
To our knowledge, there are no attempts at automated acoustic counting of overlapping signals, though there are several approaches that may be promising.
-
Train a model (e.g. LSTM-CRF) to predict the number of overlapping spectral elements at fine timescales using open-source data. Assess the model’s ability to generalize to new datasets. Train a model to predict the number of wolves in a chorus based on human annotation of the number of wolves vocalizing concurrently.
-
Train a model to align video with acoustic data, as in examples of human music instrument playing.
Unsupervised Wolf Source Separation
Using previous work in source separation and emphasizing the unsupervised MixIT training algorithm used to separate overlapping birdsong mixtures, we can attempt to separate wolf choruses into predictions for the individuals present in the chorus. Though not functionally limited in the number of sources it can handle, it is unclear how the model will perform as the number of concurrently vocalizing wolves increases.
Unsupervised Meaning Discovery in Wolf Vocalizations
The CETI project has produced machine learning models, with little or no understanding of a species vocal repertoire, can be used to reveal meaningful units in the sounds. The approach in this paper, APPROACHING AN UNKNOWN COMMUNICATION SYSTEM BY LATENT SPACE EXPLORATION AND CAUSAL INFERENCE, with modification for wolf vocalizations, is promising.
Our Data Set
-
Wolf Recordings (video optional) annotated with start and stop times of wolf howl events.
- Chorus Howls: 19.6 hours of uninterrupted recordings spanning 20 years
- Individual Howls: 5.3 hours of uninterrupted recordings spanning 20 years
-
Ambient and Similar-to-Wolf Data:
-
The classifier was also trained on 10 hours of ambient recordings from 5 locations in the GYE, as well as on elk vocalizations, coyotes, planes, vehicles to enable optimal model performance in classifying wolf vs. non-wolf/background sounds. Airplanes are one of the top false positives.
-
Our Github Repositories (email us for access)
Related Scientific Research
-
Acoustic Identification of Wild Gray Wolves, Canis lupus, Using Low Quality Recordings
-
Citizen science contribution to national wolf population monitoring: what have we learned?
-
Tracking cryptic animals using acoustic multilateration: A system for long-range wolf detection
-
Testing a New Passive Acoustic Recording Unit to Monitor Wolves
-
Bioacoustic Detection of Wolves: Identifying Subspecies and Individuals by Howls
-
Singing in a wolf chorus: structure and complexity of a multicomponent acoustic behaviour
-
The potential for acoustic individual identification in mammals
-
Tracking cryptic animals using acoustic multilateration: A system for long-range wolf detection
-
The contribution of source filter theory to mammal vocal communication research
-
Cross Modal Perception of Body Size in Domestic Dogs (Canis familiaris)
-
Size communication in domestic dog, Canis familiaris, growls
-
Wolf Howling Is Mediated by Relationship Quality Rather Than Underlying Emotional Stress
-
Not afraid of the big bad wolf: calls from large predators do not silence mesopredators
-
Acoustic analysis of wolf howls recorded in Apennine areas with different vegetation covers
-
Visualizing sound: counting wolves by using a spectral view of the chorus howling
-
Wolf howls encode both sender- and context-specific information
-
Individually distinct vocalizations in timber wolves, Canis lupus
-
Automated identification of avian vocalizations with deep convolutional neural networks
-
Chorus Howling by Wolves: Acoustic Structure, Pack Size and the Beau Geste Effect
-
Timber wolf howling playback studies: Discrimination of pup from adult howls
-
Radiographic analysis of canine vocal tract anatomy and its implications for human language origins
-
Voice-Sensitive Regions in the Dog and Human Brain Are Revealed by Comparative fMRI
-
Does size matter? Examining the drivers of mammalian vocalizations
-
Long-duration, false-colour spectrograms for detecting species in large audio data-sets
-
Recognition of familiarity on the basis of howls: a playback experiment in a captive group of wolves
Some Types of Wolf Vocalizations
Situations that can evoke wolf howls