1
|
Schuller BW, Akman A, Chang Y, Coppock H, Gebhard A, Kathan A, Rituerto-González E, Triantafyllopoulos A, Pokorny FB. Ecology & computer audition: Applications of audio technology to monitor organisms and environment. Heliyon 2024; 10:e23142. [PMID: 38163154 PMCID: PMC10755287 DOI: 10.1016/j.heliyon.2023.e23142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2023] [Revised: 11/08/2023] [Accepted: 11/27/2023] [Indexed: 01/03/2024] Open
Abstract
Among the 17 Sustainable Development Goals (SDGs) proposed within the 2030 Agenda and adopted by all the United Nations member states, the 13th SDG is a call for action to combat climate change. Moreover, SDGs 14 and 15 claim the protection and conservation of life below water and life on land, respectively. In this work, we provide a literature-founded overview of application areas, in which computer audition - a powerful but in this context so far hardly considered technology, combining audio signal processing and machine intelligence - is employed to monitor our ecosystem with the potential to identify ecologically critical processes or states. We distinguish between applications related to organisms, such as species richness analysis and plant health monitoring, and applications related to the environment, such as melting ice monitoring or wildfire detection. This work positions computer audition in relation to alternative approaches by discussing methodological strengths and limitations, as well as ethical aspects. We conclude with an urgent call to action to the research community for a greater involvement of audio intelligence methodology in future ecosystem monitoring approaches.
Collapse
Affiliation(s)
- Björn W. Schuller
- GLAM – Group on Language, Audio, & Music, Imperial College London, UK
- EIHW – Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
- audEERING GmbH, Gilching, Germany
| | - Alican Akman
- GLAM – Group on Language, Audio, & Music, Imperial College London, UK
| | - Yi Chang
- GLAM – Group on Language, Audio, & Music, Imperial College London, UK
| | - Harry Coppock
- GLAM – Group on Language, Audio, & Music, Imperial College London, UK
| | - Alexander Gebhard
- EIHW – Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
| | - Alexander Kathan
- EIHW – Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
| | - Esther Rituerto-González
- EIHW – Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
- GPM – Group of Multimedia Processing, University Carlos III of Madrid, Spain
| | | | - Florian B. Pokorny
- EIHW – Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
- Division of Phoniatrics, Medical University of Graz, Austria
| |
Collapse
|
2
|
Winiarska D, Szymański P, Osiejuk TS. Detection ranges of forest bird vocalisations: guidelines for passive acoustic monitoring. Sci Rep 2024; 14:894. [PMID: 38195687 PMCID: PMC10776575 DOI: 10.1038/s41598-024-51297-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2023] [Accepted: 01/03/2024] [Indexed: 01/11/2024] Open
Abstract
Passive acoustic monitoring has proven to have many advantages for monitoring efforts and research activities. However, there are considerations to be taken into account regarding the placement of autonomous sound recorders. Detection ranges differ among species and in response to variable conditions such as weather or the location of vocalising animals. It is thus important to the success of a research project to understand, with a certain degree of confidence, the distances at which birds might be detected. In two types of forests in Poland, we played back the vocalisations of 31 species of European forest birds exemplifying different singing characteristics. Based on recordings obtained along a 500-m transect, we estimated the probability of detection and maximum detection distance of each vocalisation. We broadcasted the recording at three heights of singing and repeated playbacks three times during the breeding season to evaluate the effect of vegetation growth. Our results revealed that environmental and meteorological factors had a significant influence on both detection probability and maximum detection distances. This work provides comprehensive measurements of detection distance for 31 bird species and can be used to plan passive acoustic monitoring research in Europe, taking into account species traits and individual characteristics of the study area.
Collapse
Affiliation(s)
- Dominika Winiarska
- Department of Behavioural Ecology, Institute of Environmental Biology, Faculty of Biology, Adam Mickiewicz University, Poznań, Poland.
| | - Paweł Szymański
- Department of Behavioural Ecology, Institute of Environmental Biology, Faculty of Biology, Adam Mickiewicz University, Poznań, Poland
| | - Tomasz S Osiejuk
- Department of Behavioural Ecology, Institute of Environmental Biology, Faculty of Biology, Adam Mickiewicz University, Poznań, Poland
| |
Collapse
|
3
|
Alberti S, Stasolla G, Mazzola S, Casacci LP, Barbero F. Bioacoustic IoT Sensors as Next-Generation Tools for Monitoring: Counting Flying Insects through Buzz. INSECTS 2023; 14:924. [PMID: 38132598 PMCID: PMC10743731 DOI: 10.3390/insects14120924] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Revised: 11/30/2023] [Accepted: 12/02/2023] [Indexed: 12/23/2023]
Abstract
The global loss of biodiversity is an urgent concern requiring the implementation of effective monitoring. Flying insects, such as pollinators, are vital for ecosystems, and establishing their population dynamics has become essential in conservation biology. Traditional monitoring methods are labour-intensive and show time constraints. In this work, we explore the use of bioacoustic sensors for monitoring flying insects. Data collected at four Italian farms using traditional monitoring methods, such as hand netting and pan traps, and bioacoustic sensors were compared. The results showed a positive correlation between the average number of buzzes per hour and insect abundance measured by traditional methods, primarily by pan traps. Intraday and long-term analysis performed on buzzes revealed temperature-related patterns of insect activity. Passive acoustic monitoring proved to be effective in estimating flying insect abundance, while further development of the algorithm is required to correctly identify insect taxa. Overall, innovative technologies, such as bioacoustic sensors, do not replace the expertise and data quality provided by professionals, but they offer unprecedented opportunities to ease insect monitoring to support conservation biodiversity efforts.
Collapse
Affiliation(s)
- Simona Alberti
- Department of Life Sciences and Systems Biology, University of Turin, Via Accademia Albertina 13, 10123 Turin, Italy;
| | | | - Simone Mazzola
- 3Bee srl, Via Alessandro Volta 4, 20056 Trezzo Sull’Adda, Italy;
| | - Luca Pietro Casacci
- Department of Life Sciences and Systems Biology, University of Turin, Via Accademia Albertina 13, 10123 Turin, Italy;
| | - Francesca Barbero
- Department of Life Sciences and Systems Biology, University of Turin, Via Accademia Albertina 13, 10123 Turin, Italy;
| |
Collapse
|
4
|
Lim KM, Lee CP, Lee ZY, Alqahtani A. EnViTSA: Ensemble of Vision Transformer with SpecAugment for Acoustic Event Classification. SENSORS (BASEL, SWITZERLAND) 2023; 23:9084. [PMID: 38005472 PMCID: PMC10674441 DOI: 10.3390/s23229084] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Revised: 09/25/2023] [Accepted: 09/29/2023] [Indexed: 11/26/2023]
Abstract
Recent successes in deep learning have inspired researchers to apply deep neural networks to Acoustic Event Classification (AEC). While deep learning methods can train effective AEC models, they are susceptible to overfitting due to the models' high complexity. In this paper, we introduce EnViTSA, an innovative approach that tackles key challenges in AEC. EnViTSA combines an ensemble of Vision Transformers with SpecAugment, a novel data augmentation technique, to significantly enhance AEC performance. Raw acoustic signals are transformed into Log Mel-spectrograms using Short-Time Fourier Transform, resulting in a fixed-size spectrogram representation. To address data scarcity and overfitting issues, we employ SpecAugment to generate additional training samples through time masking and frequency masking. The core of EnViTSA resides in its ensemble of pre-trained Vision Transformers, harnessing the unique strengths of the Vision Transformer architecture. This ensemble approach not only reduces inductive biases but also effectively mitigates overfitting. In this study, we evaluate the EnViTSA method on three benchmark datasets: ESC-10, ESC-50, and UrbanSound8K. The experimental results underscore the efficacy of our approach, achieving impressive accuracy scores of 93.50%, 85.85%, and 83.20% on ESC-10, ESC-50, and UrbanSound8K, respectively. EnViTSA represents a substantial advancement in AEC, demonstrating the potential of Vision Transformers and SpecAugment in the acoustic domain.
Collapse
Affiliation(s)
- Kian Ming Lim
- Faculty of Information Science and Technology, Multimedia University, Melaka 75450, Malaysia;
| | - Chin Poo Lee
- Faculty of Information Science and Technology, Multimedia University, Melaka 75450, Malaysia;
| | - Zhi Yang Lee
- DZH International Sdn. Bhd., Kuala Lumpur 55100, Malaysia
| | - Ali Alqahtani
- Department of Computer Science, King Khalid University, Abha 61421, Saudi Arabia;
- Center for Artificial Intelligence (CAI), King Khalid University, Abha 61421, Saudi Arabia
| |
Collapse
|
5
|
Zhang S, Gao Y, Cai J, Yang H, Zhao Q, Pan F. A Novel Bird Sound Recognition Method Based on Multifeature Fusion and a Transformer Encoder. SENSORS (BASEL, SWITZERLAND) 2023; 23:8099. [PMID: 37836929 PMCID: PMC10575132 DOI: 10.3390/s23198099] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 09/16/2023] [Accepted: 09/18/2023] [Indexed: 10/15/2023]
Abstract
Birds play a vital role in the study of ecosystems and biodiversity. Accurate bird identification helps monitor biodiversity, understand the functions of ecosystems, and develop effective conservation strategies. However, previous bird sound recognition methods often relied on single features and overlooked the spatial information associated with these features, leading to low accuracy. Recognizing this gap, the present study proposed a bird sound recognition method that employs multiple convolutional neural-based networks and a transformer encoder to provide a reliable solution for identifying and classifying birds based on their unique sounds. We manually extracted various acoustic features as model inputs, and feature fusion was applied to obtain the final set of feature vectors. Feature fusion combines the deep features extracted by various networks, resulting in a more comprehensive feature set, thereby improving recognition accuracy. The multiple integrated acoustic features, such as mel frequency cepstral coefficients (MFCC), chroma features (Chroma) and Tonnetz features, were encoded by a transformer encoder. The transformer encoder effectively extracted the positional relationships between bird sound features, resulting in enhanced recognition accuracy. The experimental results demonstrated the exceptional performance of our method with an accuracy of 97.99%, a recall of 96.14%, an F1 score of 96.88% and a precision of 97.97% on the Birdsdata dataset. Furthermore, our method achieved an accuracy of 93.18%, a recall of 92.43%, an F1 score of 93.14% and a precision of 93.25% on the Cornell Bird Challenge 2020 (CBC) dataset.
Collapse
Affiliation(s)
- Shaokai Zhang
- College of Electronics and Information Engineering, Sichuan University, Chengdu 610041, China; (S.Z.); (Y.G.); (J.C.)
| | - Yuan Gao
- College of Electronics and Information Engineering, Sichuan University, Chengdu 610041, China; (S.Z.); (Y.G.); (J.C.)
| | - Jianmin Cai
- College of Electronics and Information Engineering, Sichuan University, Chengdu 610041, China; (S.Z.); (Y.G.); (J.C.)
| | - Hangxiao Yang
- College of Computer Science, Sichuan University, Chengdu 610041, China; (H.Y.); (Q.Z.)
| | - Qijun Zhao
- College of Computer Science, Sichuan University, Chengdu 610041, China; (H.Y.); (Q.Z.)
| | - Fan Pan
- College of Electronics and Information Engineering, Sichuan University, Chengdu 610041, China; (S.Z.); (Y.G.); (J.C.)
| |
Collapse
|
6
|
Diaz SDU, Gan JL, Tapang GA. Acoustic indices as proxies for bird species richness in an urban green space in Metro Manila. PLoS One 2023; 18:e0289001. [PMID: 37506131 PMCID: PMC10381043 DOI: 10.1371/journal.pone.0289001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2022] [Accepted: 07/10/2023] [Indexed: 07/30/2023] Open
Abstract
We assessed eight acoustic indices as proxies for bird species richness in the National Science Complex (NSC), University of the Philippines Diliman. The acoustic indices were the normalized Acoustic Complexity Index (nACI), Acoustic Diversity Index (ADI), inverse Acoustic Evenness Index (1-AEI), Bioacoustic Index (BI), Acoustic Entropy Index (H), Temporal Entropy Index (Ht), Spectral Entropy Index (Hf), and Acoustic Richness Index (AR). Low-cost, automated sound recorders using a Raspberry Pi were placed in three sites at the NSC to continuously collect 5-min sound samples from July 2020 to January 2022. We selected 840 5-min sound samples, equivalent to 70 hours, through stratified sampling and pre-processed them before conducting acoustic index analysis on the raw and pre-processed data. We measured Spearman's correlation between each acoustic index and bird species richness obtained from manual spectrogram scanning and listening to recordings. We compared the correlation coefficients between the raw and pre-processed.wav files to assess the robustness of the indices using Fisher's z-transformation. Additionally, we used GLMMs to determine how acoustic indices predict bird species richness based on season and time of day. The Spearman's rank correlation and GLMM analysis showed significant, weak negative correlations between the nACI, 1-AEI, Ht, and AR with bird species richness. The weak correlations suggest that the performance of acoustic indices are dependent on various factors, such as the local noise conditions, bird species composition, season, and time of day. Thus, ground-truthing of the acoustic indices should be done before applying them in studies. Among the eight indices, the nACI was the best-performing index, performing consistently across sites and independently of season and time of day. We highlight the importance of pre-processing sound data from urban settings and other noisy environments before acoustic index analysis, as this strengthens the correlation between index values and bird species richness.
Collapse
Affiliation(s)
- Skyla Dennise U Diaz
- Institute of Biology, College of Science, University of the Philippines Diliman, Quezon City, Philippines
| | - Jelaine L Gan
- Institute of Biology, College of Science, University of the Philippines Diliman, Quezon City, Philippines
| | - Giovanni A Tapang
- National Institute of Physics, College of Science, University of the Philippines Diliman, Quezon City, Philippines
| |
Collapse
|
7
|
Clink DJ, Kier I, Ahmad AH, Klinck H. A workflow for the automated detection and classification of female gibbon calls from long-term acoustic recordings. Front Ecol Evol 2023. [DOI: 10.3389/fevo.2023.1071640] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/11/2023] Open
Abstract
Passive acoustic monitoring (PAM) allows for the study of vocal animals on temporal and spatial scales difficult to achieve using only human observers. Recent improvements in recording technology, data storage, and battery capacity have led to increased use of PAM. One of the main obstacles in implementing wide-scale PAM programs is the lack of open-source programs that efficiently process terabytes of sound recordings and do not require large amounts of training data. Here we describe a workflow for detecting, classifying, and visualizing female Northern grey gibbon calls in Sabah, Malaysia. Our approach detects sound events using band-limited energy summation and does binary classification of these events (gibbon female or not) using machine learning algorithms (support vector machine and random forest). We then applied an unsupervised approach (affinity propagation clustering) to see if we could further differentiate between true and false positives or the number of gibbon females in our dataset. We used this workflow to address three questions: (1) does this automated approach provide reliable estimates of temporal patterns of gibbon calling activity; (2) can unsupervised approaches be applied as a post-processing step to improve the performance of the system; and (3) can unsupervised approaches be used to estimate how many female individuals (or clusters) there are in our study area? We found that performance plateaued with >160 clips of training data for each of our two classes. Using optimized settings, our automated approach achieved a satisfactory performance (F1 score ~ 80%). The unsupervised approach did not effectively differentiate between true and false positives or return clusters that appear to correspond to the number of females in our study area. Our results indicate that more work needs to be done before unsupervised approaches can be reliably used to estimate the number of individual animals occupying an area from PAM data. Future work applying these methods across sites and different gibbon species and comparisons to deep learning approaches will be crucial for future gibbon conservation initiatives across Southeast Asia.
Collapse
|
8
|
Drones and sound recorders increase the number of bird species identified: A combined surveys approach. ECOL INFORM 2023. [DOI: 10.1016/j.ecoinf.2023.101988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
9
|
Lapp S, Larkin JL, Parker HA, Larkin JT, Shaffer DR, Tett C, McNeil DJ, Fiss CJ, Kitzes J. Automated recognition of ruffed grouse drumming in field recordings. WILDLIFE SOC B 2022. [DOI: 10.1002/wsb.1395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Affiliation(s)
- Samuel Lapp
- University of Pittsburgh 103 Clapp Hall, Fifth and Ruskin Avenues Pittsburgh PA 15260 USA
| | - Jeffery L. Larkin
- Department of Biology Indiana University of Pennsylvania 1011 South Drive, Indiana, PA 15701 USA
- American Bird Conservancy The Plains VA 20198 USA
| | - Halie A. Parker
- Department of Biology Indiana University of Pennsylvania 1011 South Drive, Indiana, PA 15701 USA
| | - Jeffery T. Larkin
- Department of Environmental Conservation University of Massachusetts‐Amherst 160 Holdsworth Way Amherst MA 01003‐9285 USA
| | - Dakotah R. Shaffer
- Department of Biology Indiana University of Pennsylvania 1011 South Drive, Indiana, PA 15701 USA
| | - Carolyn Tett
- University of Pittsburgh 103 Clapp Hall, Fifth and Ruskin Avenues Pittsburgh PA 15260 USA
| | - Darin J. McNeil
- Department of Forestry and Natural Resources University of Kentucky Lexington KY 40546 USA
| | - Cameron J. Fiss
- Department of Environmental Biology State University of New York College of Environmental Science and Forestry 1 Forestry Dr. Syracuse NY 13210 USA
| | - Justin Kitzes
- University of Pittsburgh 103 Clapp Hall, Fifth and Ruskin Avenues Pittsburgh PA 15260 USA
| |
Collapse
|
10
|
Zhou J, Wang Y, Zhang C, Wu W, Ji Y, Zou Y. Eyebirds: Enabling the Public to Recognize Water Birds at Hand. Animals (Basel) 2022; 12:3000. [PMID: 36359124 PMCID: PMC9658372 DOI: 10.3390/ani12213000] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 10/25/2022] [Accepted: 10/26/2022] [Indexed: 09/29/2023] Open
Abstract
Enabling the public to easily recognize water birds has a positive effect on wetland bird conservation. However, classifying water birds requires advanced ornithological knowledge, which makes it very difficult for the public to recognize water bird species in daily life. To break the knowledge barrier of water bird recognition for the public, we construct a water bird recognition system (Eyebirds) by using deep learning, which is implemented as a smartphone app. Eyebirds consists of three main modules: (1) a water bird image dataset; (2) an attention mechanism-based deep convolution neural network for water bird recognition (AM-CNN); (3) an app for smartphone users. The waterbird image dataset currently covers 48 families, 203 genera and 548 species of water birds worldwide, which is used to train our water bird recognition model. The AM-CNN model employs attention mechanism to enhance the shallow features of bird images for boosting image classification performance. Experimental results on the North American bird dataset (CUB200-2011) show that the AM-CNN model achieves an average classification accuracy of 85%. On our self-built water bird image dataset, the AM-CNN model also works well with classification accuracies of 94.0%, 93.6% and 86.4% at three levels: family, genus and species, respectively. The user-side app is a WeChat applet deployed in smartphones. With the app, users can easily recognize water birds in expeditions, camping, sightseeing, or even daily life. In summary, our system can bring not only fun, but also water bird knowledge to the public, thus inspiring their interests and further promoting their participation in bird ecological conservation.
Collapse
Affiliation(s)
- Jiaogen Zhou
- Jiangsu Provincial Engineering Research Center for Intelligent Monitoring and Ecological Management of Pond and Reservoir Water Environment, Huaiyin Normal University, Huaian 223300, China
| | - Yang Wang
- Department of Computer Science and Technology, Tongji University, Shanghai 201804, China
| | - Caiyun Zhang
- Jiangsu Provincial Engineering Research Center for Intelligent Monitoring and Ecological Management of Pond and Reservoir Water Environment, Huaiyin Normal University, Huaian 223300, China
| | - Wenbo Wu
- Research Center of Information Technology, Beijing Academy of Agriculture and Forestry Sciences, Beijing 100097, China
| | - Yanzhu Ji
- Key Laboratory of Zoological Systematics and Evolution, Institute of Zoology, Chinese Academy of Sciences, Beijing 100101, China
| | - Yeai Zou
- Dongting Lake Station for Wetland Ecosystem Research, Institute of Subtropical Agriculture, Chinese Academy of Sciences, Changsha 410125, China
| |
Collapse
|
11
|
Mutanu L, Gohil J, Gupta K, Wagio P, Kotonya G. A Review of Automated Bioacoustics and General Acoustics Classification Research. SENSORS (BASEL, SWITZERLAND) 2022; 22:8361. [PMID: 36366061 PMCID: PMC9658612 DOI: 10.3390/s22218361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/01/2022] [Revised: 10/19/2022] [Accepted: 10/21/2022] [Indexed: 06/16/2023]
Abstract
Automated bioacoustics classification has received increasing attention from the research community in recent years due its cross-disciplinary nature and its diverse application. Applications in bioacoustics classification range from smart acoustic sensor networks that investigate the effects of acoustic vocalizations on species to context-aware edge devices that anticipate changes in their environment adapt their sensing and processing accordingly. The research described here is an in-depth survey of the current state of bioacoustics classification and monitoring. The survey examines bioacoustics classification alongside general acoustics to provide a representative picture of the research landscape. The survey reviewed 124 studies spanning eight years of research. The survey identifies the key application areas in bioacoustics research and the techniques used in audio transformation and feature extraction. The survey also examines the classification algorithms used in bioacoustics systems. Lastly, the survey examines current challenges, possible opportunities, and future directions in bioacoustics.
Collapse
Affiliation(s)
- Leah Mutanu
- Department of Computing, United States International University Africa, Nairobi P.O. Box 14634-0800, Kenya
| | - Jeet Gohil
- Department of Computing, United States International University Africa, Nairobi P.O. Box 14634-0800, Kenya
| | - Khushi Gupta
- Department of Computer Science, Sam Houston State University, Huntsville, TX 77341, USA
| | - Perpetua Wagio
- Department of Computing, United States International University Africa, Nairobi P.O. Box 14634-0800, Kenya
| | - Gerald Kotonya
- School of Computing and Communications, Lancaster University, Lacaster LA1 4WA, UK
| |
Collapse
|
12
|
Introducing the Software CASE (Cluster and Analyze Sound Events) by Comparing Different Clustering Methods and Audio Transformation Techniques Using Animal Vocalizations. Animals (Basel) 2022; 12:ani12162020. [PMID: 36009611 PMCID: PMC9404437 DOI: 10.3390/ani12162020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 07/28/2022] [Accepted: 08/04/2022] [Indexed: 11/17/2022] Open
Abstract
Simple Summary Unsupervised clustering algorithms are widely used in ecology and conservation to classify animal vocalizations, but also offer various advantages in basic research, contributing to the understanding of acoustic communication. Nevertheless, there are still some challenges to overcome. For instance, the quality of the clustering result depends on the audio transformation technique previously used to adjust the audio data. Moreover, it is difficult to verify the reliability of the clustering result. To analyze bioacoustic data using a clustering algorithm, it is, therefore, essential to select a reasonable algorithm from the many existing algorithms and prepare the recorded vocalizations so that the resulting values characterize a vocalization as accurately as possible. Frequency-modulated vocalizations, whose frequencies change over time, pose a particular problem. In this paper, we present the software CASE, which includes various clustering methods and provides an overview of their strengths and weaknesses concerning the classification of bioacoustic data. This software uses a multidimensional feature-extraction method to achieve better clustering results, especially for frequency-modulated vocalizations. Abstract Unsupervised clustering algorithms are widely used in ecology and conservation to classify animal sounds, but also offer several advantages in basic bioacoustics research. Consequently, it is important to overcome the existing challenges. A common practice is extracting the acoustic features of vocalizations one-dimensionally, only extracting an average value for a given feature for the entire vocalization. With frequency-modulated vocalizations, whose acoustic features can change over time, this can lead to insufficient characterization. Whether the necessary parameters have been set correctly and the obtained clustering result reliably classifies the vocalizations subsequently often remains unclear. The presented software, CASE, is intended to overcome these challenges. Established and new unsupervised clustering methods (community detection, affinity propagation, HDBSCAN, and fuzzy clustering) are tested in combination with various classifiers (k-nearest neighbor, dynamic time-warping, and cross-correlation) using differently transformed animal vocalizations. These methods are compared with predefined clusters to determine their strengths and weaknesses. In addition, a multidimensional data transformation procedure is presented that better represents the course of multiple acoustic features. The results suggest that, especially with frequency-modulated vocalizations, clustering is more applicable with multidimensional feature extraction compared with one-dimensional feature extraction. The characterization and clustering of vocalizations in multidimensional space offer great potential for future bioacoustic studies. The software CASE includes the developed method of multidimensional feature extraction, as well as all used clustering methods. It allows quickly applying several clustering algorithms to one data set to compare their results and to verify their reliability based on their consistency. Moreover, the software CASE determines the optimal values of most of the necessary parameters automatically. To take advantage of these benefits, the software CASE is provided for free download.
Collapse
|
13
|
Clark FE, Dunn JC. From Soundwave to Soundscape: A Guide to Acoustic Research in Captive Animal Environments. Front Vet Sci 2022; 9:889117. [PMID: 35782565 PMCID: PMC9244380 DOI: 10.3389/fvets.2022.889117] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2022] [Accepted: 05/23/2022] [Indexed: 11/17/2022] Open
Abstract
Sound is a complex feature of all environments, but captive animals' soundscapes (acoustic scenes) have been studied far less than those of wild animals. Furthermore, research across farms, laboratories, pet shelters, and zoos tends to focus on just one aspect of environmental sound measurement: its pressure level or intensity (in decibels). We review the state of the art of captive animal acoustic research and contrast this to the wild, highlighting new opportunities for the former to learn from the latter. We begin with a primer on sound, aimed at captive researchers and animal caregivers with an interest (rather than specific expertise) in acoustics. Then, we summarize animal acoustic research broadly split into measuring sound from animals, or their environment. We guide readers from soundwave to soundscape and through the burgeoning field of conservation technology, which offers new methods to capture multiple features of complex, gestalt soundscapes. Our review ends with suggestions for future research, and a practical guide to sound measurement in captive environments.
Collapse
Affiliation(s)
- Fay E. Clark
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, Cambridge, United Kingdom
- School of Psychological Science, Faculty of Life Sciences, University of Bristol, Bristol, United Kingdom
- *Correspondence: Fay E. Clark
| | - Jacob C. Dunn
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, Cambridge, United Kingdom
- Biological Anthropology, Department of Archaeology, University of Cambridge, Cambridge, United Kingdom
- Department of Cognitive Biology, University of Vienna, Vienna, Austria
| |
Collapse
|
14
|
Hou Y, Yu X, Yang J, Ouyang X, Fan D. Acoustic Sensor-Based Soundscape Analysis and Acoustic Assessment of Bird Species Richness in Shennongjia National Park, China. SENSORS (BASEL, SWITZERLAND) 2022; 22:4117. [PMID: 35684738 PMCID: PMC9185234 DOI: 10.3390/s22114117] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/22/2022] [Revised: 05/18/2022] [Accepted: 05/25/2022] [Indexed: 06/15/2023]
Abstract
Passive acoustic sensor-based soundscape analysis has become an increasingly important ecological method for evaluation of ecosystem conditions using acoustic indices. Understanding the soundscape composition and correlations between acoustic indices and species richness of birds, the most important sound source in the ecosystem, are of great importance for measuring biodiversity and the level of anthropogenic disturbance. In this study, based on yearlong sound data obtained from five acoustic sensors deployed in Dalongtan, Shennongjia National Park, we analyzed the soundscape composition by comparing the distributions of the soundscape power in different frequency ranges, and examined the correlations between acoustic indices and bird species richness by means of the Spearman rank correlation coefficient method. The diurnal dynamic characteristics of acoustic indices in different seasons were also described. Results showed that the majority of sounds were in the frequency of 2-8 kHz, in which over 50% sounds were in 2-6 kHz, commonly considered the bioacoustic frequency range. The Acoustics Complexity Index, Bioacoustic Index, and Normalized Difference Soundscape Index were significantly correlated with bird species richness, suggesting that these indices can be used for evaluation of bird species richness; Apparent diurnal dynamic patterns of bird acoustic activities were observed in spring, summer, and autumn; however, the intensity and duration of bird acoustic activities in summer is larger/longer than in spring and autumn.
Collapse
Affiliation(s)
- Yanan Hou
- Institute of Forest Resource Information Techniques, Chinese Academy of Forestry, Beijing 100091, China; (Y.H.); (X.O.); (D.F.)
- Department of Computer Science Engineering, Chengdu Neusoft University, Chengdu 611844, China
| | - Xinwen Yu
- Institute of Forest Resource Information Techniques, Chinese Academy of Forestry, Beijing 100091, China; (Y.H.); (X.O.); (D.F.)
- Key Laboratory of Forestry Remote Sensing and Information System, National Forest and Grassland Administration, Beijing 100091, China
| | - Jingyuan Yang
- Shennongjia National Park Administration, Shennongjia 442421, China;
| | - Xuan Ouyang
- Institute of Forest Resource Information Techniques, Chinese Academy of Forestry, Beijing 100091, China; (Y.H.); (X.O.); (D.F.)
- Key Laboratory of Forestry Remote Sensing and Information System, National Forest and Grassland Administration, Beijing 100091, China
| | - Dongpu Fan
- Institute of Forest Resource Information Techniques, Chinese Academy of Forestry, Beijing 100091, China; (Y.H.); (X.O.); (D.F.)
- Key Laboratory of Forestry Remote Sensing and Information System, National Forest and Grassland Administration, Beijing 100091, China
| |
Collapse
|
15
|
Wightman PH, Henrichs DW, Collier BA, Chamberlain MJ. Comparison of methods for automated identification of wild turkey gobbles. WILDLIFE SOC B 2022. [DOI: 10.1002/wsb.1246] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Affiliation(s)
- Patrick H. Wightman
- Warnell School of Forestry and Natural Resources University of Georgia Athens 30602 GA USA
| | - Darren W. Henrichs
- Department of Oceanography Texas A&M University College Station 77843 TX USA
| | - Bret A. Collier
- School of Renewable Natural Resources Louisiana State University Agricultural Center Baton Rouge 70803 LA USA
| | - Michael J. Chamberlain
- Warnell School of Forestry and Natural Resources University of Georgia Athens 30602 GA USA
| |
Collapse
|
16
|
|
17
|
Romero-Mujalli D, Bergmann T, Zimmermann A, Scheumann M. Utilizing DeepSqueak for automatic detection and classification of mammalian vocalizations: a case study on primate vocalizations. Sci Rep 2021; 11:24463. [PMID: 34961788 PMCID: PMC8712519 DOI: 10.1038/s41598-021-03941-1] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2021] [Accepted: 12/09/2021] [Indexed: 11/16/2022] Open
Abstract
Bioacoustic analyses of animal vocalizations are predominantly accomplished through manual scanning, a highly subjective and time-consuming process. Thus, validated automated analyses are needed that are usable for a variety of animal species and easy to handle by non-programing specialists. This study tested and validated whether DeepSqueak, a user-friendly software, developed for rodent ultrasonic vocalizations, can be generalized to automate the detection/segmentation, clustering and classification of high-frequency/ultrasonic vocalizations of a primate species. Our validation procedure showed that the trained detectors for vocalizations of the gray mouse lemur (Microcebus murinus) can deal with different call types, individual variation and different recording quality. Implementing additional filters drastically reduced noise signals (4225 events) and call fragments (637 events), resulting in 91% correct detections (Ntotal = 3040). Additionally, the detectors could be used to detect the vocalizations of an evolutionary closely related species, the Goodman’s mouse lemur (M. lehilahytsara). An integrated supervised classifier classified 93% of the 2683 calls correctly to the respective call type, and the unsupervised clustering model grouped the calls into clusters matching the published human-made categories. This study shows that DeepSqueak can be successfully utilized to detect, cluster and classify high-frequency/ultrasonic vocalizations of other taxa than rodents, and suggests a validation procedure usable to evaluate further bioacoustics software.
Collapse
Affiliation(s)
- Daniel Romero-Mujalli
- Institute of Zoology, University of Veterinary Medicine Hannover, Bünteweg 17, 30559, Hannover, Germany.
| | - Tjard Bergmann
- Institute of Zoology, University of Veterinary Medicine Hannover, Bünteweg 17, 30559, Hannover, Germany
| | | | - Marina Scheumann
- Institute of Zoology, University of Veterinary Medicine Hannover, Bünteweg 17, 30559, Hannover, Germany
| |
Collapse
|
18
|
Brown A, Montgomery J, Garg S. Automatic construction of accurate bioacoustics workflows under time constraints using a surrogate model. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107944] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
19
|
Bravo Sanchez FJ, Hossain MR, English NB, Moore ST. Bioacoustic classification of avian calls from raw sound waveforms with an open-source deep learning architecture. Sci Rep 2021; 11:15733. [PMID: 34344970 PMCID: PMC8333097 DOI: 10.1038/s41598-021-95076-6] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2020] [Accepted: 07/13/2021] [Indexed: 11/22/2022] Open
Abstract
The use of autonomous recordings of animal sounds to detect species is a popular conservation tool, constantly improving in fidelity as audio hardware and software evolves. Current classification algorithms utilise sound features extracted from the recording rather than the sound itself, with varying degrees of success. Neural networks that learn directly from the raw sound waveforms have been implemented in human speech recognition but the requirements of detailed labelled data have limited their use in bioacoustics. Here we test SincNet, an efficient neural network architecture that learns from the raw waveform using sinc-based filters. Results using an off-the-shelf implementation of SincNet on a publicly available bird sound dataset (NIPS4Bplus) show that the neural network rapidly converged reaching accuracies of over 65% with limited data. Their performance is comparable with traditional methods after hyperparameter tuning but they are more efficient. Learning directly from the raw waveform allows the algorithm to select automatically those elements of the sound that are best suited for the task, bypassing the onerous task of selecting feature extraction techniques and reducing possible biases. We use publicly released code and datasets to encourage others to replicate our results and to apply SincNet to their own datasets; and we review possible enhancements in the hope that algorithms that learn from the raw waveform will become useful bioacoustic tools.
Collapse
Affiliation(s)
- Francisco J Bravo Sanchez
- School of Engineering and Technology, Central Queensland University, North Rockhampton, QLD, Australia
| | - Md Rahat Hossain
- School of Engineering and Technology, Central Queensland University, North Rockhampton, QLD, Australia
| | - Nathan B English
- School of Health, Medical and Applied Sciences, Flora, Fauna and Freshwater Research, Central Queensland University, Townsville, QLD, Australia
| | - Steven T Moore
- School of Engineering and Technology, Central Queensland University, North Rockhampton, QLD, Australia.
| |
Collapse
|
20
|
Brooks DR, Nocera JJ. Using autonomous recording units and change-point analysis to determine reproductive activity in an aerial insectivore. BIOACOUSTICS 2021. [DOI: 10.1080/09524622.2021.1921617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Affiliation(s)
- Delaney R. Brooks
- Faculty of Forestry and Environmental Management, University of New Brunswick, Fredericton, Canada
| | - Joseph J. Nocera
- Faculty of Forestry and Environmental Management, University of New Brunswick, Fredericton, Canada
| |
Collapse
|
21
|
Clink DJ, Klinck H. Unsupervised acoustic classification of individual gibbon females and the implications for passive acoustic monitoring. Methods Ecol Evol 2020. [DOI: 10.1111/2041-210x.13520] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/29/2023]
Affiliation(s)
- Dena J. Clink
- Center for Conservation Bioacoustics Cornell Laboratory of Ornithology Cornell University Ithaca NY USA
| | - Holger Klinck
- Center for Conservation Bioacoustics Cornell Laboratory of Ornithology Cornell University Ithaca NY USA
| |
Collapse
|
22
|
|
23
|
Byerly PA, Waddle JH, Romero Premeaux A, Leberg PL. Effects of barrier island salt marsh restoration on marsh bird occurrence in the northern Gulf of Mexico. Restor Ecol 2020. [DOI: 10.1111/rec.13222] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Affiliation(s)
- Paige A. Byerly
- Department of Biology University of Louisiana at Lafayette 410 E. St. Mary Blvd. Billeaud Hall, Room 108, Lafayette Louisiana 70503 U.S.A
- USGS Wetlands and Aquatics Research Center 7920 NW 71 St, Gainesville Florida 32653 U.S.A
| | - J. Hardin Waddle
- USGS Wetlands and Aquatics Research Center 7920 NW 71 St, Gainesville Florida 32653 U.S.A
| | - Alexis Romero Premeaux
- Department of Biology University of Louisiana at Lafayette 410 E. St. Mary Blvd. Billeaud Hall, Room 108, Lafayette Louisiana 70503 U.S.A
| | - Paul L. Leberg
- Department of Biology University of Louisiana at Lafayette 410 E. St. Mary Blvd. Billeaud Hall, Room 108, Lafayette Louisiana 70503 U.S.A
| |
Collapse
|
24
|
Darras KF, Deppe F, Fabian Y, Kartono AP, Angulo A, Kolbrek B, Mulyani YA, Prawiradilaga DM. High microphone signal-to-noise ratio enhances acoustic sampling of wildlife. PeerJ 2020; 8:e9955. [PMID: 33150056 PMCID: PMC7585376 DOI: 10.7717/peerj.9955] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2020] [Accepted: 08/25/2020] [Indexed: 11/20/2022] Open
Abstract
BACKGROUND Automated sound recorders are a popular sampling tool in ecology. However, the microphones themselves received little attention so far, and specifications that determine the recordings' sound quality are seldom mentioned. Here, we demonstrate the importance of microphone signal-to-noise ratio for sampling sonant animals. METHODS We tested 12 different microphone models in the field and measured their signal-to-noise ratios and detection ranges. We also measured the vocalisation activity of birds and bats that they recorded, the bird species richness, the bat call types richness, as well as the performance of automated detection of bird and bat calls. We tested the relationship of each one of these measures with signal-to-noise ratio in statistical models. RESULTS Microphone signal-to-noise ratio positively affects the sound detection space areas, which increased by a factor of 1.7 for audible sound, and 10 for ultrasound, from the lowest to the highest signal-to-noise ratio microphone. Consequently, the sampled vocalisation activity increased by a factor of 1.6 for birds, and 9.7 for bats. Correspondingly, the species pool of birds and bats could not be completely detected by the microphones with lower signal-to-noise ratio. The performance of automated detection of bird and bat calls, as measured by its precision and recall, increased significantly with microphone signal-to-noise ratio. DISCUSSION Microphone signal-to-noise ratio is a crucial characteristic of a sound recording system, positively affecting the acoustic sampling performance of birds and bats. It should be maximised by choosing appropriate microphones, and be quantified independently, especially in the ultrasound range.
Collapse
Affiliation(s)
- Kevin F.A. Darras
- Agroecology, University of Göttingen, Göttingen, Niedersachsen, Germany
| | - Franziska Deppe
- Agroecology, University of Göttingen, Göttingen, Niedersachsen, Germany
| | - Yvonne Fabian
- Agroecology, University of Göttingen, Göttingen, Niedersachsen, Germany
- Agroscope FAL Reckenholz, Swiss Federal Research Station for Agroecology and Agriculture, Zurich, Switzerland
| | - Agus P. Kartono
- Department of Forest Resources, Conservation and Ecotourism, Bogor Institute of Agriculture, Bogor, Indonesia
| | - Andres Angulo
- Agroecology, University of Göttingen, Göttingen, Niedersachsen, Germany
| | - Bjørn Kolbrek
- Celestion International, Ipswich, Suffolk, United Kingdom
| | - Yeni A. Mulyani
- Department of Forest Resources, Conservation and Ecotourism, Bogor Institute of Agriculture, Bogor, Indonesia
| | | |
Collapse
|
25
|
Gentry KE, Lewis RN, Glanz H, Simões PI, Nyári ÁS, Reichert MS. Bioacoustics in cognitive research: Applications, considerations, and recommendations. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2020; 11:e1538. [PMID: 32548958 DOI: 10.1002/wcs.1538] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2019] [Revised: 05/15/2020] [Accepted: 05/18/2020] [Indexed: 12/23/2022]
Abstract
The multifaceted ability to produce, transmit, receive, and respond to acoustic signals is widespread in animals and forms the basis of the interdisciplinary science of bioacoustics. Bioacoustics research methods, including sound recording and playback experiments, are applicable in cognitive research that centers around the processing of information from the acoustic environment. We provide an overview of bioacoustics techniques in the context of cognitive studies and make the case for the importance of bioacoustics in the study of cognition by outlining some of the major cognitive processes in which acoustic signals are involved. We also describe key considerations associated with the recording of sound and its use in cognitive applications. Based on these considerations, we provide a set of recommendations for best practices in the recording and use of acoustic signals in cognitive studies. Our aim is to demonstrate that acoustic recordings and stimuli are valuable tools for cognitive researchers when used appropriately. In doing so, we hope to stimulate opportunities for innovative cognitive research that incorporates robust recording protocols. This article is categorized under: Neuroscience > Cognition Psychology > Theory and Methods Neuroscience > Behavior Neuroscience > Cognition.
Collapse
Affiliation(s)
- Katherine E Gentry
- Division of Habitat and Species Conservation, Florida Fish and Wildlife Conservation Commission, Tallahassee, Florida, USA
| | - Rebecca N Lewis
- Department of Earth and Environmental Sciences, The University of Manchester, Manchester, UK.,Chester Zoo, Chester, UK
| | - Hunter Glanz
- Statistics Department, California Polytechnic State University, San Luis Obispo, California, USA
| | - Pedro I Simões
- Departmento de Zoologia, Universidade Federal de Pernambuco, Recife, Brazil
| | - Árpád S Nyári
- Department of Ecology and Evolutionary Biology, University of Tennessee, Knoxville, Tennessee, USA
| | - Michael S Reichert
- Department of Integrative Biology, Oklahoma State University, Stillwater, Oklahoma, USA
| |
Collapse
|
26
|
Xie J, Hu K, Zhu M, Guo Y. Data-driven analysis of global research trends in bioacoustics and ecoacoustics from 1991 to 2018. ECOL INFORM 2020. [DOI: 10.1016/j.ecoinf.2020.101068] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
27
|
Stehle M, Lasseck M, Khorramshahi O, Sturm U. Evaluation of acoustic pattern recognition of nightingale (Luscinia megarhynchos) recordings by citizens. RESEARCH IDEAS AND OUTCOMES 2020. [DOI: 10.3897/rio.6.e50233] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Acoustic pattern recognition methods introduce new perspectives for species identification, biodiversity monitoring and data validation in citizen science but are rarely evaluated in real world scenarios. In this case study we analysed the performance of a machine learning algorithm for automated bird identification to reliably identify common nightingales (Luscinia megarhynchos) in field recordings taken by users of the smartphone app Naturblick. We found that the performance of the automated identification tool was overall robust in our selected recordings. Although most of the recordings had a relatively low confidence score, a large proportion of the recordings were identified correctly.
Collapse
|
28
|
Bezerra DMM, Simões CRMDA, Araújo CBD, Machado CCC, Maia RR, Alves RRN, Araujo HFPD. Habitat use, density, and conservation status of the white-browed guan (Penelope jacucaca Spix, 1825). J Nat Conserv 2019. [DOI: 10.1016/j.jnc.2019.125733] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
29
|
Darras K, Batáry P, Furnas BJ, Grass I, Mulyani YA, Tscharntke T. Autonomous sound recording outperforms human observation for sampling birds: a systematic map and user guide. ECOLOGICAL APPLICATIONS : A PUBLICATION OF THE ECOLOGICAL SOCIETY OF AMERICA 2019; 29:e01954. [PMID: 31206926 DOI: 10.1002/eap.1954] [Citation(s) in RCA: 42] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/02/2018] [Revised: 04/17/2019] [Accepted: 04/23/2019] [Indexed: 06/09/2023]
Abstract
Autonomous sound recording techniques have gained considerable traction in the last decade, but the question remains whether they can replace human observation surveys to sample sonant animals. For birds in particular, survey methods have been tested extensively using point counts and sound recording surveys. Here, we review the latest evidence for this taxon within the frame of a systematic map. We compare sampling effectiveness of these two survey methods, the output variables they produce, and their practicality. When assessed against the standard of point counts, autonomous sound recording proves to be a powerful tool that samples at least as many species. This technology can monitor birds in an exhaustive, standardized, and verifiable way. Moreover, sound recorders give access to entire soundscapes from which new data types can be derived (vocal activity, acoustic indices). Variables such as abundance, density, occupancy, or species richness can be obtained to yield data sets that are comparable to and compatible with point counts. Finally, autonomous sound recorders allow investigations at high temporal and spatial resolution and coverage, which are more cost effective and cannot be achieved by human observations alone, even though small-scale studies might be more cost effective when carried out with point counts. Sound recorders can be deployed in many places, they are more scalable and reliable, making them the better choice for bird surveys in an increasingly data-driven time. We provide an overview of currently available recorders and discuss their specifications to guide future study designs.
Collapse
Affiliation(s)
- Kevin Darras
- Agroecology, Department of Crop Sciences, University of Goettingen, Grisebachstrasse 6, 37077, Göttingen, Germany
| | - Péter Batáry
- Agroecology, Department of Crop Sciences, University of Goettingen, Grisebachstrasse 6, 37077, Göttingen, Germany
- Lendület Landscape and Conservation Ecology, Institute of Ecology and Botany, MTA Centre for Ecological Research, Alkotmány u. 2-4, 2163, Vácrátót, Hungary
| | - Brett J Furnas
- Wildlife Investigations Laboratory, California Department of Fish and Wildlife, 1701 Nimbus Road, Suite D, Sacramento, California, 95670, USA
| | - Ingo Grass
- Agroecology, Department of Crop Sciences, University of Goettingen, Grisebachstrasse 6, 37077, Göttingen, Germany
| | - Yeni A Mulyani
- Department of Forest Resources Conservation and Ecotourism, Faculty of Forestry, Bogor Agricultural University, Bogor, Indonesia
| | - Teja Tscharntke
- Agroecology, Department of Crop Sciences, University of Goettingen, Grisebachstrasse 6, 37077, Göttingen, Germany
| |
Collapse
|
30
|
Bergler C, Schröter H, Cheng RX, Barth V, Weber M, Nöth E, Hofer H, Maier A. ORCA-SPOT: An Automatic Killer Whale Sound Detection Toolkit Using Deep Learning. Sci Rep 2019; 9:10997. [PMID: 31358873 PMCID: PMC6662697 DOI: 10.1038/s41598-019-47335-w] [Citation(s) in RCA: 34] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2018] [Accepted: 07/12/2019] [Indexed: 11/09/2022] Open
Abstract
Large bioacoustic archives of wild animals are an important source to identify reappearing communication patterns, which can then be related to recurring behavioral patterns to advance the current understanding of intra-specific communication of non-human animals. A main challenge remains that most large-scale bioacoustic archives contain only a small percentage of animal vocalizations and a large amount of environmental noise, which makes it extremely difficult to manually retrieve sufficient vocalizations for further analysis - particularly important for species with advanced social systems and complex vocalizations. In this study deep neural networks were trained on 11,509 killer whale (Orcinus orca) signals and 34,848 noise segments. The resulting toolkit ORCA-SPOT was tested on a large-scale bioacoustic repository - the Orchive - comprising roughly 19,000 hours of killer whale underwater recordings. An automated segmentation of the entire Orchive recordings (about 2.2 years) took approximately 8 days. It achieved a time-based precision or positive-predictive-value (PPV) of 93.2% and an area-under-the-curve (AUC) of 0.9523. This approach enables an automated annotation procedure of large bioacoustics databases to extract killer whale sounds, which are essential for subsequent identification of significant communication patterns. The code will be publicly available in October 2019 to support the application of deep learning to bioaoucstic research. ORCA-SPOT can be adapted to other animal species.
Collapse
Affiliation(s)
- Christian Bergler
- Friedrich-Alexander-University Erlangen-Nuremberg, Department of Computer Science, Pattern Recognition Lab, Martensstr. 3, 91058, Erlangen, Germany.
| | - Hendrik Schröter
- Friedrich-Alexander-University Erlangen-Nuremberg, Department of Computer Science, Pattern Recognition Lab, Martensstr. 3, 91058, Erlangen, Germany
| | - Rachael Xi Cheng
- Department of Ecological Dynamics, Leibniz Institute for Zoo and Wildlife Research (IZW) in the Forschungsverbund Berlin e.V., Alfred-Kowalke-Straße 17, 10315, Berlin, Germany
| | - Volker Barth
- Anthro-Media, Nansenstr. 19, 12047, Berlin, Germany
| | | | - Elmar Nöth
- Friedrich-Alexander-University Erlangen-Nuremberg, Department of Computer Science, Pattern Recognition Lab, Martensstr. 3, 91058, Erlangen, Germany.
| | - Heribert Hofer
- Department of Ecological Dynamics, Leibniz Institute for Zoo and Wildlife Research (IZW) in the Forschungsverbund Berlin e.V., Alfred-Kowalke-Straße 17, 10315, Berlin, Germany
- Department of Biology, Chemistry, Pharmacy, Freie Universität Berlin, Takustrasse 3, 14195, Berlin, Germany
- Department of Veterinary Medicine, Freie Universität Berlin, Oertzenweg 19b, 14195, Berlin, Germany
| | - Andreas Maier
- Friedrich-Alexander-University Erlangen-Nuremberg, Department of Computer Science, Pattern Recognition Lab, Martensstr. 3, 91058, Erlangen, Germany
| |
Collapse
|
31
|
Xie J, Zhu M. Handcrafted features and late fusion with deep learning for bird sound classification. ECOL INFORM 2019. [DOI: 10.1016/j.ecoinf.2019.05.007] [Citation(s) in RCA: 34] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
32
|
Mcloughlin MP, Stewart R, McElligott AG. Automated bioacoustics: methods in ecology and conservation and their potential for animal welfare monitoring. J R Soc Interface 2019; 16:20190225. [PMID: 31213168 PMCID: PMC6597774 DOI: 10.1098/rsif.2019.0225] [Citation(s) in RCA: 42] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2019] [Accepted: 05/16/2019] [Indexed: 11/12/2022] Open
Abstract
Vocalizations carry emotional, physiological and individual information. This suggests that they may serve as potentially useful indicators for inferring animal welfare. At the same time, automated methods for analysing and classifying sound have developed rapidly, particularly in the fields of ecology, conservation and sound scene classification. These methods are already used to automatically classify animal vocalizations, for example, in identifying animal species and estimating numbers of individuals. Despite this potential, they have not yet found widespread application in animal welfare monitoring. In this review, we first discuss current trends in sound analysis for ecology, conservation and sound classification. Following this, we detail the vocalizations produced by three of the most important farm livestock species: chickens ( Gallus gallus domesticus), pigs ( Sus scrofa domesticus) and cattle ( Bos taurus). Finally, we describe how these methods can be applied to monitor animal welfare with new potential for developing automated methods for large-scale farming.
Collapse
Affiliation(s)
- Michael P. Mcloughlin
- Centre for Digital Music, School of Electronic Engineering and Computer Science, Queen Mary University of London, Mile End Campus, London, UK
| | - Rebecca Stewart
- Centre for Digital Music, School of Electronic Engineering and Computer Science, Queen Mary University of London, Mile End Campus, London, UK
| | - Alan G. McElligott
- Centre for Research in Ecology, Evolution and Behaviour, Department of Life Sciences, University of Roehampton, London, UK
| |
Collapse
|
33
|
Acoustic recordings provide detailed information regarding the behavior of cryptic wildlife to support conservation translocations. Sci Rep 2019; 9:5172. [PMID: 30914700 PMCID: PMC6435668 DOI: 10.1038/s41598-019-41455-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2018] [Accepted: 02/26/2019] [Indexed: 12/02/2022] Open
Abstract
For translocated animals, behavioral competence may be key to post-release survival. However, monitoring behavior is typically limited to tracking movements or inferring behavior at a gross scale via collar-mounted sensors. Animal-bourne acoustic monitoring may provide a unique opportunity to monitor behavior at a finer scale. The giant panda is an elusive species of Ursid that is vulnerable to extinction. Translocation is an important aspect of the species’ recovery, and survival and recruitment for pandas likely hinge on behavioral competence. Here we tested the efficacy of a collar-mounted acoustic recording unit (ARU) to remotely monitor the behavior of panda mothers and their dependent young. We found that trained human listeners could reliably identify 10 behaviors from acoustic recordings. Through visual inspection of spectrograms we further identified 5 behavioral categories that may be detectable by automated pattern recognition, an approach that is essential for the practical application of ARU. These results suggest that ARU are a viable method for remotely observing behaviors, including feeding. With targeted effort directed towards instrumentation and computing advances, ARU could be used to document how behavioral competence supports or challenges post-release survival and recruitment, and allow for research findings to be adaptively integrated into future translocation efforts.
Collapse
|
34
|
Stiffler LL, Schroeder KM, Anderson JT, McRae SB, Katzner TE. Quantitative acoustic differentiation of cryptic species illustrated with King and Clapper rails. Ecol Evol 2018; 8:12821-12831. [PMID: 30619585 PMCID: PMC6309001 DOI: 10.1002/ece3.4711] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2018] [Revised: 10/01/2018] [Accepted: 10/14/2018] [Indexed: 11/12/2022] Open
Abstract
Reliable species identification is vital for survey and monitoring programs. Recently, the development of digital technology for recording and analyzing vocalizations has assisted in acoustic surveying for cryptic, rare, or elusive species. However, the quantitative tools that exist for species differentiation are still being refined. Using vocalizations recorded in the course of ecological studies of a King Rail (Rallus elegans) and a Clapper Rail (Rallus crepitans) population, we assessed the accuracy and effectiveness of three parametric (logistic regression, discriminant function analysis, quadratic discriminant function analysis) and six nonparametric (support vector machine, CART, Random Forest, k-nearest neighbor, weighted k-nearest neighbor, and neural networks) statistical classification methods for differentiating these species by their kek mating call. We identified 480 kek notes of each species and quantitatively characterized them with five standardized acoustic parameters. Overall, nonparametric classification methods outperformed parametric classification methods for species differentiation (nonparametric tools were between 57% and 81% accurate, parametric tools were between 57% and 60% accurate). Of the nine classification methods, Random Forest was the most accurate and precise, resulting in 81.1% correct classification of kek notes to species. This suggests that the mating calls of these sister species are likely difficult for human observers to tell apart. However, it also implies that appropriate statistical tools may allow reasonable species-level classification accuracy of recorded calls and provide an alternative to species classification where other capture- or genotype-based survey techniques are not possible.
Collapse
Affiliation(s)
- Lydia L. Stiffler
- Division of Forestry and Natural ResourcesWest Virginia UniversityMorgantownWest Virginia
- Present address:
Warnell School of Forestry and Natural ResourcesUniversity of GeorgiaAthensGeorgia
| | - Katie M. Schroeder
- Department of BiologyEast Carolina UniversityGreenvilleNorth Carolina
- Present address:
Department of BiologyUniversity of Massachusetts AmherstAmherstMassachusetts
| | - James T. Anderson
- Division of Forestry and Natural ResourcesWest Virginia UniversityMorgantownWest Virginia
| | - Susan B. McRae
- Department of BiologyEast Carolina UniversityGreenvilleNorth Carolina
| | - Todd E. Katzner
- U.S. Geological SurveyForest & Rangeland Ecosystem Science CenterBoiseIdaho
| |
Collapse
|
35
|
Penone C, Kerbiriou C, Julien JF, Marmet J, Le Viol I. Body size information in large-scale acoustic bat databases. PeerJ 2018; 6:e5370. [PMID: 30155347 PMCID: PMC6110253 DOI: 10.7717/peerj.5370] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2018] [Accepted: 07/13/2018] [Indexed: 12/03/2022] Open
Abstract
Background Citizen monitoring programs using acoustic data have been useful for detecting population and community patterns. However, they have rarely been used to study broad scale patterns of species traits. We assessed the potential of acoustic data to detect broad scale patterns in body size. We compared geographical patterns in body size with acoustic signals in the bat species Pipistrellus pipistrellus. Given the correlation between body size and acoustic characteristics, we expected to see similar results when analyzing the relationships of body size and acoustic signals with climatic variables. Methods We assessed body size using forearm length measurements of 1,359 bats, captured by mist nets in France. For acoustic analyses, we used an extensive dataset collected through the French citizen bat survey. We isolated each bat echolocation call (n = 4,783) and performed automatic measures of signals, including the frequency of the flattest part of the calls (characteristic frequency). We then examined the relationship between forearm length, characteristic frequencies, and two components resulting from principal component analysis for geographic (latitude, longitude) and climatic variables. Results Forearm length was positively correlated with higher precipitation, lower seasonality, and lower temperatures. Lower characteristic frequencies (i.e., larger body size) were mostly related to lower temperatures and northern latitudes. While conducted on different datasets, the two analyses provided congruent results. Discussion Acoustic data from citizen science programs can thus be useful for the detection of large-scale patterns in body size. This first analysis offers a new perspective for the use of large acoustic databases to explore biological patterns and to address both theoretical and applied questions.
Collapse
Affiliation(s)
- Caterina Penone
- Institute of Plant Sciences, University of Bern, Bern, Switzerland
| | - Christian Kerbiriou
- CESCO UMR7204 MNHN-UPMC-CNRS-Sorbonne Université, Université Pierre et Marie Curie (Paris VI), Paris, France.,Marine Station, Muséum national d'Histoire naturelle, Concarneau, France
| | - Jean-François Julien
- CESCO UMR7204 MNHN-UPMC-CNRS-Sorbonne Université, Muséum national d'Histoire naturelle, Paris, France
| | - Julie Marmet
- CESCO UMR7204 MNHN-UPMC-CNRS-Sorbonne Université, Muséum national d'Histoire naturelle, Paris, France
| | - Isabelle Le Viol
- Marine Station, Muséum national d'Histoire naturelle, Concarneau, France.,CESCO UMR7204 MNHN-UPMC-CNRS-Sorbonne Université, Muséum national d'Histoire naturelle, Paris, France
| |
Collapse
|
36
|
Brown A, Garg S, Montgomery J. Scalable preprocessing of high volume environmental acoustic data for bioacoustic monitoring. PLoS One 2018; 13:e0201542. [PMID: 30075012 PMCID: PMC6075764 DOI: 10.1371/journal.pone.0201542] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2018] [Accepted: 07/17/2018] [Indexed: 11/18/2022] Open
Abstract
In this work, we examine the problem of efficiently preprocessing and denoising high volume environmental acoustic data, which is a necessary step in many bird monitoring tasks. Preprocessing is typically made up of multiple steps which are considered separately from each other. These are often resource intensive, particularly because the volume of data involved is high. We focus on addressing two challenges within this problem: how to combine existing preprocessing tasks while maximising the effectiveness of each step, and how to process this pipeline quickly and efficiently, so that it can be used to process high volumes of acoustic data. We describe a distributed system designed specifically for this problem, utilising a master-slave model with data parallelisation. By investigating the impact of individual preprocessing tasks on each other, and their execution times, we determine an efficient and accurate order for preprocessing tasks within the distributed system. We find that, using a single core, our pipeline executes 1.40 times faster compared to manually executing all preprocessing tasks. We then apply our pipeline in the distributed system and evaluate its performance. We find that our system is capable of preprocessing bird acoustic recordings at a rate of 174.8 seconds of audio per second of real time with 32 cores over 8 virtual machines, which is 21.76 times faster than a serial process.
Collapse
Affiliation(s)
- Alexander Brown
- School of Technology, Environments and Design, University of Tasmania, Hobart, Tasmania, Australia
| | - Saurabh Garg
- School of Technology, Environments and Design, University of Tasmania, Hobart, Tasmania, Australia
| | - James Montgomery
- School of Technology, Environments and Design, University of Tasmania, Hobart, Tasmania, Australia
| |
Collapse
|
37
|
Improving Classification Algorithms by Considering Score Series in Wireless Acoustic Sensor Networks. SENSORS 2018; 18:s18082465. [PMID: 30061506 PMCID: PMC6111609 DOI: 10.3390/s18082465] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/11/2018] [Revised: 07/22/2018] [Accepted: 07/27/2018] [Indexed: 11/17/2022]
Abstract
The reduction in size, power consumption and price of many sensor devices has enabled the deployment of many sensor networks that can be used to monitor and control several aspects of various habitats. More specifically, the analysis of sounds has attracted a huge interest in urban and wildlife environments where the classification of the different signals has become a major issue. Various algorithms have been described for this purpose, a number of which frame the sound and classify these frames, while others take advantage of the sequential information embedded in a sound signal. In the paper, a new algorithm is proposed that, while maintaining the frame-classification advantages, adds a new phase that considers and classifies the score series derived after frame labelling. These score series are represented using cepstral coefficients and classified using standard machine-learning classifiers. The proposed algorithm has been applied to a dataset of anuran calls and its results compared to the performance obtained in previous experiments on sensor networks. The main outcome of our research is that the consideration of score series strongly outperforms other algorithms and attains outstanding performance despite the noisy background commonly encountered in this kind of application.
Collapse
|
38
|
Darras K, Batáry P, Furnas B, Celis‐Murillo A, Van Wilgenburg SL, Mulyani YA, Tscharntke T. Comparing the sampling performance of sound recorders versus point counts in bird surveys: A meta‐analysis. J Appl Ecol 2018. [DOI: 10.1111/1365-2664.13229] [Citation(s) in RCA: 63] [Impact Index Per Article: 10.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2022]
Affiliation(s)
- Kevin Darras
- Department of Crop Sciences, AgroecologyUniversity of Goettingen Göttingen Germany
| | - Péter Batáry
- Department of Crop Sciences, AgroecologyUniversity of Goettingen Göttingen Germany
- MTA Centre for Ecological ResearchGINOP Sustainable Ecosystems Group Tihany Hungary
| | - Brett Furnas
- Wildlife Investigations LaboratoryCalifornia Department of Fish and Wildlife Sacramento California
| | - Antonio Celis‐Murillo
- Illinois Natural History SurveyPrairie Research InstituteUniversity of Illinois at Urbana‐Champaign Champaign Illinois
| | | | - Yeni A. Mulyani
- Department of Forest Resources Conservation and EcotourismFaculty of ForestryBogor Agricultural University Bogor Indonesia
| | - Teja Tscharntke
- Department of Crop Sciences, AgroecologyUniversity of Goettingen Göttingen Germany
| |
Collapse
|
39
|
Williams EM, O'Donnell CFJ, Armstrong DP. Cost-benefit analysis of acoustic recorders as a solution to sampling challenges experienced monitoring cryptic species. Ecol Evol 2018; 8:6839-6848. [PMID: 30038779 PMCID: PMC6053556 DOI: 10.1002/ece3.4199] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2018] [Revised: 04/06/2018] [Accepted: 04/22/2018] [Indexed: 11/06/2022] Open
Abstract
The inferences that can be made from any study are limited by the quality of the sampling design. By bad luck, when monitoring species that are difficult to detect (cryptic), sampling designs become dictated by what is feasible rather than what is desired. We calibrated and conducted a cost-benefit analysis of four acoustic recorder options that were being considered as potential solutions to several sampling restrictions experienced while monitoring the Australasian bittern, a cryptic wetland bird. Such sampling restrictions are commonly experienced while monitoring many different endangered species, particularly those that are cryptic. The recorder options included mono and stereo devices, with two sound file processing options (visual and audible analysis). Recording devices provided call-count data similar to those collected by field observers but at a fraction of the cost, which meant that "idealistic" sampling regimes, previously thought to be too expensive, became feasible for bitterns. Our study is one of the few to assess the monetary value of recording devices in the context of data quality, allowing trade-offs (and potential solutions) commonly experienced while monitoring cryptic endangered species to be shown and compared more clearly. The ability to overcome challenges of monitoring cryptic species in this way increases research possibilities for data deficient species and is applicable to any species with similar monitoring challenges.
Collapse
Affiliation(s)
- Emma M. Williams
- Matuku EcologyChristchurchNew Zealand
- Wildlife Ecology GroupMassey UniversityPalmerston NorthNew Zealand
- Department of ConservationBiodiversity GroupChristchurchNew Zealand
| | | | | |
Collapse
|
40
|
Oliver RY, Ellis DPW, Chmura HE, Krause JS, Pérez JH, Sweet SK, Gough L, Wingfield JC, Boelman NT. Eavesdropping on the Arctic: Automated bioacoustics reveal dynamics in songbird breeding phenology. SCIENCE ADVANCES 2018; 4:eaaq1084. [PMID: 29938220 PMCID: PMC6010323 DOI: 10.1126/sciadv.aaq1084] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/02/2017] [Accepted: 05/14/2018] [Indexed: 06/01/2023]
Abstract
Bioacoustic networks could vastly expand the coverage of wildlife monitoring to complement satellite observations of climate and vegetation. This approach would enable global-scale understanding of how climate change influences phenomena such as migratory timing of avian species. The enormous data sets that autonomous recorders typically generate demand automated analyses that remain largely undeveloped. We devised automated signal processing and machine learning approaches to estimate dates on which songbird communities arrived at arctic breeding grounds. Acoustically estimated dates agreed well with those determined via traditional surveys and were strongly related to the landscape's snow-free dates. We found that environmental conditions heavily influenced daily variation in songbird vocal activity, especially before egg laying. Our novel approaches demonstrate that variation in avian migratory arrival can be detected autonomously. Large-scale deployment of this innovation in wildlife monitoring would enable the coverage necessary to assess and forecast changes in bird migration in the face of climate change.
Collapse
Affiliation(s)
- Ruth Y. Oliver
- Department of Earth and Environmental Sciences, Columbia University, New York, NY 10027, USA
- Lamont-Doherty Earth Observatory, Columbia University, Palisades, NY 10964, USA
| | | | - Helen E. Chmura
- Department of Neurobiology, Physiology, and Behavior, University of California, Davis, Davis, CA 95616, USA
| | - Jesse S. Krause
- Department of Neurobiology, Physiology, and Behavior, University of California, Davis, Davis, CA 95616, USA
| | - Jonathan H. Pérez
- Department of Neurobiology, Physiology, and Behavior, University of California, Davis, Davis, CA 95616, USA
| | - Shannan K. Sweet
- Horticulture Section, School of Integrative Plant Science, Cornell University, Ithaca, NY 14853, USA
| | - Laura Gough
- Department of Biological Sciences, Towson University, Towson, MD 21252, USA
| | - John C. Wingfield
- Department of Neurobiology, Physiology, and Behavior, University of California, Davis, Davis, CA 95616, USA
| | - Natalie T. Boelman
- Department of Earth and Environmental Sciences, Columbia University, New York, NY 10027, USA
- Lamont-Doherty Earth Observatory, Columbia University, Palisades, NY 10964, USA
| |
Collapse
|
41
|
Fontana I, Tullo E, Carpentier L, Berckmans D, Butterworth A, Vranken E, Norton T, Berckmans D, Guarino M. Sound analysis to model weight of broiler chickens. Poult Sci 2018; 96:3938-3943. [PMID: 29050436 DOI: 10.3382/ps/pex215] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2016] [Accepted: 08/04/2017] [Indexed: 11/20/2022] Open
Abstract
The pattern of body weight gain during the commercial growing of broiler chickens is important to understand growth and feed conversion ratio of each flock.The application of sound analysis techniques has been widely studied to measure and analyze the amplitude and frequency of animal sounds. Previous studies have shown a significant correlation (P ≤ 0.001) between the frequency of vocalization and the age and weight of broilers. Therefore, the aim of this study was to identify and validate a model that describes the growth rate of broiler chickens based on the peak frequency of their vocalizations and to explore the possibility to develop a tool capable of automatically detecting the growth of the chickens based on the frequency of their vocalizations during the production cycle. It is part of an overall goal to develop a Precision Livestock Farming tool that assists farmers in monitoring the growth of broiler chickens during the production cycle. In the present study, sounds and body weight were continuously recorded in an intensive broiler farm during 5 production cycles. For each cycle the peak frequencies of the chicken vocalizations were used to estimate the weight and then they were compared with the observed weight of the birds automatically measured using on farm automated weighing devices. No significant difference is shown between expected and observed weights along the entire production cycles; this trend was confirmed by the correlation coefficient between expected and observed weights (r = 96%, P value ≤ 0.001).The identified model used to predict the weight as a function of the peak frequency confirmed that bird weight might be predicted by the frequency analysis of the sounds emitted at farm level. Even if the precision of the weighing method based on sounds investigated in this study has to be improved, it gives a reasonable indication regarding the growth of broilers opening a new scenario in monitoring systems in broiler houses.
Collapse
Affiliation(s)
- Ilaria Fontana
- Department of Environmental Science and Policy, Università degli Studi di Milano, via Celoria 10, 20133, Milan, Italy
| | - Emanuela Tullo
- Department of Environmental Science and Policy, Università degli Studi di Milano, via Celoria 10, 20133, Milan, Italy
| | - Lenn Carpentier
- Department of Biosystems, Division Animal and Human Health Engineering, M3-BIORES, Katholieke Universiteit Leuven, Kasteelpark Arenberg 30, bus 2456, 3001 Leuven, Belgium
| | | | - Andy Butterworth
- Department of Clinical Veterinary Science, University of Bristol, Langford, BS40 5DU, North Somerset UK
| | - Erik Vranken
- Department of Biosystems, Division Animal and Human Health Engineering, M3-BIORES, Katholieke Universiteit Leuven, Kasteelpark Arenberg 30, bus 2456, 3001 Leuven, Belgium.,Fancom BV, Wilhelminastraat 17, 5981 XW Panningen, The Netherlands
| | - Tomas Norton
- Department of Biosystems, Division Animal and Human Health Engineering, M3-BIORES, Katholieke Universiteit Leuven, Kasteelpark Arenberg 30, bus 2456, 3001 Leuven, Belgium
| | - Daniel Berckmans
- Department of Biosystems, Division Animal and Human Health Engineering, M3-BIORES, Katholieke Universiteit Leuven, Kasteelpark Arenberg 30, bus 2456, 3001 Leuven, Belgium
| | - Marcella Guarino
- Department of Environmental Science and Policy, Università degli Studi di Milano, via Celoria 10, 20133, Milan, Italy
| |
Collapse
|
42
|
Priyadarshani N, Castro I, Marsland S. The impact of environmental factors in birdsong acquisition using automated recorders. Ecol Evol 2018; 8:5016-5033. [PMID: 29876078 PMCID: PMC5980359 DOI: 10.1002/ece3.3889] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2017] [Revised: 09/24/2017] [Accepted: 12/11/2017] [Indexed: 11/08/2022] Open
Abstract
The use of automatic acoustic recorders is becoming a principal method to survey birds in their natural habitats, as it is relatively noninvasive while still being informative. As with any other sound, birdsong degrades in amplitude, frequency, and temporal structure as it propagates to the recorder through the environment. Knowing how different birdsongs attenuate under different conditions is useful to, for example, develop protocols for deploying acoustic recorders and improve automated detection methods, an essential part of the research field that is becoming known as ecoacoustics. This article presents playback and recapture (record) experiments carried out under different environmental conditions using twenty bird calls from eleven New Zealand bird species in a native forest and an open area, answering five research questions: (1) How does birdsong attenuation differ between forest and open space? (2) What is the relationship between transmission height and birdsong attenuation? (3) How does frequency of birdsong impact the degradation of sound with distance? (4) Is birdsong attenuation different during the night compared to the day? and (5) what is the impact of wind on attenuation? Bird calls are complex sounds; therefore, we have chosen to use them rather than simple tones to ensure that this complexity is not missed in the analysis. The results demonstrate that birdsong transmission was significantly better in the forest than in the open site. During the night, the attenuation was at a minimum in both experimental sites. Transmission height affected the propagation of the songs of many species, particularly the flightless ones. The effect of wind was severe in the open site and attenuated lower frequencies. The reverberations due to reflective surfaces masked higher frequencies (8 kHz) in the forest even at moderate distances. The findings presented here can be applied to develop protocols for passive acoustic monitoring. Even though the attenuation can be generalized to frequency bands, the structure of the birdsong is also important. Selecting a reasonable sampling frequency avoids unnecessary data accumulation because higher frequencies attenuate more in the forest. Even at moderate distances, recorders capture significantly attenuated birdsong, and hence, automated analysis methods for field recordings need to be able to detect and recognize faint birdsong.
Collapse
Affiliation(s)
- Nirosha Priyadarshani
- School of Engineering and Advanced TechnologyMassey UniversityPalmerston NorthNew Zealand
| | - Isabel Castro
- Wildlife and Ecology GroupMassey UniversityPalmerston NorthNew Zealand
| | - Stephen Marsland
- School of Mathematics and StatisticsVictoria University of WellingtonWellingtonNew Zealand
| |
Collapse
|
43
|
Retamosa Izaguirre MI, Ramírez-Alán O. Acoustic indices applied to biodiversity monitoring in a Costa Rica dry tropical forest. ACTA ACUST UNITED AC 2018. [DOI: 10.22261/jea.tnw2np] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
Abstract
Standardized methods for biodiversity monitoring are needed to evaluate conservation efforts. Acoustic indices are used in biodiversity assessments, but need to be compared to traditional wildlife methods. This work was conducted in the Santa Rosa National Park between June and November, 2015. We installed recorders and conducted bird point counts in twelve sampling sites. We compared acoustic indices (Acoustic Evenness Index [AEI], Acoustic Diversity Index [ADI], Acoustic Complexity Index [ACI], Bioacoustic Index [BIO], Normalized Difference Soundscape Index [NDSI], Total Entropy [TE], Median Amplitude Envelope [MAE], Number of peaks [NP]) with indices from bird point counts (Bird Abundance, Bird Richness, Bird Diversity and Bird Evenness), and discuss the utility of acoustic indices as indicators for biodiversity monitoring in tropical forests. ADI, ACI, BIO and TE presented a similar temporal pattern peaking between 5 am and 6 am; and an additional peak at 5 pm, except for ACI. These patterns were consistent with the daily biological rhythms. AEI, ACI, BIO and Bird Abundance were related to characteristics of younger forests (lower percentage of canopy cover) but NP, ADI, TE, Bird Diversity and Bird Evenness were related to characteristics of older forests (higher percentage of canopy cover and a lower number of patches). ACI was positively correlated to Bird Abundance and NP was positively correlated to Bird Diversity. ACI reflects biological activity, but not necessarily a more diverse bird community in this study area. This might be an indication of a strong acoustic competition, or several highly dominant bird species in younger forests. Furthermore, acoustic communities in tropical forests commonly include insects (cicadas) and frogs, which might affect resulting acoustic indices. A variety of methods are probably needed to thoroughly assess biodiversity. However, a combination of indices such as ACI and NP might be considered to monitor trends in abundance and diversity of birds in dry forests.
Collapse
|
44
|
Papin M, Pichenot J, Guérold F, Germain E. Acoustic localization at large scales: a promising method for grey wolf monitoring. Front Zool 2018; 15:11. [PMID: 29681989 PMCID: PMC5897954 DOI: 10.1186/s12983-018-0260-2] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2017] [Accepted: 03/25/2018] [Indexed: 11/15/2022] Open
Abstract
Background The grey wolf (Canis lupus) is naturally recolonizing its former habitats in Europe where it was extirpated during the previous two centuries. The management of this protected species is often controversial and its monitoring is a challenge for conservation purposes. However, this elusive carnivore can disperse over long distances in various natural contexts, making its monitoring difficult. Moreover, methods used for collecting signs of presence are usually time-consuming and/or costly. Currently, new acoustic recording tools are contributing to the development of passive acoustic methods as alternative approaches for detecting, monitoring, or identifying species that produce sounds in nature, such as the grey wolf. In the present study, we conducted field experiments to investigate the possibility of using a low-density microphone array to localize wolves at a large scale in two contrasting natural environments in north-eastern France. For scientific and social reasons, the experiments were based on a synthetic sound with similar acoustic properties to howls. This sound was broadcast at several sites. Then, localization estimates and the accuracy were calculated. Finally, linear mixed-effects models were used to identify the factors that influenced the localization accuracy. Results Among 354 nocturnal broadcasts in total, 269 were recorded by at least one autonomous recorder, thereby demonstrating the potential of this tool. Besides, 59 broadcasts were recorded by at least four microphones and used for acoustic localization. The broadcast sites were localized with an overall mean accuracy of 315 ± 617 (standard deviation) m. After setting a threshold for the temporal error value associated with the estimated coordinates, some unreliable values were excluded and the mean accuracy decreased to 167 ± 308 m. The number of broadcasts recorded was higher in the lowland environment, but the localization accuracy was similar in both environments, although it varied significantly among different nights in each study area. Conclusions Our results confirm the potential of using acoustic methods to localize wolves with high accuracy, in different natural environments and at large spatial scales. Passive acoustic methods are suitable for monitoring the dynamics of grey wolf recolonization and so, will contribute to enhance conservation and management plans.
Collapse
Affiliation(s)
- Morgane Papin
- Centre de Recherche et d'Observation sur les Carnivores (CROC), 4 rue de la banie, 57590, Lucy, France.,2Université de Lorraine, CNRS, Laboratoire Interdisciplinaire des Environnements Continentaux, F-57500 Metz, France
| | - Julian Pichenot
- Biologiste Écologue Consultant (BEC), 8A rue principale, 57590, Fonteny, France
| | - François Guérold
- 2Université de Lorraine, CNRS, Laboratoire Interdisciplinaire des Environnements Continentaux, F-57500 Metz, France
| | - Estelle Germain
- Centre de Recherche et d'Observation sur les Carnivores (CROC), 4 rue de la banie, 57590, Lucy, France
| |
Collapse
|
45
|
Stiffler LL, Anderson JT, Katzner TE. Evaluating autonomous acoustic surveying techniques for rails in tidal marshes. WILDLIFE SOC B 2018. [DOI: 10.1002/wsb.860] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Lydia L. Stiffler
- Division of Forestry and Natural Resources; P.O. Box 6125, West Virginia University Morgantown WV 26506 USA
| | - James T. Anderson
- Division of Forestry and Natural Resources; P.O. Box 6125, West Virginia University Morgantown WV 26506 USA
| | - Todd E. Katzner
- U.S. Geological Survey, Forest & Rangeland Ecosystem Science Center; 970 Lusk Street Boise ID 83706 USA
| |
Collapse
|
46
|
|
47
|
Chambert T, Waddle JH, Miller DAW, Walls SC, Nichols JD. A new framework for analysing automated acoustic species detection data: Occupancy estimation and optimization of recordings post‐processing. Methods Ecol Evol 2017. [DOI: 10.1111/2041-210x.12910] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Thierry Chambert
- Department of Ecosystem Science and ManagementPennsylvania State University University Park PA USA
- Patuxent Wildlife Research CenterU.S. Geological Survey Laurel MD USA
| | - J. Hardin Waddle
- Wetland and Aquatic Research CenterU.S. Geological Survey Lafayette LA USA
| | - David A. W. Miller
- Department of Ecosystem Science and ManagementPennsylvania State University University Park PA USA
| | - Susan C. Walls
- Wetland and Aquatic Research CenterU.S. Geological Survey Gainesville FL USA
| | - James D. Nichols
- Patuxent Wildlife Research CenterU.S. Geological Survey Laurel MD USA
| |
Collapse
|
48
|
Nguyen DT, Ogunbona PO, Li W, Tasker E, Yearwood J. Detection of ground parrot vocalisation: A multiple instance learning approach. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:1281. [PMID: 28964088 DOI: 10.1121/1.4999318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Ground parrot vocalisation can be considered as an audio event. Test-based diverse density multiple instance learning (TB-DD-MIL) is proposed for detecting this event in audio files recorded in the field. The proposed method is motivated by the advantages of multiple instance learning from incomplete training data. Spectral features suitable for encoding the vocal source information of the ground parrot vocalization are also investigated. The proposed method was benchmarked against a dataset collected in various environmental conditions and an audio detection evaluation scheme is proposed. The evaluation includes a study on performance of the various vocal source features and comparison with other classification techniques. Experimental results indicated that the most appropriate feature to encode ground parrot calls is the spectral bandwidth and the proposed TB-DD-MIL method outperformed other existing classification methods.
Collapse
Affiliation(s)
- Duc Thanh Nguyen
- Deakin University, School of Information Technology, Burwood, Australia
| | - Philip O Ogunbona
- University of Wollongong, School of Computing and Information Technology, Wollongong, Australia
| | - Wanqing Li
- University of Wollongong, School of Computing and Information Technology, Wollongong, Australia
| | | | - John Yearwood
- Deakin University, School of Information Technology, Burwood, Australia
| |
Collapse
|
49
|
An FPGA-Based WASN for Remote Real-Time Monitoring of Endangered Species: A Case Study on the Birdsong Recognition of Botaurus stellaris. SENSORS 2017; 17:s17061331. [PMID: 28594373 PMCID: PMC5492858 DOI: 10.3390/s17061331] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/31/2017] [Revised: 06/01/2017] [Accepted: 06/06/2017] [Indexed: 11/19/2022]
Abstract
Fast environmental variations due to climate change can cause mass decline or even extinctions of species, having a dramatic impact on the future of biodiversity. During the last decade, different approaches have been proposed to track and monitor endangered species, generally based on costly semi-automatic systems that require human supervision adding limitations in coverage and time. However, the recent emergence of Wireless Acoustic Sensor Networks (WASN) has allowed non-intrusive remote monitoring of endangered species in real time through the automatic identification of the sound they emit. In this work, an FPGA-based WASN centralized architecture is proposed and validated on a simulated operation environment. The feasibility of the architecture is evaluated in a case study designed to detect the threatened Botaurus stellaris among other 19 cohabiting birds species in The Parc Natural dels Aiguamolls de l’Empordà, showing an averaged recognition accuracy of 91% over 2h 55’ of representative data. The FPGA-based feature extraction implementation allows the system to process data from 30 acoustic sensors in real time with an affordable cost. Finally, several open questions derived from this research are discussed to be considered for future works.
Collapse
|
50
|
Zhao Z, Zhang SH, Xu ZY, Bellisario K, Dai NH, Omrani H, Pijanowski BC. Automated bird acoustic event detection and robust species classification. ECOL INFORM 2017. [DOI: 10.1016/j.ecoinf.2017.04.003] [Citation(s) in RCA: 57] [Impact Index Per Article: 8.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|