Animal acoustic communication research depends on our ability to record the vocal behaviour of different species. Only rarely do we have the opportunity to continuously follow the vocal behaviour of a group of individuals of the same species for a long period of time. Here, we provide a database of Egyptian fruit bat vocalizations, which were continuously recorded in the lab in several groups simultaneously for more than a year. The dataset includes almost 300,000 files, a few seconds each, containing social vocalizations and representing the complete vocal repertoire used by the bats in the experiment period. Around 90,000 files are annotated with details about the individuals involved in the vocal interactions, their behaviours and the context. Moreover, the data include the complete vocal ontogeny of pups, from birth to adulthood, in different conditions (e.g., isolated or in a group). We hope that this comprehensive database will stimulate studies that will enhance our understanding of bat, and mammal, social vocal communication.
Background & Summary
Comparative research of nonhuman animals can potentially shed light on the evolution of language and speech1. For instance, the study of animal vocal communication may reveal the roots of syntax and semantics2,3. Nonhuman vocalizations are often cryptic to a human observer, and with little prior knowledge about the animal-relevant acoustics, identifying essential information in them becomes an arduous task. Still, the discovery of nuances, which may be subtle to the human ear but important to the communicating animal, may become plausible if facilitated by large recording datasets. Here we present an extremely large collection of vocalizations of Egyptian fruit bats (Rousettus aegyptiacus). Moreover, many of the vocalizations in this database are accompanied by relevant information such as the identities of the emitter and the addressee of the vocalization, and the related behavioural context. Bats are social mammals which use rich vocal communication4,
Animal retrieval and care
All adult bats (Rousettus aegyptiacus), females in their late pregnancy and males, were caught in a natural roost near Herzliya, Israel. This roost is inhabited by a colony of 5,000 to 10,000 bats. All recorded pups were born in the experimental setup to wild-caught females. The bats were kept in acoustic chambers, large enough to allow flight (Fig. 1), and fed with a variety of fruits. Pups were separated from their mothers, and joined together (if were previously isolated; see below), after all pups were observed feeding on fruit by themselves. All experiments were reviewed and approved by the Institutional Animal Care and Use Committee of Tel Aviv University (Number L-13-016). The use of bats was approved by the Israeli National Park Authority.
Two types of chambers were used to house the bats: colony chambers (Fig. 1a) for most of the recordings, and (smaller) isolation chambers (Fig. 1b) for the recording of preweaned isolated pups. The chambers were acoustically isolated (see ref. 13 for isolation verification methods) and their walls were covered with foam to diminish echoes. The chambers were continuously monitored with IR-sensitive cameras and omnidirectional electret ultrasound microphones (Avisoft-Bioacoustics Knowles FG-O). Audio recordings were conducted using Avisoft-Bioacoustics UltraSoundGate 1216H A/D converter with a sampling rate of 250 kHz.
Two types of treatments are included in our data: colony and isolation (Table 1). In a colony treatment adult bats were housed together, usually a few females and one male, in a colony chamber (Fig. 1a), and pups were born to the females in this chamber. In the isolation treatment each pregnant female was housed alone in a private isolation chamber (Fig. 1b), and gave birth to one pup in this chamber. After weaning, pups of both treatments were housed in colony chambers without adults. The recordings were conducted from May 2012 to June 2013, and in February 2014, in chambers of different treatments in parallel, where each chamber was continuously recorded (see Table 1 for recording periods of each group, and Table 2 for treatment assignment of individual bats). Importantly, we include in this database all of the recordings in which our automatic tools identified social calls. Thus, the database can be regarded as practically complete representation of the social vocal communication used by the bats during the recording periods, and statistics about the total usage of vocalizations can be safely drawn. Recordings may only be missing on rare occasions, due to short periods of technical problems, such as power cuts or malfunctioning microphone replacements.
293,238 recorded audio files (WAV format, sampling rate: 250 kHz, depth: 16 bit). The files are compressed (and can be extracted) using 7-zip (www.7-zip.org).
One annotation file: Annotations.csv, with 91,080 annotations. These annotations were obtained from the videos (see below) and include details such as the emitter and context of each vocalization. The content of each column in the annotation file is described in Table 3, and descriptions of contexts and behaviours are depicted in Table 4. Each annotation corresponds to sequences of vocalizations in one file. Most files include a single interaction and, correspondingly, a single annotation, though some files record several interactions (and may be annotated with several annotations). Accordingly, columns 9 and 10 in Annotations.csv specify the location in the file to which each annotation refers (see Table 3).
One file describing the audio files: FileInfo.csv, which includes the exact recording time, the recording channel, and the exact time of the voiced segments in each file.
A set of audio example files.
Example videos of different interactions.
A folder with example raw video recordings.
A sample Matlab code exemplifying the segmentation and noise-filtering of raw audio recording. A similar process was used for obtaining the start and end positions of voiced segments (given in FileInfo.csv), and for filtering voiceless files; parameter adjustment might be required for specific tasks.
The recorded audio files are divided into folders of no more than 10,000 files for the convenience of use (this division has no significance). Note that two non-annotated files might be different recordings of the same call, if they were recorded at the same time, in the same treatment, in different channels, though it is not necessarily the case, and this is never the case for annotated files. Such duplicates can be excluded by the user by inspecting the recordings themselves. The annotation and metadata files are in comma-separated-value format (CSV) to ease their use with automatic tools, and to allow their direct upload into spreadsheet software. The metadata file includes descriptions of all identifiers in the annotation file. The example files include a few audio files exemplifying different recorded sounds, to facilitate the familiarity of the user with the recorded data. These examples include social call syllables, isolation calls of young pups, echolocation clicks, and examples of background noises (e.g., cage noise, microphone direct hit, etc., which are all rare in this database).
The annotation types (contexts or behaviours) were defined by YP and MT. The recordings were annotated using the videos by MT and EP, or by trained students. These observers were certified after annotating a few recording days, which were then validated by an expert (YP or MT). In annotating the recordings we adopted a conservative approach, in which we designated as ‘unknown’ any type of data for which we had any doubt. Despite the training of the observers, some noise might have been introduced during the hundreds of hours of manual annotations, thus we estimated an error rate by a post-hoc quality test: 435 annotated recordings were sampled randomly and were then carefully re-annotated by EP, MT, and YP. Errors were counted when there were either a discrepancy between the post-hoc and the original annotations, or when the post-hoc examination concluded that some doubt still exists. The error rates were 2.1% (95% Confidence-Interval [CI]: 0.8–3.4%) for the emitter identification, 2.1% (95% CI: 0.8–3.4%) for the addressee identification, and 4% (95% CI: 2.2–5.8%) for the context identification. Thus we estimate the accuracy of the annotations as 97.9% for the emitter and addressee, and 96% for the context.
The FileInfo.csv file includes automatically generated start and end positions of voiced segments (in samples) for each file. This enables an immediate analysis of the data without the need to apply any pre-processing. However, we encourage users to verify the suitability of the automatic segmentation (given in FileInfo.csv) to their needs by reviewing it in a random sample of recordings. To facilitate such review, and familiarity with the database, we provide a small library of examples of different sounds which might be encountered in the raw recordings, including social calls (which are the core interest of this study, and the most common sounds in the database), echolocation clicks (which are sporadically recorded before or after social calls), cage noises (relatively rare), and pup isolation calls (which are distinct from adults calls). For a usage which is sensitive to possible differences between microphones, one may take advantage of the Recording channel (different channels represent different microphones) field in FileInfo.csv (note that some microphones might have been replaced during the experiment, although these replacements were rare).
How to cite this article: Prat, Y. et al. An annotated dataset of Egyptian fruit bat vocalizations across varying contexts and during vocal ontogeny. Sci. Data 4:170143 doi: 10.1038/sdata.2017.143 (2017).
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Prat, Y., Taub, M., Pratt, E., & Yovel, Y. Figshare https://doi.org/10.6084/m9.figshare.c.3666502 (2017)
This study was partially supported by a research grant to Y.Y. from the European Research Council (ERC – GPSBAT).
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ The Creative Commons Public Domain Dedication waiver http://creativecommons.org/publicdomain/zero/1.0/ applies to the metadata files made available in this article.