Dataset Open Access

# To bee or not to bee: An annotated dataset for beehive sound recognition

Nolasco, Inês; Benetos, Emmanouil

### DataCite XML Export

<?xml version='1.0' encoding='utf-8'?>
<identifier identifierType="DOI">10.5281/zenodo.1321278</identifier>
<creators>
<creator>
<creatorName>Nolasco, Inês</creatorName>
<givenName>Inês</givenName>
<familyName>Nolasco</familyName>
<affiliation>Queen Mary University of London</affiliation>
</creator>
<creator>
<creatorName>Benetos, Emmanouil</creatorName>
<givenName>Emmanouil</givenName>
<familyName>Benetos</familyName>
<nameIdentifier nameIdentifierScheme="ORCID" schemeURI="http://orcid.org/">0000-0002-6820-6764</nameIdentifier>
<affiliation>Queen Mary University of London</affiliation>
</creator>
</creators>
<titles>
<title>To bee or not to bee: An annotated dataset for beehive sound recognition</title>
</titles>
<publisher>Zenodo</publisher>
<publicationYear>2018</publicationYear>
<dates>
<date dateType="Issued">2018-07-25</date>
</dates>
<resourceType resourceTypeGeneral="Dataset"/>
<alternateIdentifiers>
<alternateIdentifier alternateIdentifierType="url">https://zenodo.org/record/1321278</alternateIdentifier>
</alternateIdentifiers>
<relatedIdentifiers>
<relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.5281/zenodo.1321277</relatedIdentifier>
<relatedIdentifier relatedIdentifierType="URL" relationType="IsPartOf">https://zenodo.org/communities/opensourcebeehives</relatedIdentifier>
</relatedIdentifiers>
<rightsList>
<rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights>
</rightsList>
<descriptions>
<description descriptionType="Abstract">&lt;p&gt;&lt;strong&gt;-- Dataset documentation --&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;br&gt;
&lt;strong&gt;1- Introduction&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The present dataset was developed in the context of our work in [1] that focus on the automatic recognition of beehive sounds. The problem is posed as the classification of sound segments in two classes: Bee and noBee. The novelty of the explored approach and the need for annotated data, dictated the construction of such dataset.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2- Description&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.1- Audio recordings:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The annotated dataset was developed based on a selected set of recordings acquired in the context of two different projects: the Open Source Beehive (OSBH) project [2] and the NU-Hive project [3]. Both projects main goal is to develop a beehive monitoring system capable of identifying and predict certain events and states of the hive that are of interest to the beekeeper. Among many different variables that can be measured and that help the recognition of different states of the hive, the analysis and use of the sound the bees produce is a big focus for both projects.&lt;/p&gt;

&lt;p&gt;The recordings from the OSBH project were acquired through a citizen science initiative which asked people from the general public to record the sound from their beehives together with the registering of the hive state at the moment. Because of the amateur and collaborative nature of this project, the recordings from the OSBH project present great diversity due to the very different conditions in which the signals were acquired: different recording devices used, different environments where the hives were placed, and even different position for the microphones inside the hive. This variety of settings makes this dataset a very interesting tool to help evaluate and challenge the methods developed.&lt;/p&gt;

&lt;p&gt;The NU-Hive project is a comprehensive effort of data acquisition, concerning not only sound, but a vast amount of variables that will allow the study of bees behaviors and other unknown aspects. The selected recordings are taken from 2 hives and labeled regarding two states: queen bee is present, and queen bee not present. Contrary to the OSBH project recordings, the recordings from&amp;nbsp;the NU-Hive project are from a much more controlled and homogeneous environment. Here the occurring external sounds are mainly traffic, car honks and birds.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The annotated dataset:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;For each selected recording, time segments are labeled as Bee or noBee depending on the perceived source of the sound signal being from bees or external to the hive.&lt;/p&gt;

&lt;p&gt;The whole annotated dataset consists of 78 recordings of varying lengths which make up for a total duration of approximately 12 hours of which 25% is annotated as noBee events.&lt;/p&gt;

&lt;p&gt;About 60% of the recordings are from the NU-Hive dataset and represent 2 hives, the remaining are recordings from the OSBH dataset and 6 different hives. The recorded hives are from 3 main locations: North America, Australia and Europe.&lt;/p&gt;

&lt;p&gt;&amp;nbsp;&lt;/p&gt;

&lt;p&gt;The annotation procedure consists in hearing the selected recordings and marking the beginning and the end of every sound that could not be recognized as a beehive sound. The recognition of external sounds is based primarily on the perceived heard sounds, but a visual aid is also used by visualizing the log-mel frequency spectrum of the signal. All the above are functionalities offered by the Sonic Visualiser software, which was used by two volunteers that are neither bee-specialists nor specially trained in sound annotation tasks.&lt;/p&gt;

&lt;p&gt;By marking these pairs of moments corresponding to the beginning and end of external sound periods, we are able to get the whole recording labeled into&amp;nbsp;Bee and noBee intervals. Thus in the resulting Bee intervals only pure beehive sounds, (no external sounds) should be perceived for the entirety of the segment. The noBee intervals refer to periods where an external sound can be perceived (superimposed to the bee sounds).&lt;/p&gt;

&lt;p&gt;&amp;nbsp;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;File Structure:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Each audio file is coupled with its corresponding annotation file, identified by the same name and extension &lt;em&gt;.lab&lt;/em&gt;.&lt;br&gt;
For convenience, all the annotations are collected in a single master label file named &lt;em&gt;beeAnnotations.mlf&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The &lt;em&gt;.lab&lt;/em&gt;&amp;nbsp;files consist of :&amp;nbsp;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;First row identifies the audio file to which the annotations refer to.&lt;/li&gt;
&lt;li&gt;Each line after that describes an interval with starting time point, end time point and label. The time points are expressed in seconds.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Below is an example of such an annotation file:&amp;nbsp;&lt;/p&gt;

&lt;pre&gt;&lt;code&gt;Hive3_20_07_2017_QueenBee_H3_audio_15_30_00
0 78.45 bee
78.46 78.95 nobee
78.96 103.92 bee
103.93 112.48 nobee
112.49 152.48 bee
.
&lt;/code&gt;&lt;/pre&gt;

When using this dataset, please cite [1]:&lt;/p&gt;

&lt;p&gt;[1] I. Nolasco and E. Benetos, &amp;ldquo;To bee or not to bee: Investigating machine learning approaches to beehive sound recognition,&amp;rdquo; in Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE), 2018, submitted.&lt;/p&gt;

&lt;p&gt;[2] &amp;ldquo;Open Source Beehives Project,&amp;rdquo; https://www.osbeehives.com/.&lt;/p&gt;

&lt;p&gt;[3] S. Cecchi, A. Terenzi, S. Orcioni, P. Riolo, S. Ruschioni, and N. Isidoro, &amp;ldquo;A preliminary study of sounds emitted by honey bees in a beehive,&amp;rdquo; in Audio Engineering Society Convention 144, 2018.&lt;/p&gt;</description>
</descriptions>
</resource>

548
3,942
views