• Home
  • About
  • Repositories
  • Search
  • Web API
  • Feedback
<< Go Back

Metadata

Name
FSDKaggle2019
Repository
ZENODO
Identifier
doi:10.5281/zenodo.3612637
Description
FSDKaggle2019 is an audio dataset containing 29,266 audio files annotated with 80 labels of the AudioSet Ontology. FSDKaggle2019 has been used for the DCASE Challenge 2019 Task 2,&nbsp; which was run as a Kaggle competition titled&nbsp;Freesound Audio Tagging 2019.

Citation

If you use the FSDKaggle2019 dataset or part of it, please cite our DCASE 2019 paper:


Eduardo Fonseca, Manoj Plakal, Frederic Font, Daniel P. W. Ellis, Xavier Serra. &quot;Audio tagging with noisy labels and minimal supervision&quot;. Proceedings of the DCASE 2019 Workshop, NYC, US (2019)


You can also consider citing our ISMIR 2017 paper, which describes how we gathered the manual annotations included in FSDKaggle2019.


Eduardo Fonseca, Jordi Pons, Xavier Favory, Frederic Font, Dmitry Bogdanov, Andres Ferraro, Sergio Oramas, Alastair Porter, and Xavier Serra, &quot;Freesound Datasets: A Platform for the Creation of Open Audio Datasets&quot;, In Proceedings of the 18th International Society for Music Information Retrieval Conference, Suzhou, China, 2017


Data curators

Eduardo Fonseca, Manoj Plakal, Xavier Favory, Jordi Pons

Contact

You are welcome to contact Eduardo Fonseca should you have any questions at eduardo.fonseca@upf.edu.

&nbsp;

ABOUT FSDKaggle2019

Freesound Dataset Kaggle 2019 (or FSDKaggle2019 for short) is an audio dataset containing 29,266 audio files annotated with 80 labels of the AudioSet Ontology [1]. FSDKaggle2019 has been used for the Task 2 of the Detection and Classification of Acoustic Scenes and Events (DCASE) Challenge 2019. Please visit the DCASE2019 Challenge Task 2 website for more information. This Task was hosted on the Kaggle platform as a competition titled Freesound Audio Tagging 2019. It was organized by researchers from the Music Technology Group (MTG) of Universitat Pompeu Fabra (UPF), and from Sound Understanding team at Google AI Perception. The competition intended to provide insight towards the development of broadly-applicable sound event classifiers able to cope with label noise and minimal supervision conditions.

FSDKaggle2019 employs audio clips from the following sources:


&nbsp;Freesound Dataset (FSD): a dataset being collected at the MTG-UPF based on Freesound content organized with the AudioSet Ontology
&nbsp;The soundtracks of a pool of Flickr videos taken from the Yahoo Flickr Creative Commons 100M dataset (YFCC)


The audio data is labeled using a vocabulary of 80 labels from Google&rsquo;s AudioSet Ontology [1], covering diverse topics: Guitar and other Musical Instruments, Percussion, Water, Digestive, Respiratory sounds, Human voice, Human locomotion, Hands, Human group actions, Insect, Domestic animals, Glass, Liquid, Motor vehicle (road), Mechanisms, Doors, and a variety of Domestic sounds. The full list of categories can be inspected in&nbsp;vocabulary.csv (see Files &amp; Download below). The goal of the task was to build a multi-label audio tagging system that can predict appropriate label(s) for each audio clip in a test set.

What follows is a summary of some of the most relevant characteristics of FSDKaggle2019. Nevertheless, it is highly recommended to read our DCASE 2019 paper for a more in-depth description of the dataset and how it was built.

Ground Truth Labels

The ground truth labels are provided at the clip-level, and express the presence of a sound category in the audio clip, hence can be considered&nbsp;weak&nbsp;labels or tags. Audio clips have variable lengths (roughly from 0.3 to 30s).

The audio content from&nbsp;FSD&nbsp;has been manually labeled by humans following a data labeling process using the&nbsp;Freesound Annotator&nbsp;platform. Most labels have inter-annotator agreement but not all of them. More details about the data labeling process and the&nbsp;Freesound Annotator&nbsp;can be found in [2].

The&nbsp;YFCC&nbsp;soundtracks were labeled using automated heuristics applied to the audio content and metadata of the original Flickr clips. Hence, a substantial amount of label noise can be expected. The label noise can vary widely in amount and type depending on the category, including in- and out-of-vocabulary noises. More information about some of the types of label noise that can be encountered is available in [3].

Specifically, FSDKaggle2019 features&nbsp;three types of label quality, one for each set in the dataset:


curated train set: correct (but potentially incomplete) labels
noisy train set: noisy labels
test set: correct and complete labels


Further details can be found below in the sections for each set.

Format

All audio clips are provided as uncompressed PCM 16 bit, 44.1 kHz, mono audio files.

&nbsp;

DATA SPLIT

FSDKaggle2019 consists of&nbsp;two train sets&nbsp;and&nbsp;one test set. The idea is to limit the supervision provided for training (i.e., the manually-labeled, hence reliable, data), thus promoting approaches to deal with label noise.

Curated train set

The&nbsp;curated train set&nbsp;consists of manually-labeled data from&nbsp;FSD.&nbsp;


Number of clips/class: 75 except in a few cases (where there are less)
Total number of clips: 4970
Avg number of labels/clip: 1.2
Total duration: 10.5 hours


The duration of the audio clips ranges from 0.3 to 30s due to the diversity of the sound categories and the preferences of Freesound users when recording/uploading sounds. Labels are correct but potentially incomplete. It can happen that a few of these audio clips present additional acoustic material beyond the provided ground truth label(s).

Noisy train set

The&nbsp;noisy train set&nbsp;is a larger set of noisy web audio data from Flickr videos taken from the&nbsp;YFCC&nbsp;dataset [5].


Number of clips/class: 300
Total number of clips: 19,815
Avg number of labels/clip: 1.2
Total duration: ~80 hours


The duration of the audio clips ranges from 1s to 15s, with the vast majority lasting 15s. Labels are automatically generated and purposefully noisy. No human validation is involved. The label noise can vary widely in amount and type depending on the category, including in- and out-of-vocabulary noises.

Considering the numbers above, the per-class data distribution available for training is, for most of the classes, 300 clips from the noisy train set and 75 clips from the curated train set. This means 80% noisy / 20% curated at the clip level, while at the duration level the proportion is more extreme considering the variable-length clips.

Test set

The&nbsp;test set&nbsp;is used for system evaluation and consists of manually-labeled data from&nbsp;FSD.&nbsp;


Number of clips/class: between 50 and 150
Total number of clips: 4481
Avg number of labels/clip: 1.4
Total duration: 12.9 hours


The acoustic material present in the test set clips is labeled exhaustively using the aforementioned vocabulary of 80 classes. Most labels have inter-annotator agreement but not all of them. Except human error, the label(s) are correct and complete considering the target vocabulary; nonetheless, a few clips could still present additional (unlabeled) acoustic content out of the vocabulary.

During the&nbsp;DCASE2019 Challenge Task 2, the test set was split into two subsets, for the&nbsp;public&nbsp;and&nbsp;private&nbsp;leaderboards, and only the data corresponding to the&nbsp;public&nbsp;leaderboard was provided.&nbsp;In this current package you will find the full test set with all the test labels. To allow comparison with previous work, the file&nbsp;test_post_competition.csv&nbsp;includes a flag to determine the corresponding leaderboard (public or private) for each test clip (see more info in&nbsp;Files &amp; Download&nbsp;below).

Acoustic mismatch

As mentioned before, FSDKaggle2019 uses audio clips from two sources:


FSD: curated train set and test set, and
YFCC: noisy train set.


While the sources of audio (Freesound and Flickr) are collaboratively contributed and pretty diverse themselves, a certain acoustic mismatch can be expected between&nbsp;FSD&nbsp;and&nbsp;YFCC. We conjecture this mismatch comes from a variety of reasons.
For example, through acoustic inspection of a small sample of both data sources, we find a higher percentage of high quality recordings in FSD. In addition, audio clips in Freesound are typically recorded with the purpose of capturing audio, which is not necessarily the case in YFCC.

This mismatch can have an impact in the evaluation, considering that most of the train data come from YFCC, while all test data are drawn from FSD. This constraint (i.e., noisy training data coming from a different web audio source than the test set) is sometimes a real-world condition.

&nbsp;

LICENSE

All clips in FSDKaggle2019 are released under Creative Commons (CC) licenses. For attribution purposes and to facilitate attribution of these files to third parties, we include a mapping from the audio clips to their corresponding licenses.



Curated train set and test set. All clips in Freesound are released under different modalities of Creative Commons (CC) licenses, and each audio clip has its own license as defined by the audio clip uploader in Freesound, some of them requiring attribution to their original authors and some forbidding further commercial reuse. The licenses are specified in the files&nbsp;train_curated_post_competition.csv&nbsp;and&nbsp;test_post_competition.csv. These licenses can be CC0, CC-BY, CC-BY-NC and CC Sampling+.


Noisy train set. Similarly, the licenses of the soundtracks from&nbsp;Flickr&nbsp;used in FSDKaggle2019 are specified in the file&nbsp;train_noisy_post_competition.csv. These licenses can be CC-BY and CC BY-SA.



In addition, FSDKaggle2019 as a whole is the result of a curation process and it has an additional license. FSDKaggle2019 is released under&nbsp;CC-BY. This license is specified in the&nbsp;LICENSE-DATASET&nbsp;file downloaded with the&nbsp;FSDKaggle2019.doc zip file.

&nbsp;

FILES &amp; DOWNLOAD

FSDKaggle2019 can be downloaded as a series of zip files with the following directory structure:

root
?
????FSDKaggle2019.audio_train_curated/ Audio clips in the curated train set
?
????FSDKaggle2019.audio_train_noisy/ Audio clips in the noisy train set
?
????FSDKaggle2019.audio_test/ Audio clips in the full test set
?
????FSDKaggle2019.meta/ Files for evaluation setup
? ?
? ???? train_curated_post_competition.csv Ground truth for the curated train set
? ?
? ???? train_noisy_post_competition.csv Ground truth for the noisy train set
? ?
? ???? test_post_competition.csv Ground truth for the full test set
? ?
? ???? vocabulary.csv List of sound classes in FSDKaggle2019
?
????FSDKaggle2019.doc/
?
????README.md The dataset description file that you are reading
?
????LICENSE-DATASET License of the FSDKaggle2019 dataset as an entity


Important Note:&nbsp;the original&nbsp;train_curated.csv&nbsp;and&nbsp;train_noisy.csv&nbsp;files provided during the competition have been updated with more metadata (licenses, Freesound/Flickr ids, etc.) into&nbsp;train_curated_post_competition.csv&nbsp;and&nbsp;train_noisy_post_competition.csv. Likewise, the original&nbsp;test.csv&nbsp;that was not public during the competition is now available with ground truth and metadata as&nbsp;test_post_competition.csv.

Each row (i.e. audio clip) of the&nbsp;train_curated_post_competition.csv&nbsp;or&nbsp;train_noisy_post_competition.csv files contains the following information:


fname: the file name, e.g.,&nbsp;0006ae4e.wav
labels: the audio classification label(s) (ground truth). Note that the number of labels per clip can be one, eg,&nbsp;Bark&nbsp;or more, eg,&nbsp;&quot;Walk_and_footsteps,Slam&quot;.
freesound_id&nbsp;or&nbsp;flickr_video_URL: the Freesound id or Flickr id for the audio clip
license: the license for the audio clip


Each row (i.e. audio clip) of the&nbsp;test_post_competition.csv&nbsp;file contains the following information:


fname: the file name
labels: the audio classification label(s) (ground truth). Note that the number of labels per clip can be one, eg,&nbsp;Bark&nbsp;or more, eg,&nbsp;&quot;Walk_and_footsteps,Slam&quot;.
usage: string that indicates to which Kaggle leaderboard the clip was associated during the competition:&nbsp;Public&nbsp;or&nbsp;Private
freesound_id: the Freesound id for the audio clip
license: the license for the audio clip


Detected corrupted files in the curated train set

The following 5 audio files in the&nbsp;curated train set&nbsp;have a wrong label, due to a bug in the file renaming process:&nbsp;f76181c4.wav,&nbsp;77b925c2.wav,&nbsp;6a1f682a.wav,&nbsp;c7db12aa.wav,&nbsp;7752cc8a.wav.

The audio file&nbsp;1d44b0bd.wav&nbsp;in the&nbsp;curated train set&nbsp;was found to be corrupted (contains no signal) due to an error in format conversion.

If you find more corrupted files in FSDKaggle2019, please send an email to eduardo.fonseca@upf.edu.

Download

Each of the folders in the directory structure above is compressed into one corresponding zip file that you can download and unzip with your favorite compression tool. There is one exception: due to the large size of&nbsp;FSDKaggle2019.audio_train_noisy/, it is split into 7 files (note the last file is not&nbsp;*.z07, but&nbsp;*.zip):

FSDKaggle2019.audio_train_noisy.z01
FSDKaggle2019.audio_train_noisy.z02
FSDKaggle2019.audio_train_noisy.z03
FSDKaggle2019.audio_train_noisy.z04
FSDKaggle2019.audio_train_noisy.z05
FSDKaggle2019.audio_train_noisy.z06
FSDKaggle2019.audio_train_noisy.zip

In this case, you first have to download the 7 files. Once downloaded, we convert the split archive to a single-file archive. In other words, we merge the 7 files into one zip file called e.g.&nbsp;unsplit.zip&nbsp;in your local machine.

zip -s 0 FSDKaggle2019.audio_train_noisy.zip --out unsplit.zip

Finally, this merged file is unzipped.

unzip unsplit.zip

Baseline System

A CNN baseline system for FSDKaggle2019 is available at&nbsp;https://github.com/DCASE-REPO/dcase2019task2baseline.

&nbsp;

REFERENCES AND LINKS

[1] Jort F Gemmeke, Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R Channing Moore, Manoj Plakal, and Marvin Ritter. &quot;Audio set: An ontology and human-labeled dataset for audio events.&quot; In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, 2017. [PDF]

[2] Eduardo Fonseca, Jordi Pons, Xavier Favory, Frederic Font, Dmitry Bogdanov, Andres Ferraro, Sergio Oramas, Alastair Porter, and Xavier Serra. &quot;Freesound Datasets: A Platform for the Creation of Open Audio Datasets.&quot; In Proceedings of the International Conference on Music Information Retrieval, 2017. [PDF]

[3] Eduardo Fonseca, Manoj Plakal, Daniel P. W. Ellis, Frederic Font, Xavier Favory, and Xavier Serra. &quot;Learning Sound Event Classifiers from Web Audio with Noisy Labels.&quot; In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, 2019. [PDF]

[4] Frederic Font, Gerard Roma, and Xavier Serra. &quot;Freesound technical demo.&quot; Proceedings of the 21st ACM international conference on Multimedia, 2013.&nbsp;https://freesound.org

[5] Bart Thomee, David A. Shamma, Gerald Friedland, Benjamin Elizalde, Karl Ni, Douglas Poland, Damian Borth, and Li-Jia Li, YFCC100M: The New Data in Multimedia Research, Commun. ACM, 59(2):64&ndash;73, January 2016

Freesound Annotator:&nbsp;https://annotator.freesound.org/
Freesound:&nbsp;https://freesound.org
Eduardo Fonseca&#39;s personal website:&nbsp;http://www.eduardofonseca.net/
More datasets collected by us:&nbsp;http://www.eduardofonseca.net/datasets/

Acknowledgments

This work is partially supported by the European Union&rsquo;s Horizon 2020 research and innovation programme under grant agreement No 688382&nbsp;AudioCommons. Eduardo Fonseca is also sponsored by a&nbsp;Google Faculty Research Award 2018. We thank everyone who contributed to FSDKaggle2019 with annotations.
Data or Study Types
multiple
Source Organization
Unknown
Access Conditions
available
Year
2020
Access Hyperlink
https://doi.org/10.5281/zenodo.3612637

Distributions

  • Encoding Format: HTML ; URL: https://doi.org/10.5281/zenodo.3612637
This project was funded in part by grant U24AI117966 from the NIH National Institute of Allergy and Infectious Diseases as part of the Big Data to Knowledge program. We thank all members of the bioCADDIE community for their valuable input on the overall project.