You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository contains companion source code for working with the OpenMIC-2018 dataset, a collection of audio and crowd-sourced instrument labels produced in a collaboration between Spotify and New York Universiy's MARL and Center for Data Science. The cost of annotation was sponsored by Spotify, whose contributions to open-source research can be found online at the developer site, engineering blog, and public GitHub.
If you use this dataset, please cite the following work:
Humphrey, Eric J., Durand, Simon, and McFee, Brian. "OpenMIC-2018: An Open Dataset for Multiple Instrument Recognition." in Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), 2018. pdf
Download the Dataset
The OpenMIC-2018 dataset is made available on Zenodo. After downloading, decompress with your favorite commandline tar utility:
$ tar xvzf openmic-2018-v1.0.0.tgz -C some/dir
This will expand into some/dir/openmic-2018, with the following structure:
The openmic-2018.npz is a Python-friendly composite of the vggish features and the openmic-2018-aggregated-labels.csv. An example of how to train and evaluate a model is provided in a tutorial notebook.
Installing
To use the provided openmic Python library, first clone the repository and change directory into it:
$ git clone https://github.com/cosmir/openmic-2018.git
$ cd ./openmic-2018
Next, you'll want to pull down the VGGish model parameters via the following script.
$ ./scripts/download-deps.sh
Finally, you can now install the Python library, e.g. with pip:
$ pip install .
Errata
When initially collecting data, ten audio files were corrupted due to an issue in the source FMA dataset: