

The community's growing interest in feature and end-to-end learning is however restrained by the limited availability of large audio datasets.
AUTOMATIC MUSIC TAG ARCHIVE
For reproducibility, we provide the PyTorch implementations with the pre-trained models.read more read lessĪbstract: We introduce the Free Music Archive (FMA), an open and easily accessible dataset which can be used to evaluate several tasks in music information retrieval (MIR), a field concerned with browsing, searching, and organizing large music collections. Furthermore, all the models are evaluated with perturbed inputs to investigate the generalization capabilities concerning time stretch, pitch shift, dynamic range compression, and addition of white noise. To facilitate further research, in this paper we conduct a consistent evaluation of different music tagging models on three datasets (MagnaTagATune, Million Song Dataset, and MTG-Jamendo) and provide reference results using common evaluation metrics (ROC-AUC and PR-AUC).

However, due to the differences in experimental setups followed by researchers, such as using different dataset splits and software versions for evaluation, it is difficult to compare the proposed architectures directly with each other.

Music information retrieval (MIR) researchers proposed various architecture designs, mainly based on convolutional neural networks (CNNs), that achieve state-of-the-art results in this multi-label binary classification task. Abstract: Recent advances in deep learning accelerated the development of content-based automatic music tagging systems.
