Awesome
magnatagatune-list
List of automatic music tagging (using audio) research articles that are evaluated against MagnaTagATune Dataset
Download
MagnaTagATune is kindly hosted by MIRG of City University London. Visit here to download mp3s and others.
Folders and files
After download three files, $ cat mp3.zip.* > mp3_all.zip
to merge them, then unzip it by unzip mp3_all.zip
.
You then got 16 folders, '0' to '9' and then 'a' to 'f'.
Many works based on splitting the folders as 13:1:3 for training/validatin/testing. However it might be the best choice since each folders do NOT have same tag distributions. Minimum line is to shuffle training set, and would like to recommend to shuffle all of them and then split. Still you might need the same splitting to comparison, as in my case :-(
Tags (sorted by popularity)
Top 50
guitar, classical, slow, techno, strings, drums, electronic, rock, fast, piano, ambient, beat, violin, vocal, synth, female, indian, opera, male, singing, vocals, no vocals, harpsichord, loud, quiet, flute, woman, male vocal, no vocal, pop, soft, sitar, solo, man, classic, choir, voice, new age, dance, male voice, female vocal, beats, harp, cello, no voice, weird, country, metal, female voice, choral
Top 50 by categories
- genre:
classical, techno, electronic, rock, indian, opera, pop, classic, new age, dance, country, metal
- instrument:
guitar, strings, drums, piano, violin, vocal, synth, female, male, singing, vocals, no vocals, harpsichord, flute, no vocal, sitar, man, choir, voice, male voice, female vocal, harp, cello, femal voice, choral
- mood:
slow, fast, ambient, loud, quiet, soft, weird
- etc:
beat, solo, beats
Top 51-188
electro, drum, male vocals, jazz, violins, eastern, female vocals, instrumental, bass, modern, no piano, harpsicord, jazzy, string, baroque, foreign, orchestra, hard rock, electric, trance, folk, chorus, chant, voices, classical guitar, spanish, heavy, upbeat, no guitar, acoustic, male singer, electric guitar, electronica, oriental, funky, tribal, banjo, dark, medieval, man singing, organ, blues, irish, no singing, bells, percussion, no drums, woman singing, noise, spacey, singer, female singer, middle eastern, chanting, no flute, low, strange, calm, wind, lute, heavy metal, different, punk, oboe, celtic, sax, flutes, talking, women, arabic, hard, mellow, funk, fast beat, house, rap, not english, no violin, fiddle, female opera, water, india, guitars, no beat, chimes, drone, male opera, trumpet, duet, birds, industrial, sad, plucking, girl, silence, men, operatic, horns, repetitive, airy, world, eerie, deep, hip hop, space, light, keyboard, english, not opera, not classical, not rock, clapping, horn, acoustic guitar, disco, orchestral, no strings, old, echo, lol, soft rock, no singer, jungle, bongos, reggae, monks, clarinet, scary, synthesizer, female singing, piano solo, no voices, woodwind, happy, viola, soprano, quick, clasical
histogram of tags
So the dataset is unbalanced.
Proposed tag preprocessing
I wrote code to merge these synonyms.
synonyms = [['beat', 'beats'],
['chant', 'chanting'],
['choir', 'choral'],
['classical', 'clasical', 'classic'],
['drum', 'drums'],
['electro', 'electronic', 'electronica', 'electric'],
['fast', 'fast beat', 'quick'],
['female', 'female singer', 'female singing', 'female vocals', 'female voice', 'woman', 'woman singing', 'women'],
['flute', 'flutes'],
['guitar', 'guitars'],
['hard', 'hard rock'],
['harpsichord', 'harpsicord'],
['heavy', 'heavy metal', 'metal'],
['horn', 'horns'],
['india', 'indian'],
['jazz', 'jazzy'],
['male', 'male singer', 'male vocal', 'male vocals', 'male voice', 'man', 'man singing', 'men'],
['no beat', 'no drums'],
['no singer', 'no singing', 'no vocal','no vocals', 'no voice', 'no voices', 'instrumental'],
['opera', 'operatic'],
['orchestra', 'orchestral'],
['quiet', 'silence'],
['singer', 'singing'],
['space', 'spacey'],
['string', 'strings'],
['synth', 'synthesizer'],
['violin', 'violins'],
['vocal', 'vocals', 'voice', 'voices'],
['strange', 'weird']]`
I'm not 100% sure if these should be merged.
['opera', 'operatic'],
['hard', 'hard rock'],
Papers
This list is based on google scholar, list of papers that cited the dataset, google scholar search result
2016
not yet
2015
According to this,
- U Sandouk et al., arxiv: text-processing.
- U Sandouk et al., arxiv: text-processing.
- E Colautti et al. : it's not relevant.
(and,)
- J Nam et al., arxiv: pdf
- title: A deep bag-of-features model for music auto-tagging
- algorithm: bag-of-features (mel-specgram, onsets) --> PCA --> RBM for feature extraction --> DNN + stackedRBM
- set: 14600/1629/6499 for training/validation/test
- tags: used 160 tags
- (selected) result
algorithm | Deep-BoF (proposed) | P 2011 | P Hamel 2011 | P Hamel 2012 |
---|---|---|---|---|
AUC tag | 0.888 | 0.845 | 0.861 | 0.870 |
AUC clip | 0.956 | 0.938 | 0.943 | 0.949 |
Precision@3 | 0.511 | 0.449 | 0.467 | 0.481 |
Precision@6 | 0.358 | 0.430 | 0.327 | 0.339 |
Precision@9 | 0.275 | 0.249 | 0.255 | 0.263 |
Precision@12 | 0.225 | 0.205 | 0.211 | 0.216 |
Precision@15 | 0.190 | 0.175 | 0.181 | 0.184 |
2014
According to this,
- D Lim at al., JMLR: about metrics
- S Duan et al.: survey of tagging techniques
- G Sageder et al.: not used. onlysubset of MagnaTagATune is used to verity the proposed feature selection that is based on ismir2014 DB.
(and,)
- S Dieleman et al.: pdf
- title: End-to-end learning for music audio
- algorithm: end-to-end setting from audio (and spectrogram for comparison), 1d conv - MP - 1d conv - MP - fc layers
- set: 16-folds, 1-12/13/14-6 for training/validation/test
- (selected) result
filter length | stride | AUC (spectrograms) | AUC (raw audio) |
---|---|---|---|
1024 | 1024 | 0.8690 | 0.8366 |
256 | 256 | 0.8815 | 0.8487 |
- A Oord et al.: pdf
- title: Transfer learning by supervised pre-training for audio-based music classification
- algorithm: transfering learned MLP into problems based on other dataset.
- network: k-means feature learning --> MLP
- result
| model | nmse | mean average precision (mAP) | | ------------- |:-------------:|:-------------:|:-------------:| | linear regression | 0.965 | 0.823 | 0.0099 | | MLP (1 hidden layer) | 0.939 | 0.841 | 0.0179 | | MLP (2 hidden layers) | 0.924 | 0.837 | 0.0179 |
task | AUC |
---|---|
tag (top 50 tags) | <0.88 |
tag (all 188 tags) | <0.86 |
- SN Tran et al.: pdf: about similarity
- title: feature preprocessing with RBMs for music similarity learning
- F Gouton et al.: pdf
-
title: on evaluation validity in music autotagging
-
used MagTag5k, which is subset of MagnaTagATune. Now the link is broken.
-
2013
According to this,
- D Wolff et al. : similarity
- J Watson et al., arxiv: proposing an algorithm for ranking prediction.
- set: 16k for training, 160 tags (probably top-160 popular tags)
- feature: 104-dim MFCC, 100-dim embedding dimension
- result
algorithm | precision@1 | precision@3 |
---|---|---|
k-nn | 39.4% | 28.6% |
k-nn (Wsabie space) | 45.2% | 31.9% |
Wsabie | 48.7% | 37.5% |
Affinity Weighted Embedding | 52.7% | 39.2% |
- JA Burgoyne et al., ismir: just mentioning.
(and,)
- S Dieleman et al., ismir: pdf
- title: Multiscale Approaches To Music Audio Feature Learning
- result
algorithm | average AUC |
---|---|
Laplacian 1 frame | 0.898 |
Multiresolution spectrograms | 0.888 |
- J Stastny et al.: pdf
- title: audio data classification by means of new algorithms
- Not on tag prediction
2012
According to this,
-
D wolff et al.: just mentioning.
-
Y Li et al.: not about the DB
-
M Levy: can't get the pdf of it, seems like internal document in the school.
-
L Barrington, PhD thesis: probably not focusing on automatic tagging
-
P Hamel, PhD thesis: written in french. according to its list of contents it would be same results as in the author's ismir paper. (and,)
-
J NAM et al.,: mirex 2012 submission
- title: mirex 2012 submission: audio classification using sparse feature learning
- algorithm: RBM + SVM
- result
- AROC (ranking): 0.7244,
- F-measure (annotation): 0.1123
2011
According to this,
- AJ Quinn et al., SIGCHI: a general survey on human computation
- YH Yang et al. IEEE ASLP: just mentining. experiment was performed on smaller private dataset
- P Hamel et al.:
- title: temporal pooling and multiscale learning for automatic annotation and ranking of music audio
- features: MFCC
- set: 14600/1629/6499 for training/validation/test
- tags: not specified, probably using all 160 tags.
- comparison: against four algirithms [19][17][18][9] (see the references in the paper), all of which use MFCC or 'cepstral transform that is closely related to MFCCs' + spectral features (spectrao centroid, rolloff, flux). They also used GMM of tags, VQ, PAMIR, or individual SVM.
- features: 128 MFCC --> 120 principle components, which is called PMSC
- then mean,var,max,min, 3rd and 4th centered moments -pooling
- model 1: MLP with L2 regularisation and cross-entropy for loss function.
- model 2: multi-time-scale learning
- evaluation metric: AUC
- result:
table 1.
algorithm | AUC | average training time |
---|---|---|
MFCC(20) | 0.77+-0.04 | 5.9gh |
MEL-spectrum(128) | 0.853+-0.008 | 5.2h |
PMSC(120) | 0.876+-0.004 | 1.5h |
table 2.
measure | Manzagol | Zhi | Mandel | Marsyas | mel-spec+PFC | PMSC+PFC | PSMC+MTSL |
---|---|---|---|---|---|---|---|
average AUC-tag | 0.750 | 0.673 | 0.821 | 0.831 | 0.820 | 0.845 | 0.861 |
average AUC-clip | 0.810 | 0.748 | 0.886 | 0.933 | 0.930 | 0.938 | 0.943 |
precision@3 | 0.255 | 0.224 | 0.323 | 0.440 | 0.430 | 0.449 | 0.467 |
precision@6 | 0.194 | 0.192 | 0.245 | 0.314 | 0.305 | 0.320 | 0.327 |
precision@9 | 0.159 | 0.168 | 0.197 | 0.244 | 0.240 | 0.249 | 0.255 |
precision@12 | 0.136 | 0.146 | 0.167 | 0.201 | 0.198 | 0.205 | 0.211 |
precision@15 | 0.119 | 0.127 | 0.145 | 0.172 | 0.170 | 0.175 | 0.181 |
left four: rsults from MIREX 2009
- J Weston et al., JNMR: http://static.googleusercontent.com/media/research.google.com/en//pubs/archive/37179.pdf
- D Wolff, ismir: about similarity metric
- J Weston et al., arXiv: same as above JNMR, but with tables at within text. http://arxiv.org/pdf/1105.5196.pdf
- M Mann et al., ismir: mood predction in reduced dimensions.
- D Wolff et al., : about similarity
- T Aizenberg et al.: about melodic salience modelling
(and,)
- S Dieleman et al., ismir 2011:
- title: audio-based music classification with a pretrained convolutional network
2010
According to this,
- E Law et al., MLKD: learning to tag from open voca labels
- M Mandel et al.: learning tags that vary withnin a song
- A Quinn et al.: not related.
- E Law et al., ECML: learning to tag using noisy labels
- Y Panagakis et al.: sparse multi-label linear embedding NNTF for automatic music tagging
- F Maillet: about music recommendation
(and,)
- K Seyerlehner et al.: pdf
- title: Automatic music tag classification based on block-level
- metric: f-score and G-mean (geometric mean of true negative rate and recall)
- result:
feature set | f-score | g-mean | f-score (S2) | g-mean(S20 |
---|---|---|---|---|
SAF | 0.3775 | 0.6101 | 0.3962 | 0.6252 |
BLF-PCA | 0.4163 | 0.6410 | 0.4201 | 0.6439 |
feature set | avg. f-score | avg. g-mean | avg. f-score (S2) | avg. g-mean(S20 |
---|---|---|---|---|
SAF | 0.1777 | 0.3573 | 0.1932 | 0.3784 |
BLF-PCA | 0.2136 | 0.4019 | 0.2185 | 0.4081 |
-
J Bergstra et al., ismir 2010: pdf
- used few selected tags only
-
SR Ness et al.: [pdf](Improving automatic music tag annotation using stacked generalization of probabilistic svm outputs)
- title: Improving automatic music tag annotation using stacked generalization of probabilistic svm outputs
- audio features (+Affinity) + SVM
- result
global average (table 2)
algorithm | precision | recall | accuracy | F-score |
---|---|---|---|---|
Audio SVM | 0.307 | 0.315 | 0.969 | 0.311 |
Affinity SVM | 0.351 | 0.354 | 0.971 | 0.353 |
affinity svm - per-tag evaluation (table 4) (table 3 is ommited as table 4 outperforms in overall)
number of tags | precision | recall | accuracy | F-score |
---|---|---|---|---|
20 | 0.418 | 0.691 | 0.856 | 0.518 |
30 | 0.346 | 0.671 | 0.862 | 0.453 |
40 | 0.394 | 0.397 | 0.914 | 0.395 |
50 | 0.369 | 0.372 | 0.923 | 0.371 |
100 | 0.259 | 0.262 | 0.951 | 0.260 |
all (188) | 0.184 | 0.186 | 0.971 | 0.185 |
2009
N/A