MARBLE 🏔️

Leaderboard

This site provides the overall results of MARBLE benchmark under the constrained evaluation track. Details of the setting can be found at the submit page.

Noted

  1. The results of sequence labelling tasks (Beat Tracking and Source Seperation) are not available for specific models due to the cost limitation of the models.
  2. The results of Jukebox-5B model on MTG tasks hit the computational wall of one week on a single consumer GPU (RTX3090) in MARBLE.
TaggingKeyGenreRhythmEmotionInstrumentPitchTechSingerInstrumentMoodThemeGenreTop50Source SeparationOn Available Results
DatasetMTTGSGTZANGTZANEMONsynthNSynthVocalSetVocalSetMTGMTGMTGMTGMUSDBALL TASKS
TaskTaggingKeyGenreRhythmEmotionInstrumentPitchTechSingerInstrumentMoodThemeGenreTop50Source SeparationOn Available Results
Metric
ROC
AP
AccRefined
Acc
F1beat
R2V
R2A
Acc
Acc
Acc
Acc
ROC
AP
ROC
AP
ROC
AP
ROC
AP
SDRvocals
SDRdrums
SDRbass
SDRother
Average Score
MAP-MERT-v0-95M90.738.264.174.888.352.969.970.492.373.677.076.618.775.913.786.918.582.828.85.65.64.03.062.3
MAP-MERT-v0-95M-public90.738.467.372.888.159.172.870.492.375.678.077.519.676.213.387.218.883.028.95.55.53.73.063.0
MAP-MERT-v1-95M91.039.363.574.888.355.576.370.792.674.283.777.519.476.413.487.118.883.029.05.55.53.83.163.3
MAP-MERT-v1-330M91.139.561.777.687.959.075.872.694.476.987.178.119.876.514.086.718.683.429.95.35.63.63.064.2
MAP-Music2Vec90.036.250.674.168.252.171.069.393.171.181.476.119.276.714.387.118.883.029.25.55.54.13.059.9
MusiCNN90.337.814.473.5-44.068.872.664.170.357.074.017.274.012.686.017.582.027.5-----
CLMR89.536.014.865.2-44.470.367.947.058.149.973.517.073.512.684.616.281.326.4-----
Jukebox-5B91.440.663.877.9-57.073.070.491.676.782.678.522.077.615.388.020.583.430.4-----
MULE91.240.164.975.5-60.773.174.688.575.587.576.619.278.015.488.020.483.730.6-----

We will provide multiple leaderboads for different task categories in near future, including: