Home

Awesome

BrainHub: Multimodal Brain Understanding Benchmark

Updates

Motivation

Unlike texts, images, or audio, whose contents are intuitively aligned with human perception and judgment, we lack sufficient knowledge of the information contained in captured brain responses, as they are not directly interpretable or interoperable to humans. We could translate the brain's responses into other understandable modalities as an indirect method of ascertaining its ability to describe, recognize, and localize instances, as well as discern spatial relationships among multiple exemplars. These abilities are important for brain-machine interfaces and other brain-related research. Therefore, we construct BrainHub, a brain understanding benchmark, based on NSD and COCO.

Tasks and Metrics

The objectives are categorized into concept recognition and spatial localization, including:

Evaluation

There are 982 test images, 80 classes, 4,913 captions, and 5,829 boundingboxes. For grounding evaluation, we further group the 80 classes of COCO into four salience categories according to their salience in images: Salient (S), Salient Creatures (SC), Salient Objects (SO), and Inconspicuous (I). The illustration shows the statistics and mapping of our categories, w.r.t. COCO classes.

We provide the processed text and boundingbox groundtruth. The demo evaluation script is provided here. If you would like to evaluate your produced results, please modify the result path accordingly.

We also provide baseline results associated with BrainHub, including the captioning results from SDRecon, BrainCap, and OneLLM, as well as the captioning and grounding results from UMBRAE.

For contributing, please (a) update the leaderboard and (b) upload the results to the desired path with the required file name, such as caption/comparison/umbrae/sub01_decoded_caption.json.

Leaderboard

Captioning

This is the quantitative comparison for subject 1 (S1). For results on other subjects, refer to the paper. 'UMBRAE-S1' refers to model trained with S1 only, while 'UMBRAE' denotes the model with cross-subject training.

MethodEvalBLEU1BLEU4METEORROUGECIDErSPICECLIPSRefCLIPS
UMBRAES159.4419.0319.4543.7161.0612.7967.7873.54
UMBRAE-S1S157.6316.7618.4142.1551.9311.8366.4472.12
BrainCapS155.9614.5116.6840.6941.309.0664.3169.90
OneLLMS147.049.5113.5535.0522.996.2654.8061.28
SDReconS136.213.4310.0325.1313.835.0261.0766.36
MethodEvalBLEU1BLEU4METEORROUGECIDErSPICECLIPSRefCLIPS
UMBRAES259.3718.4119.1743.8655.9312.0866.4672.36
UMBRAE-S2S257.1817.1818.1141.8550.6211.5064.8771.06
BrainCapS253.8013.0315.9039.9635.608.4762.4868.19
SDReconS234.713.029.6024.2213.384.5859.5265.30
MethodEvalBLEU1BLEU4METEORROUGECIDErSPICECLIPSRefCLIPS
UMBRAES560.3619.0320.0444.8161.3213.1968.3974.11
UMBRAE-S5S558.9918.7319.0443.3057.0912.7066.4872.69
BrainCapS555.2814.6216.4540.8741.059.2463.8969.64
SDReconS534.963.499.9324.7713.855.1960.8366.30
MethodEvalBLEU1BLEU4METEORROUGECIDErSPICECLIPSRefCLIPS
UMBRAES757.2017.1318.2942.1652.7311.6365.9071.83
UMBRAE-S7S755.7115.7517.5140.6447.0711.2663.6670.09
BrainCapS754.2514.0015.9440.0237.498.5762.5268.48
SDReconS734.993.269.5424.3313.014.7458.6864.59

Grounding

MethodEvalacc@0.5 (A)IoU (A)acc@0.5 (S)IoU (S)acc@0.5 (I)IoU (I)
UMBRAES118.9321.2830.2330.184.8310.18
UMBRAE-S1S113.7217.5621.5225.144.008.08
UMBRAES218.2720.7728.2229.195.861025
UMBRAE-S2S215.2118.6823.6026.594.748.81
UMBRAES518.1920.8528.7430.025.029.41
UMBRAE-S5S514.7218.4522.9326.344.468.60
UMBRAES716.7419.6325.6927.905.589.31
UMBRAE-S7S713.6017.8321.0725.194.288.64

Citation

@inproceedings{xia2024umbrae,
  author    = {Xia, Weihao and de Charette, Raoul and Öztireli, Cengiz and Xue, Jing-Hao},
  title     = {UMBRAE: Unified Multimodal Brain Decoding},
  booktitle = {European Conference on Computer Vision (ECCV)},
  year      = {2024},
}