MMMU Save

This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI"

Project README

MMMU

🌐 Homepage | πŸ€— Dataset | πŸ€— Paper | πŸ“– arXiv | GitHub

This repo contains the evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI"

πŸ””News

  • πŸš€[2024-01-31]: We added Human Expert performance on the Leaderboard!🌟
  • πŸ”₯[2023-12-04]: Our evaluation server for test set is now availble on EvalAI. We welcome all submissions and look forward to your participation! πŸ˜†

Introduction

We introduce MMMU: a new benchmark designed to evaluate multimodal models on massive multi-discipline tasks demanding college-level subject knowledge and deliberate reasoning. MMMU includes 11.5K meticulously collected multimodal questions from college exams, quizzes, and textbooks, covering six core disciplines: Art & Design, Business, Science, Health & Medicine, Humanities & Social Science, and Tech & Engineering. These questions span 30 subjects and 183 subfields, comprising 32 highly heterogeneous image types, such as charts, diagrams, maps, tables, music sheets, and chemical structures. Unlike existing benchmarks, MMMU focuses on advanced perception and reasoning with domain-specific knowledge, challenging models to perform tasks akin to those faced by experts. Our evaluation of 14 open-source LMMs and the proprietary GPT-4V(ision) highlights the substantial challenges posed by MMMU. Even the advanced GPT-4V only achieves a 56% accuracy, indicating significant room for improvement. We believe MMMU will stimulate the community to build next-generation multimodal foundation models towards expert artificial general intelligence (AGI).

Alt text

Dataset Creation

MMMU was created to challenge multimodal models with tasks that demand college-level subject knowledge and deliberate reasoning, pushing the boundaries of what these models can achieve in terms of expert-level perception and reasoning. Please refer to our huggingface πŸ€— Dataset for more details.

Evaluation

Please refer to our eval folder for more details.

πŸ† Mini-Leaderboard

Model Val (900) Test (10.5K)
Expert (Best) 88.6 -
Expert (Medium) 82.6 -
Expert (Worst) 76.2 -
Gemini Ultra* 59.4 -
Claude 3 Opus* 59.4 -
GPT-4V(ision) (Playground) 56.8 55.7
Claude 3 Sonnet* 53.1 -
HPT Pro* 52.0 -
Qwen-VL-MAX* 51.4 46.8
InternVL-Chat-V1.2* 51.6 46.2
Reka Flash* 51.3 -
LLaVA-1.6-34B* 51.1 44.7
Claude 3 Haiku* 50.2 -
Adept Fuyu-Heavy* 48.3 -
Gemini Pro* 47.9 -
Marco-VL-Plus* 46.2 44.3
Yi-VL-34B* 45.9 41.6
Qwen-VL-PLUS* 45.2 40.8
HPT Air* 44.0 -
Marco-VL* 41.2 40.4
OmniLMM-12B* 41.1 40.4
Weitu-VL-1.0-15B* - 38.4
InternLM-XComposer2-VL* 43.0 38.2
Yi-VL-6B* 39.1 37.8
InfiMM-Zephyr-7B* 39.4 35.5
InternVL-Chat-V1.1* 39.1 35.3
SVIT* 38.0 34.1
MiniCPM-V* 37.2 34.1
Emu2-Chat* 36.3 34.1
BLIP-2 FLAN-T5-XXL 35.4 34.0
InstructBLIP-T5-XXL 35.7 33.8
LLaVA-1.5-13B 36.4 33.6
Bunny-3B* 38.2 33.0
Qwen-VL-7B-Chat 35.9 32.9
SPHINX* 32.9 32.9
mPLUG-OWL2* 32.7 32.1
BLIP-2 FLAN-T5-XL 34.4 31.0
InstructBLIP-T5-XL 32.9 30.6
Gemini Nano2* 32.6 -
CogVLM 32.1 30.1
Otter 32.2 29.1
LLaMA-Adapter2-7B 29.8 27.7
MiniGPT4-Vicuna-13B 26.8 27.6
Adept Fuyu-8B 27.9 27.4
Kosmos2 24.4 26.6
OpenFlamingo2-9B 28.7 26.3
Frequent Choice 22.1 23.9
Random Choice 26.8 25.8

*: results provided by the authors.

🎯 We have released a full suite comprising 150 development samples and 900 validation samples. However, the 10,500 test questions are available without their answers. Use the development set for few-shot/in-context learning, and the validation set for debugging models, selecting hyperparameters, and quick evaluations. The answers and explanations for the test set questions are withheld. You can submit your model's predictions for the test set on EvalAI.

Disclaimers

The guidelines for the annotators emphasized strict compliance with copyright and licensing rules from the initial data source, specifically avoiding materials from websites that forbid copying and redistribution. Should you encounter any data samples potentially breaching the copyright or licensing regulations of any site, we encourage you to contact us. Upon verification, such samples will be promptly removed.

Contact

Citation

BibTeX:

@inproceedings{yue2023mmmu,
  title={MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI},
  author={Xiang Yue and Yuansheng Ni and Kai Zhang and Tianyu Zheng and Ruoqi Liu and Ge Zhang and Samuel Stevens and Dongfu Jiang and Weiming Ren and Yuxuan Sun and Cong Wei and Botao Yu and Ruibin Yuan and Renliang Sun and Ming Yin and Boyuan Zheng and Zhenzhu Yang and Yibo Liu and Wenhao Huang and Huan Sun and Yu Su and Wenhu Chen},
  booktitle={Proceedings of CVPR},
  year={2024},
}
Open Source Agenda is not affiliated with "MMMU" Project. README Source: MMMU-Benchmark/MMMU