task_categories:
- question-answering
- visual-question-answering
language:
- en
tags:
- Multimodal Search
- Multimodal Long Context
size_categories:
- n<1K
configs:
- config_name: default
data_files:
- split: train
path: '*.arrow'
dataset_info:
features:
- name: question
dtype: string
- name: answer
sequence: string
- name: num_images
dtype: int64
- name: arxiv_id
dtype: string
- name: video_url
dtype: string
- name: category
dtype: string
- name: difficulty
dtype: string
- name: subtask
dtype: string
- name: img_1
dtype: image
- name: img_2
dtype: image
- name: img_3
dtype: image
- name: img_4
dtype: image
- name: img_5
dtype: image
splits:
- name: train
num_examples: 311
MMSearch-Plusβ¨: Benchmarking Provenance-Aware Search for Multimodal Browsing Agents
Official repository for the paper "MMSearch-Plus: Benchmarking Provenance-Aware Search for Multimodal Browsing Agents".
π For more details, please refer to the project page with examples: https://mmsearch-plus.github.io/.
[π Webpage] [π Paper] [π€ Huggingface Dataset] [π Leaderboard]
π₯ News
- [2025.09.26] π₯ We update the arXiv paper and release all MMSearch-Plus data samples in huggingface dataset.
- [2025.08.29] π We release the arXiv paper.
π ToDo
- Agentic rollout framework code
- Evaluation script
- Set-of-Mark annotations
Usage
β οΈ Important: This dataset is encrypted to prevent data contamination. However, decryption is handled transparently by the dataset loader.
Dataset Usage
For better compatibility with newer versions of the datasets library, we provide explicit decryption functions, downloadable from our GitHub/HF repo.
wget https://github.com/mmsearch-plus/MMSearch-Plus/blob/main/decrypt_after_load.py
import os
from datasets import load_dataset
from decrypt_after_load import decrypt_dataset
encrypted_dataset = load_dataset("Cie1/MMSearch-Plus", split='train')
decrypted_dataset = decrypt_dataset(
encrypted_dataset=encrypted_dataset,
canary='your_canary_string' # Set the canary string (hint: it's the name of this repo without username)
)
# Access a sample
sample = decrypted_dataset[0]
print(f"Question: {sample['question']}")
print(f"Answer: {sample['answer']}")
print(f"Category: {sample['category']}")
print(f"Number of images: {sample['num_images']}")
# Access images (PIL Image objects)
sample['img_1'].show() # Display the first image
π About MMSearch-Plus
MMSearch-Plus is a challenging benchmark designed to test multimodal browsing agents' ability to perform genuine visual reasoning. Unlike existing benchmarks where many tasks can be solved with text-only approaches, MMSearch-Plus requires models to extract and use fine-grained visual cues through iterative image-text retrieval.
Key Features
π Genuine Multimodal Reasoning: 311 carefully curated tasks that cannot be solved without visual understanding
π― Fine-grained Visual Analysis: Questions require extracting spatial cues and temporal traces from images to find out-of-image facts like events, dates, and venues
π οΈ Agent Framework: Model-agnostic web agent with standard browsing tools (text search, image search, zoom-in)
π Set-of-Mark (SoM) Module: Enables provenance-aware cropping and targeted searches with human-verified bounding box annotations
Dataset Structure
Each sample contains:
- Quuestion text and images
- Ground truth answers and alternative valid responses
- Metadata including arXiv id (if an event is a paper), video URL (if an event is a video), area and subfield
Performance Results
Evaluation of closed- and open-source MLLMs shows:
- Best accuracy is achieved by o3 with full rollout: 36.0% (indicating significant room for improvement)
- SoM integration provides consistent gains up to +3.9 points
- Models struggle with multi-step visual reasoning and cross-modal information integration
The overview of three paradigms for multimodal browsing tasks that demand fine-grained visual reasoning.
The overview of an example trajectory for a task in MMSearch-Plus.
π Leaderboard
Contributing to the Leaderboard
π¨ The Leaderboard is continuously being updated, welcoming the contribution of your excellent LMMs!
π Citation
If you find MMSearch-Plus useful for your research and applications, please kindly cite using this BibTeX:
@article{tao2025mmsearch,
title={MMSearch-Plus: A Simple Yet Challenging Benchmark for Multimodal Browsing Agents},
author={Tao, Xijia and Teng, Yihua and Su, Xinxing and Fu, Xinyu and Wu, Jihao and Tao, Chaofan and Liu, Ziru and Bai, Haoli and Liu, Rui and Kong, Lingpeng},
journal={arXiv preprint arXiv:2508.21475},
year={2025}
}