You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

🧠 EgoExoBench: A Cross-Perspective Video Understanding Benchmark

Dataset Summary

EgoExoBench is a benchmark designed to evaluate cross-perspective understanding capabilities of multimodal large models (MLLMs).
It contains synchronized and asynchronous egocentric (first-person) and exocentric (third-person) video pairs, along with multiple-choice questions that assess semantic alignment, viewpoint association, and temporal reasoning between the two perspectives.

Features

Each sample contains:

  • Question: A natural-language question testing cross-perspective reasoning.
  • Options: Multiple-choice answers (A/B/C/D).
  • Answer: Correct answer label.
  • Videos: Egocentric and Exocentric videos.

Evaluation Metric

Accuracy (%)

Data Splits

Split #Samples
Test 7,330

Example Usage

from datasets import load_dataset

dataset = load_dataset("YourUsername/EgoExoBench")

print(dataset["test"][0])

Citation

If you use EgoExoBench in your research, please cite:

@misc{he2025egoexobench,
      title={EgoExoBench: A Benchmark for First- and Third-person View Video Understanding in MLLMs}, 
      author={Yuping He and Yifei Huang and Guo Chen and Baoqi Pei and Jilan Xu and Tong Lu and Jiangmiao Pang},
      year={2025},
      eprint={2507.18342},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2507.18342}
}
Downloads last month
96