Datasets:

Modalities:
Image
Text
Formats:
parquet
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:

Update dataset card: add task category, tags, and correct links

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +14 -10
README.md CHANGED
@@ -1,4 +1,16 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
2
  dataset_info:
3
  features:
4
  - name: task_category
@@ -40,21 +52,13 @@ configs:
40
  data_files:
41
  - split: train
42
  path: data/train-*
43
- license: cc-by-sa-4.0
44
- task_categories:
45
- - visual-question-answering
46
- language:
47
- - en
48
- pretty_name: ROME
49
- size_categories:
50
- - n<1K
51
  ---
52
 
53
  ![LRM-Eval](src/LRM-Eval.png)
54
 
55
- 🏠[Home & Leaderboard](https://github.com/flageval-baai/LRM-Eval) | 🤗[Data](https://huggingface.co/datasets/) | 🤗[Evaluation Response](https://huggingface.co/datasets/) | 💻[Code](https://github.com/flageval-baai/ROME-evaluation) | 📄[Paper](https://arxiv.org/pdf/2509.17177)
56
 
57
- This repository contains a visual reasoning benchmark named ROME from the paper [A Preliminary Contamination-Free Evaluation of Reasoning Models](https://arxiv.org/).
58
 
59
  ROME include 8 subtasks (281 high-quality questions in total). Each sample has been verified to ensure that images are necessary to answer correctly:
60
  * Academic
 
1
  ---
2
+ language:
3
+ - en
4
+ license: cc-by-sa-4.0
5
+ size_categories:
6
+ - n<1K
7
+ task_categories:
8
+ - image-text-to-text
9
+ pretty_name: ROME
10
+ tags:
11
+ - benchmark
12
+ - reasoning
13
+ - vlm
14
  dataset_info:
15
  features:
16
  - name: task_category
 
52
  data_files:
53
  - split: train
54
  path: data/train-*
 
 
 
 
 
 
 
 
55
  ---
56
 
57
  ![LRM-Eval](src/LRM-Eval.png)
58
 
59
+ 🏠[Project Page & Leaderboard](https://flageval-baai.github.io/LRM-Eval/) | 💻[Code](https://github.com/flageval-baai/ROME-evaluation) | 📄[Paper](https://huggingface.co/papers/2509.17177) | 🤗[Data](https://huggingface.co/datasets/FlagEval/ROME) | 🤗[Evaluation Response](https://huggingface.co/datasets/)
60
 
61
+ This repository contains a visual reasoning benchmark named ROME from the paper [FlagEval Findings Report: A Preliminary Evaluation of Large Reasoning Models on Automatically Verifiable Textual and Visual Questions](https://huggingface.co/papers/2509.17177).
62
 
63
  ROME include 8 subtasks (281 high-quality questions in total). Each sample has been verified to ensure that images are necessary to answer correctly:
64
  * Academic