--- dataset_info: features: - name: obs_uid dtype: string - name: usr_uid dtype: string - name: caption dtype: string - name: image dtype: image - name: clicks_path sequence: sequence: int32 length: 2 - name: clicks_time sequence: timestamp[s] splits: - name: train num_bytes: 1611467 num_examples: 3848 download_size: 241443505 dataset_size: 1611467 --- ### Dataset Description CapMIT1003 is a dataset of captions and click-contingent image explorations collected during captioning tasks. CapMIT1003 is based on the same stimuli from the well-known MIT1003 benchmark, for which eye-tracking data under free-viewing conditions is available, which offers a promising opportunity to concurrently study human attention under both tasks. ### Usage You can load CapMIT1003 as follows: ```python from datasets import load_dataset capmit1003_dataset = load_dataset("azugarini/CapMIT1003", trust_remote_code=True) print(capmit1003_dataset["train"][0]) #print first example ``` ### Citation Information If you use this dataset in your research or work, please cite the following paper: ``` @article{zanca2023contrastive, title={Contrastive Language-Image Pretrained Models are Zero-Shot Human Scanpath Predictors}, author={Zanca, Dario and Zugarini, Andrea and Dietz, Simon and Altstidl, Thomas R and Ndjeuha, Mark A Turban and Schwinn, Leo and Eskofier, Bjoern}, journal={arXiv preprint arXiv:2305.12380}, year={2023} ```