This repository shares the caption annotations of the 1200 GOD training stimulus images used in the following Kamitani lab's representative papers.
- Horikawa, T., & Kamitani, Y. (2017). Generic decoding of seen and imagined objects using hierarchical visual features. Nature Communications, 8, 15037. https://doi.org/10.1038/ncomms15037
- Shen, G., Horikawa, T., Majima, K., & Kamitani, Y. (2019). Deep image reconstruction from human brain activity. PLOS Computational Biology, 15(1), 1006633. https://doi.org/10.1371/journal.pcbi.1006633
- Horikawa, T., & Kamitani, Y. (2022). Attention modulates neural representation to render reconstructions according to subjective appearance. Communications Biology, 5(1), 34. https://doi.org/10.1038/s42003-021-02975-5
- Cheng, F., Horikawa, T., Majima, K., Tanaka, M., Abdelhack, M., Aoki, S. C., Hirano, J., & Kamitani, Y. (2023). Reconstructing visual illusory experiences from human brain activity (p. 2023.06.15.545037). bioRxiv. https://doi.org/10.1101/2023.06.15.545037
Caption data and category information are stored in *.jsonl
file. Please see example.ipynb for detailed data usage.
df = pd.read_json("data/ImageNetTraining_captions.jsonl", orient="records", lines=True)
display(df)
image_id category_id captions
0 n01518878_10042 n01518878 [A brown and tan ostridge walking near a metal...
1 n01518878_12028 n01518878 [The head of an ostridge looking over a metal ...
2 n01518878_14075 n01518878 [A brown ostridge with its beak open walking i...
3 n01518878_14910 n01518878 [An ostrich walks across a dirt road towards g...
4 n01518878_5958 n01518878 [An ostrich standing in a field with trees beh...
... ... ... ...
1195 n13111881_28475 n13111881 [A small bonsai tree sits in the dirt in a sma...
1196 n13111881_6054 n13111881 [Miniature manicured trees on display table ou...
1197 n13111881_6560 n13111881 [Young woman watering her home plants joyfully...
1198 n13111881_7817 n13111881 [A bonsai tree in a bowl near a red wall., A s...
1199 n13111881_9170 n13111881 [A bonsai tree with gnarly white branches and ...
1200 rows × 3 columns
The caption data was collected in 2018 using Amazon Mechanical Turk (AMT), a major crowdsourcing service. Workers have been filtered by the following two conditions;
- Workers are granted "Masters".
- Workers have HIT Approval Rate greater than 95% for all Requesters' HITs. For further detail on the filtering conditions, please refer to here.
The captions for test stimulus images are not publicly available to prevent data crawling. If you wish to be shared the test image captions, please contact us via email ([email protected]).
The stimulus images cannot be shared due to copyright problems. If you need the images, please contact us from here. fMRI data is open to the public on Figshare, etc. Please see the GitHub repository of the above representative papers for details.