This study brings major advances in encoding and decoding cortical activity that supports human natural vision. For encoding, it demonstrates the unique promise of using deep learning to model and visualize the functional representations at the level of single cortical locations along the entire visual pathway, and to create a computational workbench for high-throughput vision research. For decoding, the study presents a stand-alone, efficient, reliable, and generalizable strategy to decode cortical fMRI activity to directly reconstruct the visual and semantic experiences during natural vision. These unique capabilities highlight a promising emerging direction of using the artificial brain to under-stand the biological brain.
The Laboratory of Integrated Brain Imaging (LIBI) at Purdue University acquired 3T fMRI responses from three subjects during watching natural movies (Wen et al., 2017). The movie stimuli contain diverse yet representative of real-life visual experiences, e.g. people in action, moving animals, nature scenes, outdoor or indoor scenes etc. The stimuli include two sets of movie segments: 1) 18 training movie segments, and 2) 5 testing movie segments. The duration of each segment is 8 minutes. During each fMRI scanning session, one segment was presented to the subjects. For each subject, the training movie segments were presented twice and the testing movie segments were presented ten times. In total, there are 11.47 hours of fMRI responses to 3.07 hours of movie stimuli for each subject.
This is publication series that contains data from three subjects, datasets with the stimuli and source code.
Wen, H., Shi, J., Zhang, Y., Lu, KH., Cao JY. & Liu, Z. (2017). Neural Encoding and Decoding with Deep Learning for Dynamic Natural Vision. Cerebral cortex. In press.
Content List ( Show bundle contents ) Bundle
This is a video-fMRI dataset contains the videos with stimuli acquired by the Laboratory of Integrated Brain Imaging (LIBI).
This is a video-fMRI dataset for subject 1 (out of three) acquired by the Laboratory of Integrated Brain Imaging (LIBI),
This is a video-fMRI dataset for subject 2 (out of three) acquired by the Laboratory of Integrated Brain Imaging (LIBI).
This is a video-fMRI dataset for subject 3 (out of three) acquired by the Laboratory of Integrated Brain Imaging (LIBI).
This document includes the main source code (Matlab or Python) related to our study.
Cite this work
Researchers should cite this work as follows:
- Wen, H., Shi, J., Zhang, Y., Lu, K., Cao, J., Liu, Z. (2017). Data for Neural Encoding and Decoding with Deep Learning for Dynamic Natural Vision Tests. Purdue University Research Repository. doi:10.4231/R7SF2TCW
Laboratory of Integrated Brain Imaging
This publication belongs to the Laboratory of Integrated Brain Imaging group.