Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published October 2021 | Accepted Version + Supplemental Material
Journal Article Open

Cortical networks of dynamic scene category representation in the human brain

Abstract

Humans have an impressive ability to rapidly process global information in natural scenes to infer their category. Yet, it remains unclear whether and how scene categories observed dynamically in the natural world are represented in cerebral cortex beyond few canonical scene-selective areas. To address this question, here we examined the representation of dynamic visual scenes by recording whole-brain blood oxygenation level-dependent (BOLD) responses while subjects viewed natural movies. We fit voxelwise encoding models to estimate tuning for scene categories that reflect statistical ensembles of objects and actions in the natural world. We find that this scene-category model explains a significant portion of the response variance broadly across cerebral cortex. Cluster analysis of scene-category tuning profiles across cortex reveals nine spatially-segregated networks of brain regions consistently across subjects. These networks show heterogeneous tuning for a diverse set of dynamic scene categories related to navigation, human activity, social interaction, civilization, natural environment, non-human animals, motion-energy, and texture, suggesting that the organization of scene category representation is quite complex.

Additional Information

© 2021 Elsevier Ltd. Received 22 September 2020, Revised 28 June 2021, Accepted 14 July 2021, Available online 24 July 2021. Reviewed 27 May 2021; Action editor Ana B. Chica. The authors declare no competing financial interests. The work was supported in part by a National Eye Institute Grant (EY019684), by a Marie Curie Actions Career Integration Grant (PCIG13-GA-2013-618101), by a European Molecular Biology Organization Installation Grant (IG 3028), by a TUBA GEBIP 2015 fellowship, and by a Science Academy BAGEP 2017 award. We thank D. Stansbury, A. Huth, and S. Nishimoto for assistance in various aspects of this research. We report how we determined our sample size, all data exclusions, all inclusion/exclusion criteria, whether inclusion/exclusion criteria were established prior to data analysis, all manipulations, and all measures in the study. No part of the study procedures or analyses was pre-registered prior to the research being conducted. CRediT author statement: Emin Çelik: Conceptualization, Methodology, Software, Validation, Formal analysis, Writing – Review & Editing, Visualization. Umit Keles: Conceptualization, Methodology, Software, Formal analysis, Writing – Original Draft, Visualization. İbrahim Kiremitçi: Validation, Formal analysis. Jack L. Gallant: Writing – Review & Editing. Tolga Çukur: Conceptualization, Investigation, Resources, Writing – Review & Editing, Supervision, Project Administration, Funding Acquisition. Open practices: The study in this article earned Open Data and Open Materials badges for transparent practices. Data for this study can be found at https://crcns.org/data-sets/vc/vim-2 and https://crcns.org/data-sets/vc/vim-4.

Attached Files

Accepted Version - 1-s2.0-S0010945221002549-main.pdf

Supplemental Material - 1-s2.0-S0010945221002549-mmc1.pdf

Files

1-s2.0-S0010945221002549-main.pdf
Files (3.6 MB)
Name Size Download all
md5:2fdf86c8f1c943f6899ac3003b8a778c
3.0 MB Preview Download
md5:cf2403ae6d014a9074431a33fa1c4146
558.2 kB Preview Download

Additional details

Created:
August 22, 2023
Modified:
October 23, 2023