Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Oct-Dec;38(7-8):425-439.
doi: 10.1080/02643294.2022.2034609. Epub 2022 Feb 13.

A role for visual areas in physics simulations

Affiliations

A role for visual areas in physics simulations

Aarit Ahuja et al. Cogn Neuropsychol. 2021 Oct-Dec.

Abstract

To engage with the world, we must regularly make predictions about the outcomes of physical scenes. How do we make these predictions? Recent computational evidence points to simulation-the idea that we can introspectively manipulate rich, mental models of the world-as one explanation for how such predictions are accomplished. However, questions about the potential neural mechanisms of simulation remain. We hypothesized that the process of simulating physical events would evoke imagery-like representations in visual areas of those same events. Using functional magnetic resonance imaging, we find that when participants are asked to predict the likely trajectory of a falling ball, motion-sensitive brain regions are activated. We demonstrate that this activity, which occurs even though no motion is being sensed, resembles activity patterns that arise while participants perceive the ball's motion. This finding thus suggests that mental simulations recreate sensory depictions of how a physical scene is likely to unfold.

Keywords: Simulation; fMRI; imagery; intuitive physics.

PubMed Disclaimer

Conflict of interest statement

Disclosure Statement

No potential competing interest was reported by the authors

Figures

Figure 1:
Figure 1:. Task Design.
(A) An example of a board that constituted the primary stimulus in the ball fall task. Participants had to determine which of the two catchers the ball would land in if dropped (B) A schematic depicting the blocked design of the task variants (Simulation, Perception, Control, and Native), as well as the internal composition of a block. (C) A schematic depicting the trial outlines for each of the three variants of interest. The Native variant was not included in the subsequent fMRI analyses and is hence not shown here.
Figure 2:
Figure 2:. Board Designations for Behavioral Analyses.
(A) An example of a board on which the ball only hit two planks. Such boards were classified as having a short simulation length. (B) An example of a board on which the ball hit four planks. Such boards were classified as having a long simulation length. (C) An example of a board where slightly jittering the position of each plank (four jittered examples are shown to the right) had a minimal impact on the ball’s final position. Such boards were classified as having a low simulation uncertainty. (D) An example of a board where slightly jittering the position of each plank (four jittered examples are shown to the right) greatly impacted the ball’s final position. Such boards were classified as having a high simulation uncertainty.
Figure 3:
Figure 3:. fMRI Analysis Pipeline.
(A) A schematic of the motion localizer task. The display alternated between blocks of moving and static dots, flanked by periods of fixation. (B) A hypothetical activation map of motion-sensitive voxels derived from a Motion > Static localizer contrast, plotted as a 3D point cloud to demonstrate ROI selection. (C) An example participant’s t-values in the ROI from (B) for each of the three task conditions, contrasted to baseline. Based on these t-maps, we assessed whether the voxel-wise representation in the Simulation condition was more similar to the Perception condition (S-P) or the Control condition (S-C). (D) A line graph showing an example comparison of S-P and S-C similarities. The participant shown in (C) is highlighted in color in (D), and the grey lines represent a hypothetical group effect if the analysis were repeated for all 12 participants.
Figure 4:
Figure 4:. Behavioral Results.
(A) Participants’ mean task accuracy across the three variants of interest. Participants were extremely good at all three variants. (B) Participants’ mean reaction times as a function of simulation length and uncertainty. We found that simulation length and simulation uncertainty affected participants’ reaction times on the task. (C) Participants’ average accuracy on the task as a function of simulation length and uncertainty. We found that simulation uncertainty affected participants’ reaction times, whereas simulation length did not. Both of these behavioral effects were previously reported in Ahuja & Sheinberg, 2019. Error bars in all figures represent the standard error of the mean performance for the twelve subjects.
Figure 5:
Figure 5:. Localizer Results.
Activation maps for a second level Motion > Static contrast at a p < 0.05 threshold (family-wise error [FWE] cluster corrected for multiple comparisons, extent threshold 187). We observed several canonically motion sensitive regions such as area MT and PPC in this contrast. These voxels were used to define ROIS for subsequent RSA analyses.
Figure 6:
Figure 6:. RSA Results.
(A) Pairwise comparisons of S-P and S-C representational similarities in a motion-sensitive ROI. Each pair of points represents one participant. We found that for all participants, the representational similarity between the Simulation and Perception conditions was greater (evidenced by a higher Spearman correlation) than the representational similarity between the Simulation and Control conditions (B) The same analysis as in (A), repeated for an MT ROI. (C) The same analysis as in (A), repeated for a PPC ROI.
Figure 7:
Figure 7:. Searchlight Results.
Clusters of voxels that were highlighted by a searchlight analysis for consistently exhibiting the main effect from Figure 5. The searchlight largely revealed the same regions as we had previously isolated using the motion localizer task (slices here are the same as in Figure 4).

References

    1. Ahuja A, & Sheinberg DL (2019). Behavioral and oculomotor evidence for visual simulation of object movement. Journal of Vision, 19(6), 13–13. 10.1167/19.6.13 - DOI - PMC - PubMed
    1. Albright TD (1984). Direction and orientation selectivity of neurons in visual area MT of the macaque. Journal of Neurophysiology, 52(6), 1106–1130. doi: 10.1152/jn.1984.52.6.1106 - DOI - PubMed
    1. Ballard DH, Hayhoe MM, Pook PK, & Rao RP (1997). Deictic codes for the embodiment of cognition. Behavioral and Brain Sciences, 20(4). - PubMed
    1. Bates CJ, Yildirim I, Tenenbaum JB, & Battaglia P (2019). Modeling human intuitions about liquid flow with particle-based simulation. PLOS Computational Biology, 15(7), e1007210. 10.1371/journal.pcbi.1007210 - DOI - PMC - PubMed
    1. Battaglia PW, Hamrick JB, & Tenenbaum JB (2013). Simulation as an engine of physical scene understanding. Proceedings of the National Academy of Sciences, 110(45), 18327–18332. 10.1073/pnas.1306572110 - DOI - PMC - PubMed

Publication types