Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Sep 15:7:349-365.
doi: 10.1146/annurev-vision-093019-112249. Epub 2021 Jul 16.

Remembering the Past to See the Future

Affiliations

Remembering the Past to See the Future

Nicole C Rust et al. Annu Rev Vis Sci. .

Abstract

In addition to the role that our visual system plays in determining what we are seeing right now, visual computations contribute in important ways to predicting what we will see next. While the role of memory in creating future predictions is often overlooked, efficient predictive computation requires the use of information about the past to estimate future events. In this article, we introduce a framework for understanding the relationship between memory and visual prediction and review the two classes of mechanisms that the visual system relies on to create future predictions. We also discuss the principles that define the mapping from predictive computations to predictive mechanisms and how downstream brain areas interpret the predictive signals computed by the visual system.

Keywords: memory; prediction; vision.

PubMed Disclaimer

Figures

Figure 1
Figure 1
Examples of prediction. (a) Catching a ball. Shown is the lag between the estimated position of a fast-moving ball attributed to the latency of processing in the retina (purple dashed) relative to its actual position (green). Lag compensation in both the eye and the brain allows us to accurately estimate ball position. (b) Novelty. Curiosity-based exploration is crucial for efficient learning.
Figure 2
Figure 2
The information bottleneck technique links optimal prediction and memory. The information bottleneck technique is a method for computing the maximal amount of information that a compressed signal, like the brain’s code for the visual stimulus, can carry about a relevant variable in the original input. Within our prediction framework, that relevant feature is the future stimulus. In the diagram, the input is the past stimulus measured within a window of time preceding the neural response (past), and the relevant variable is the future stimulus (future) starting at some time in the future, Δt. For a particular Δt, we can trace out the maximal amount of information that a neural population could possibly carry about the future stimulus given how much information that population encoded about the past stimulus. There are several notable regions in this information plane spanned by the past and future information (left). First (①), there is an inaccessible region: You cannot know more about the future than about the past; i.e., there is no fortune-telling. The brain’s code can occupy any other region in the plane. Second (②), sitting near the bound means that the neural code contains the maximal amount of predictive information possible for a given level of fidelity of past information. Third (③), neural responses that reflect information about the past but fall away from the bound are not optimized for prediction as a consequence of encoding unpredictable parts of the input stimulus. Fourth (④), the saturation point, reflecting the maximal information that you can glean about the future, is set by the correlation structure in the stimulus. It is important to note that memories of the past can, themselves, be faulty. The x axis in the bottleneck plots reflects precisely that fidelity. For a system with a given memory time window, neural systems can be so noisy that they carry no information about the past stimulus (the origin). Conversely, they could represent the past stimulus with high precision (e.g., with finer and finer stimulus resolution, moving outward along the x axis). Increasing the timescale for memory of the past can improve prediction, up to the limits set by the longest correlation times in the stimulus itself (e.g., dashed versus solid lines). In this example, expanding the length of the past stimulus history saturates after the memory of the past is expanded to four blocks of time in the past.
Figure 3
Figure 3
Computational proposals for prediction and memory. (a) A depiction of the initial, sensory-evoked response (gray arrows and green circles) at the first time point (t = 0). All other panels depict the processing that happens at the next time point (t = Δt) for three different models of prediction. (b) A predictive architecture in which memories of recent stimulus history are stored via adaptation (red arrows), which can take the form of gain adaptation (bottom layer, lighter green circles) or adaptation of the feedforward synapses that connect the layers. In this class of model, connectivity is exclusively feedforward. (c) A predictive architecture in which memories of recent stimulus history are maintained via persistent, recurrent activity (red arrows) and extrapolation of recent events into future predictions happens via feedback (blue arrows). (d) A predictive architecture that extrapolates the current sensory input forward in time using hippocampal pattern completion and integrates this information with incoming sensory signals via feedback (blue arrows).

References

    1. Aly M, Turk-Browne NB. 2016. Attention promotes episodic encoding by stabilizing hippocampal representations. PNAS 113:E420–29 - PMC - PubMed
    1. Aly M, Turk-Browne NB. 2017. How hippocampal memory shapes, and is shaped by, attention. In The Hippocampus from Cells to Systems: Structure, Connectivity, and Functional Contributions to Memory and Flexible Cognition, ed. Hannula DE, Duff MC, pp. 369–403. Berlin: Springer
    1. Anstis S 2007. The flash-lag effect during illusory chopstick rotation. Perception 36:1043–48 - PubMed
    1. Bastos AM, Usrey WM, Adams RA, Mangun GR, Fries P, Friston KJ. 2012. Canonical microcircuits for predictive coding. Neuron 76:695–711 - PMC - PubMed
    1. Bellemare MG, Srinivasan S, Ostrovski G, Schaul T, Saxton D, Munos R. 2016. Unifying count-based exploration and intrinsic motivation. In Advances in Neural Information Processing Systems 29 (NIPS 2016), ed. D Lee, Sugiyama M, Luxburg U, Guyon I, Garnett R, pp. 1471–79. N.p.: NeurIPS

Publication types

LinkOut - more resources