Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2019 May 28:13:28.
doi: 10.3389/fnbot.2019.00028. eCollection 2019.

Neuromorphic Stereo Vision: A Survey of Bio-Inspired Sensors and Algorithms

Affiliations
Review

Neuromorphic Stereo Vision: A Survey of Bio-Inspired Sensors and Algorithms

Lea Steffen et al. Front Neurorobot. .

Abstract

Any visual sensor, whether artificial or biological, maps the 3D-world on a 2D-representation. The missing dimension is depth and most species use stereo vision to recover it. Stereo vision implies multiple perspectives and matching, hence it obtains depth from a pair of images. Algorithms for stereo vision are also used prosperously in robotics. Although, biological systems seem to compute disparities effortless, artificial methods suffer from high energy demands and latency. The crucial part is the correspondence problem; finding the matching points of two images. The development of event-based cameras, inspired by the retina, enables the exploitation of an additional physical constraint-time. Due to their asynchronous course of operation, considering the precise occurrence of spikes, Spiking Neural Networks take advantage of this constraint. In this work, we investigate sensors and algorithms for event-based stereo vision leading to more biologically plausible robots. Hereby, we focus mainly on binocular stereo vision.

Keywords: bio-inspired 3D-perception; brain-inspired robotics; cooperative algorithms; event-based technologies; human-like vision; neuromorphic visual sensors.

PubMed Disclaimer

Figures

Figure 1
Figure 1
The correspondence problem. The scene comprises four identical objects, recorded from two perspectives. R is the right and L the left camera. The rectangles L1L4 represent the imaging of L and R1R4 of R. There are 16 possible matches, visualized by dots, showing how the shots of L and R might correspond to each other. Only four of them, accentuated in red, are correct matches.
Figure 2
Figure 2
The human retina, reduced to essential layers for neuromorphic visual sensors. The photoreceptor layer, the outer plexiform layer including bipolar cells and the inner plexiform layer made up of ganglion cells. Additionally, horizontal cells and amacrine cells connect these layers.
Figure 3
Figure 3
The three components of Mahowalds silicon retina are modeled on photoreceptor cells, bipolar cells, and horizontal cells. Every module is marked by the first letter of its biological paragon.
Figure 4
Figure 4
Artificial building blocks and their biological models of the Parvo-Magno Retina from Zaghloul and Boahen. The left circuit shows the outer retinal layer. A phototransistor takes current via an nMOS transistor. Its source is connected to Vc, representing the biological photoreceptor (P). Its gate is connected to Vh, portraying horizontal cells (H). The circuit in the center represents the amacrine cell modulation. A bipolar terminal B excites a network of wide-field amacrine cells (WA) and also narrow-field amacrine cells (NA) through a current mirror. The amacrine cells, in turn have an inhibitory effect on B. By the right circuit the spiking ganglion cells are represented. Current Iin from the inner retinal circuit charges up a membrane capacitor G, based on biological ganglion cells. If its membrane voltage crosses a threshold a spike (Sp) is emitted and a reset (Rst) discharges the membrane. Inspired by Zaghloul and Boahen (2006).
Figure 5
Figure 5
The AER-bus-system. Three neurons on the sending chip produce spikes [see (I)]. These are interpreted as binary events [see (II)] and by means of the address encoder (AE), a binary address is generated. This address is transmitted via the bus-system and the address decoder (AD) determines the correct position on the receiving chip [see (III)]. Hence a spike is emitted on the affected neuron of the receiver [see (IV)].
Figure 6
Figure 6
The three-layered pixel circuit of the DVS, consisting of a photoreceptor, inspired by the biological cones, a differential circuit, based on the bipolar cell and a comparator modeled after the ganglion cell.
Figure 7
Figure 7
The two-section circuit constituting each ATIS pixel. The change-detector circuit (CD), which is also part of the DVS, is supplemented in the ATIS by the exposure measurement circuit (EM). In this way the camera is able to obtain, additionally to transient, also sustainable image information. Written informed consent for publication of the image was obtained from the individual in that graphic.
Figure 8
Figure 8
The circuits building up the DAVIS. Each of its pixels is a hybrid of the DVS-circuit and an active-pixel-sensor (APS). Like the ATIS sensor, the additional component of the DAVIS is capable of generating gray-scale images. However, the principle of operation of the APS is synchronous and thus similar to conventional vision sensors, distinguishing both sensors severely.
Figure 9
Figure 9
Event streams of two sensors for the same point of the real world. The top row shows the deviation of illumination between the two corresponding pixels of two retinas. The resulting streams of OFF and ON events are shown below. Below it can be seen that events of both sensors occurring within the timeframe δt can be both, correct and incorrect correspondences.
Figure 10
Figure 10
Structure of the network inspired by Dikov et al. (2017) with increasing degree of detail. The network's 3D-structure is represented in the left part; An entire row of pixels is mapped from both EBS to one plane of the SNN. This 2D-layer is where the disparities of the affected pixels are calculated. The center shows the neurons of a 2D-layer, connected according to the constraints of cooperative algorithms, outlines in chapter 4.1.1. Green represents inhibiting and blue exciting synapses. In the right part the outline of micro ensembles are visualized. The cooperative manner of this network relies on micro-ensembles. The retina coordinates are visualized as light blue squares, the blocking neurons as blue circles and the collecting neuron as a red circle. Reprinted by permission from Springer Artificial Neural Networks and Machine Learning (Kaiser et al., 2018).

Similar articles

Cited by

References

    1. Adachi-Usami E., Lehmann D. (1983). Monocular and binocular evoked average potential field topography: upper and lower hemiretinal stimuli. Exp. Brain Res. 50, 341–346. 10.1007/BF00239198 - DOI - PubMed
    1. Akolkar H., Meyer C., Clady X., Marre O., Bartolozzi C., Panzeri S., et al. . (2015). What can neuromorphic event-driven precise timing add to spike-based pattern recognition? Neural Comput. 27, 561–593. 10.1162/NECO_a_00703 - DOI - PubMed
    1. Andreopoulos A., Kashyap H. J., Nayak T. K., Amir A., Flickner M. D. (2018). A low power, high throughput, fully event-based stereo system, in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). (Salt Lake City, UT: ), 7532–7542.
    1. Balasubramanian R., Gan L. (2014). Development of retinal amacrine cells and their dendritic stratification. Curr. Ophthalmol. Rep. 2, 100–106. 10.1007/s40135-014-0048-2 - DOI - PMC - PubMed
    1. Barnard S. T., Fischler M. A. (1982). Computational stereo. ACM Comput. Surv. 14, 553–572. 10.1145/356893.356896 - DOI

LinkOut - more resources