Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2016 Jan 6:9:484.
doi: 10.3389/fnins.2015.00484. eCollection 2015.

Energy Scaling Advantages of Resistive Memory Crossbar Based Computation and Its Application to Sparse Coding

Affiliations

Energy Scaling Advantages of Resistive Memory Crossbar Based Computation and Its Application to Sparse Coding

Sapan Agarwal et al. Front Neurosci. .

Abstract

The exponential increase in data over the last decade presents a significant challenge to analytics efforts that seek to process and interpret such data for various applications. Neural-inspired computing approaches are being developed in order to leverage the computational properties of the analog, low-power data processing observed in biological systems. Analog resistive memory crossbars can perform a parallel read or a vector-matrix multiplication as well as a parallel write or a rank-1 update with high computational efficiency. For an N × N crossbar, these two kernels can be O(N) more energy efficient than a conventional digital memory-based architecture. If the read operation is noise limited, the energy to read a column can be independent of the crossbar size (O(1)). These two kernels form the basis of many neuromorphic algorithms such as image, text, and speech recognition. For instance, these kernels can be applied to a neural sparse coding algorithm to give an O(N) reduction in energy for the entire algorithm when run with finite precision. Sparse coding is a rich problem with a host of applications including computer vision, object tracking, and more generally unsupervised learning.

Keywords: energy; memristor; neuromorphic computing; resistive memory; sparse coding.

PubMed Disclaimer

Figures

Figure 1
Figure 1
Analog resistive memories can be used to reduce the energy of a vector-matrix multiply. The conductance of each resistive memory represents a weight. Analog input values are represented by the input voltages or input pulse lengths, and outputs are represented by current values. This allows all the read operations, multiplication operations, and sum operations to occur in a single step. A conventional architecture must perform these operations sequentially for each weight resulting in a higher energy and delay.
Figure 2
Figure 2
A typical SRAM array. Each row/wordline must be accessed sequentially.
Figure 3
Figure 3
A parallel write is illustrated. Weight Wij is updated by xi × yj. In order to achieve a multiplicative effect the xi are encoded in time while the yi are encoded in the height of a voltage pulse. The resistive memory will only train when xi is non-zero. The height of yi determines the strength of training when xi is non-zero.
Figure 4
Figure 4
A system would consist of individual crossbar based cores that communicate with each other through a communications bus and routing system.
Figure 5
Figure 5
If the data is not the same shape as the array, the input data will come from a single router, and the output data will need to go to a single computation unit. At best, the extra wire length to go to the input/output units plus the row/column wire length will be O[max(N,M)].

References

    1. Arora S., Ge R., Ma T., Moitra A. (2015). Simple, efficient, and neural algorithms for sparse coding. arXiv preprint arXiv:1503.00778. - PubMed
    1. Burr G. W., Shelby R. M., Sidler S., di Nolfo C., Jang J., Boybat I., et al. (2015). Experimental demonstration and tolerancing of a large-scale neural network (165 000 Synapses) using phase-change memory as the synaptic weight element. IEEE Trans. Electron Dev. 62, 3498–3507. 10.1109/TED.2015.2439635 - DOI
    1. Cassidy A. S., Alvarez-Icaza R., Akopyan F., Sawada J., Arthur J. V., Merolla P. A., et al. (2014). Real-time scalable cortical computing at 46 giga-synaptic OPS/watt with ~100 × Speedup in Time-to-Solution and ~100,000 × reduction in energy-to-solution, in Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (New Orleans, LA: IEEE Press; ), 27–38.
    1. Chen A. (2013). A comprehensive crossbar array model with solutions for line resistance and nonlinear device characteristics. IEEE Trans. Electron Dev. 60, 1318–1326. 10.1109/TED.2013.2246791 - DOI
    1. Chen H.-Y., Gao B., Li H., Liu R., Huang P., Chen Z., et al. (2014). Towards high-speed, write-disturb tolerant 3D vertical RRAM arrays, in Digest of Technical Papers, 2014 Symposium on: IEEE VLSI Technology (VLSI-Technology) (Honolulu, HI: ), 1–2.