Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2020 Oct 13;117(41):25505-25516.
doi: 10.1073/pnas.1917551117. Epub 2020 Oct 2.

Robust parallel decision-making in neural circuits with nonlinear inhibition

Affiliations

Robust parallel decision-making in neural circuits with nonlinear inhibition

Birgit Kriener et al. Proc Natl Acad Sci U S A. .

Abstract

An elemental computation in the brain is to identify the best in a set of options and report its value. It is required for inference, decision-making, optimization, action selection, consensus, and foraging. Neural computing is considered powerful because of its parallelism; however, it is unclear whether neurons can perform this max-finding operation in a way that improves upon the prohibitively slow optimal serial max-finding computation (which takes [Formula: see text] time for N noisy candidate options) by a factor of N, the benchmark for parallel computation. Biologically plausible architectures for this task are winner-take-all (WTA) networks, where individual neurons inhibit each other so only those with the largest input remain active. We show that conventional WTA networks fail the parallelism benchmark and, worse, in the presence of noise, altogether fail to produce a winner when N is large. We introduce the nWTA network, in which neurons are equipped with a second nonlinearity that prevents weakly active neurons from contributing inhibition. Without parameter fine-tuning or rescaling as N varies, the nWTA network achieves the parallelism benchmark. The network reproduces experimentally observed phenomena like Hick's law without needing an additional readout stage or adaptive N-dependent thresholds. Our work bridges scales by linking cellular nonlinearities to circuit-level decision-making, establishes that distributed computation saturating the parallelism benchmark is possible in networks of noisy, finite-memory neurons, and shows that Hick's law may be a symptom of near-optimal parallel decision-making with noisy input.

Keywords: neural circuits; noisy computation; optimal decision-making; speed–accuracy trade-off.

PubMed Disclaimer

Conflict of interest statement

The authors declare no competing interest.

Figures

Fig. 1.
Fig. 1.
Parallelism benchmarks and setup of the WTA circuit. (A) Schematic illustration of how log(N) integration time is sufficient to maintain fixed accuracy (here measured as number of correct trials over total number of trials). Thin lines denote sample time series from different trials for each option (N options are in different colors; here N=4). Thick curves denote sample histograms (distributions) after integrating for time T. The tails of the distribution shrink exponentially (shown by gray shading). Thus, integrating for time log(N) means a per-option error probability of 1/N, so that the total error probability, given by N times the per-option error probability, is fixed. (B) The log(N) integration time is necessary to maintain fixed accuracy when estimating the means of noisy time series with Gaussian fluctuations. Plot shows the integration time to fixed accuracy (dashed and solid lines, 80% and 99% accuracy, respectively) in (nonneural) simulations with Gaussian input fluctuations. Thin black lines are logarithmic fits. (C) WTA network architectures. (Top) The time series bi serve as input to N=4 neurons (or pools of neurons), ordered by the size of their constant mean (with gap Δb1b2 between the top two inputs). Each neuron pool inhibits all others and excites itself. (Bottom) Mathematically equivalent network with a global inhibitory neuron pool: All neuron pools excite the inhibitory neuron pool, which inhibits them. This requires only N synapses, compared to N2 for the mutual inhibition circuit (Top). (D) Neural firing rates and synaptic activations (coloring as in C). Convergence time TWTA is the time taken for the firing rate of the winner neuron pool to reach a fraction c (usually chosen as 0.8; see Methods) of its expected asymptotic activity x (black line). Dashed line denotes threshold for a neuron pool to contribute inhibition in the nonlinear inhibition model (nWTA; see Methods).
Fig. 2.
Fig. 2.
Conventional WTA networks fail the parallelism benchmark. (AC) Results for weak inhibition. (A) Activity dynamics of the most active neuron (pool) for networks of size N=10,20,40 (light to dark red), respectively, with noisy input; α=0.6, β=1, Δ=0.1, ση=0.35, τη=0.05τ. The asymptotic activity level grows with N. (B) WTA simulation results on decision time without (solid gray) and with (dashed gray) noise. Also shown are noisy serial strategy (green) and noisy parallelism benchmark (purple). Note that benchmarks are shown up to an overall constant. Error bars are smaller than the line width. Deterministic serial strategy (i.e., O(N)) is not shown, for simplicity. (C) (Top) Critical noise amplitude versus N: WTA dynamics exists below a given curve and fails above it (dashed curve, numerical simulation; solid curve, analytical). Darker curves correspond to a larger gap between the top two inputs (Δ={0.01,0.06,,0.26}, α=11/2N, τη=0.005τ). (Bottom) Average accuracy, that is, fraction of correct trials, as a function of N (simulations averaged over 150 trials). (DF) Results for strong inhibition. (D) As in A but for strong inhibition. WTA breaks down rapidly (note absence of WTA for N = 40, dark curve). (E) WTA decision time in the absence of noise (gray) along with the deterministic serial strategy (green). (F) As in C, showing critical noise amplitude versus N. Darker curves denote stronger self-excitation (Top; α={0.1,0.6,0.9,1.1}; Δ=0.1) or a widening gap (Bottom; Δ={0.01,0.06,,0.26}; α=0.6). β=1, in all curves.
Fig. 3.
Fig. 3.
The nWTA is robust to noise and achieves the parallelism benchmark. (A) The nWTA decision time without noise (gray) and the deterministic serial strategy (green). (B) Critical noise amplitude ση* as a function of N for varying α={0.5,0.7,0.9,1.1}, Δ=0.075 (Left) and Δ={0,0.0125,0.025,0.05,0.075,0.1,0.15}, α=0.5, β=0.51 (Right). Below each curve, WTA behavior exists, while, above, it does not. (C) Heat maps showing fraction of trials with a WTA solution (single winner; Left), accuracy of the WTA solution (Middle), and convergence time TWTA/τ (Right) as function of network size N and noise amplitude ση. Δ=0.05, α=0.5, and β=0.6, averaged over 1,500 trials each. (D) Speed–accuracy curves for Δ=0.075, ση=0.12, and α=0.5 for varying N={23,24,,215} (light to dark red) and β={0.51,0.52,,0.6,0.65,0.7,0.8,0.9} (light to dark gray circles): 1,500 trials, of which only trials that produced a WTA solution were included. Curves are nonmonotonic, so that certain parameters are strictly better than others for both speed and accuracy. (E) N scaling of decision time to achieve fixed accuracy of 0.99 for nWTA (gray) and the parallelism benchmark (purple). Thin black line shows logarithmic fit. Δ=0.075; τη=0.05τ; θ=0.2; ση=0.12; see SI Appendix, section S2 and Fig. S6 for similar results with different parameters and noise levels. (F) N scaling of accuracy at fixed decision time. Gray denotes nWTA for integration time 12.5τ; purple denotes parallelism benchmark for integration time 2τ.
Fig. 4.
Fig. 4.
The nWTA network self-adjusts to maintain accuracy and exhibits Hick’s law without any parameter change. (A) Accuracy for fixed parameter values as a function of N for large N. Dark and light blue show (α, β) = (0.5, 0.51), (0.6, 0.41), respectively (one accuracy trace not visible). Input parameters are as in Fig. 3 DF, 5,000 trials per data point. (B) Decision time for the parameter values shown in A, along with logarithmic fits (thin black). (C) Accuracy for fixed parameter values for small N. Blue symbols show nWTA network (light, dark show (α, β) = (0.3, 0.71), (0.6, 0.41), respectively), while purple squares show AB model with perfect integration. Δ=0.05,σ=0.2, τη=0.05τ. Inset shows selected parameter values as well as average accuracy for all parameter combinations (averaged across N=2 to 10). (D) Decision time for nWTA networks and AB model shown in C. Thin black lines show logarithmic fits for nWTA and constant fit for AB.
Fig. 5.
Fig. 5.
Self-terminating WTA dynamics as a minimal-parameter, neural model of multi-AFC decision-making. (A) Reward rate for the fixed parameters chosen to maximize reward rate (dark blue circles, α=0.41,β=0.7) along with reward rate when parameters are individually optimized for each N (light blue). Thin black lines show fit to (T0+log(N+1))1. Inset shows accuracy for fixed parameters chosen to maximize reward rate. Accuracy at highest reward rate is high but not constant, reflecting a speed–accuracy trade-off. (B) Reward rate vs. accuracy for N={2,10}. Gray, nWTA (using best α, β for given accuracy); dark purple: AB model; light purple: AB implementation of relative (max-vs-next) form of MSPRT. Inset for N=10 panel shows speed–accuracy curves for α,β that yield optimal reward rate for N=10; see SI Appendix, Fig. S8F [speed (100/τ) includes T0=300 ms; dark blue: α=0.46 fixed, β{0.55,0.6,,1}; light blue: β=0.6 fixed, α{0.41,0.46,,0.96} varied; other parameters are as follows: Δ=0.05, ση=0.2, τη=0.05τ]. (C) Example activation trajectories xi(t) for nWTA networks with different (α,β) but with α+β (i.e., integration time constant) held fixed. (Left) Lower self-excitation and higher inhibition; (Right) vice versa. (D) Example activation trajectories for network with same inhibition as C, Right, but stronger self-excitation. Trials are faster, but final activation of winner is higher. (E) Example activation trajectories for N=2,6,10 (light to dark red) with fixed parameters. Activity at the time of convergence remains constant. (F) (Top) Decision time for fixed parameters as a function of gap between largest and second largest input (related to task difficulty), showing that network self-adjusts to integrate longer for difficult tasks. Colors as in E. (Bottom) As in Top but for noise in input. Note that, in the small-N regime, conventional WTA models show similar performance to nWTA for many of these results, although less robustly; see SI Appendix, section S3 and Fig. S8.

References

    1. Franks N. R., Pratt S. C., Mallon E. B., Britton N. F., Sumpter D. J. T., Information flow, opinion polling and collective intelligence in house–hunting social insects. Philos. Trans. R. Soc. Lond. B Biol. Sci. 357, 1567–1583 (2002). - PMC - PubMed
    1. Gold J. I., Shadlen M. N., The neural basis of decision making. Annu. Rev. Neurosci. 30, 535–574 (2007). - PubMed
    1. Bogacz R., Gurney K., The basal ganglia and cortex implement optimal decision making between alternative actions. Neural Comput. 19, 442–477 (2007). - PubMed
    1. Chittka L., Skorupski P., Raine N. E., Speed–accuracy tradeoffs in animal decision making. Trends Ecol. Evol. 24, 400–407 (2009). - PubMed
    1. Stevens C. F., What the fly’s nose tells the fly’s brain. Proc. Natl. Acad. Sci. U.S.A. 112, 9460–9465 (2015). - PMC - PubMed

Publication types