<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Neurosci.</journal-id>
<journal-title>Frontiers in Computational Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5188</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fncom.2018.00074</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Optimal Localist and Distributed Coding of Spatiotemporal Spike Patterns Through STDP and Coincidence Detection</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Masquelier</surname> <given-names>Timoth&#x000E9;e</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/21932/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Kheradpisheh</surname> <given-names>Saeed R.</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/300123/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Centre de Recherche Cerveau et Cognition, UMR5549 CNRS&#x02014;Universit&#x000E9; Toulouse 3</institution>, <addr-line>Toulouse</addr-line>, <country>France</country></aff>
<aff id="aff2"><sup>2</sup><institution>Instituto de Microelectr&#x000F3;nica de Sevilla (IMSE-CNM), CSIC, Universidad de Sevilla</institution>, <addr-line>Sevilla</addr-line>, <country>Spain</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Computer Science, Faculty of Mathematical Sciences and Computer, Kharazmi University</institution>, <addr-line>Tehran</addr-line>, <country>Iran</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Anthony N. Burkitt, The University of Melbourne, Australia</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Thomas Wennekers, Plymouth University, United Kingdom; Florence Isabelle Kleberg, Frankfurt Institute for Advanced Studies, Germany</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Timoth&#x000E9;e Masquelier <email>timothee.masquelier&#x00040;cnrs.fr</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>18</day>
<month>09</month>
<year>2018</year>
</pub-date>
<pub-date pub-type="collection">
<year>2018</year>
</pub-date>
<volume>12</volume>
<elocation-id>74</elocation-id>
<history>
<date date-type="received">
<day>26</day>
<month>06</month>
<year>2018</year>
</date>
<date date-type="accepted">
<day>17</day>
<month>08</month>
<year>2018</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2018 Masquelier and Kheradpisheh.</copyright-statement>
<copyright-year>2018</copyright-year>
<copyright-holder>Masquelier and Kheradpisheh</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Repeating spatiotemporal spike patterns exist and carry information. Here we investigated how a single spiking neuron can optimally respond to one given pattern (localist coding), or to either one of several patterns (distributed coding, i.e., the neuron&#x00027;s response is ambiguous but the identity of the pattern could be inferred from the response of multiple neurons), but not to random inputs. To do so, we extended a theory developed in a previous paper (Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>), which was limited to localist coding. More specifically, we computed analytically the signal-to-noise ratio (SNR) of a multi-pattern-detector neuron, using a threshold-free leaky integrate-and-fire (LIF) neuron model with non-plastic unitary synapses and homogeneous Poisson inputs. Surprisingly, when increasing the number of patterns, the SNR decreases slowly, and remains acceptable for several tens of independent patterns. In addition, we investigated whether spike-timing-dependent plasticity (STDP) could enable a neuron to reach the theoretical optimal SNR. To this aim, we simulated a LIF equipped with STDP, and repeatedly exposed it to multiple input spike patterns, embedded in equally dense Poisson spike trains. The LIF progressively became selective to every repeating pattern with no supervision, and stopped discharging during the Poisson spike trains. Furthermore, tuning certain STDP parameters, the resulting pattern detectors were optimal. Tens of independent patterns could be learned by a single neuron using a low adaptive threshold, in contrast with previous studies, in which higher thresholds led to localist coding only. Taken together these results suggest that coincidence detection and STDP are powerful mechanisms, fully compatible with distributed coding. Yet we acknowledge that our theory is limited to single neurons, and thus also applies to feed-forward networks, but not to recurrent ones.</p></abstract>
<kwd-group>
<kwd>neural coding</kwd>
<kwd>localist coding</kwd>
<kwd>distributed coding</kwd>
<kwd>coincidence detection</kwd>
<kwd>leaky integrate-and-fire neuron</kwd>
<kwd>spatiotemporal spike pattern</kwd>
<kwd>unsupervised learning</kwd>
<kwd>spike-timing-dependent plasticity (STDP)</kwd>
</kwd-group>
<counts>
<fig-count count="7"/>
<table-count count="1"/>
<equation-count count="5"/>
<ref-count count="61"/>
<page-count count="10"/>
<word-count count="7807"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>In a neural network, either biological or artificial, two forms of coding can be used: localist or distributed. With localist coding, each neuron codes (i.e., maximally responds) for one and only one category of stimulus (or stimulus feature). As a result, the category of the stimulus (or the presence of a certain feature) can be inferred from the response of this sole neuron, ignoring the other neurons&#x00027; responses. Conversely, with distributed coding each neuron responds to multiple stimulus categories (or features) in a similar way. Therefore, the response of each neuron is ambiguous, and the category of the stimulus, or the presence of a certain feature, can only be inferred from the responses of multiple neurons. Thus the distinction between the two schemes is the number of different stimuli to which a given neuron responds&#x02014;not the number of neurons which respond to a given stimulus. Indeed, a localist network can have redundancy, and use multiple &#x0201C;copies&#x0201D; of each category specific neuron (Thorpe, <xref ref-type="bibr" rid="B54">1989</xref>; Bowers, <xref ref-type="bibr" rid="B3">2009</xref>).</p>
<p>Does the brain use localist or distributed coding? This question has been, and still is, intensively debated. In practice, discriminating between the two schemes from electrophysiological recordings is tricky (Quian Quiroga and Kreiman, <xref ref-type="bibr" rid="B46">2010</xref>), since the set of tested stimuli is always limited, the responses are noisy, the thresholds are arbitrary and the boundaries between categories are fuzzy. Here we do not attempt to do a complete review of the experimental literature; but rather to summarize it. It is commonly believed that distributed coding is prevalent (Rolls et al., <xref ref-type="bibr" rid="B48">1997</xref>; O&#x00027;Reilly, <xref ref-type="bibr" rid="B42">1998</xref>; Hung et al., <xref ref-type="bibr" rid="B19">2005</xref>; Quiroga et al., <xref ref-type="bibr" rid="B47">2008</xref>), but there is also evidence for localist coding, at least for familiar stimuli, reviewed in Bowers (<xref ref-type="bibr" rid="B3">2009</xref>, <xref ref-type="bibr" rid="B4">2017</xref>), Thorpe (<xref ref-type="bibr" rid="B55">2009</xref>, <xref ref-type="bibr" rid="B56">2011</xref>), and Roy (<xref ref-type="bibr" rid="B50">2017</xref>).</p>
<p>The question of localist vs. distributed coding is also relevant for artificial neural networks, and in particular for the recently popular deep neural networks. Most of the time, these networks are trained in a supervised manner, using the backpropagation algorithm (LeCun et al., <xref ref-type="bibr" rid="B30">2015</xref>). The last layer contains exactly one neuron per category, and backpropagation forces each neuron to respond more strongly when the stimulus belongs to the neuron&#x00027;s category. In other words, localist coding is imposed in the last layer. Conversely, the hidden layers are free to choose their coding scheme, which is supposedly optimal for the categorization task at hand. It is thus very interesting to analyze the chosen coding scheme. It is not easy to do such analysis on the brain (as explained above), but we can do it rigorously for computational models by computing the responses to huge amounts of images, and even synthesizing images that maximize the responses. Results indicate that some hidden neurons respond to one object category only (Zhou et al., <xref ref-type="bibr" rid="B61">2015</xref>; Nguyen et al., <xref ref-type="bibr" rid="B40">2016</xref>; Olah et al., <xref ref-type="bibr" rid="B41">2017</xref>), while others respond to multiple different objects (Nguyen et al., <xref ref-type="bibr" rid="B40">2016</xref>; Olah et al., <xref ref-type="bibr" rid="B41">2017</xref>). Thus it appears that both localist and distributed codes can be optimal, depending on the task, the layer number, and the network parameters (number of layers, neurons, etc.).</p>
<p>Let us come back to the brain, in which computation is presumably implemented by spiking neurons performing coincidence detection (Abeles, <xref ref-type="bibr" rid="B1">1982</xref>; K&#x000F6;nig et al., <xref ref-type="bibr" rid="B28">1996</xref>; Brette, <xref ref-type="bibr" rid="B6">2015</xref>). This observation raises an important question, which we tried to address in this theoretical paper: can coincidence detector neurons implement both localist and distributed codes? In this context, different stimuli correspond to different spatiotemporal input spike patterns. Here each pattern was generated randomly, leading to chance-level overlap between patterns. In addition, each pattern was jittered at each presentation, resulting in categories of similar, yet different, stimuli. Can a neuron respond to one, or several of these patterns, and not to random inputs? What is the required connectivity to do so in an optimal way? And finally, can this required connectivity emerge with spike-timing-dependent plasticity (STDP), in an unsupervised manner?</p>
<p>To address these questions, we extended a theory that we developed in a previous paper, but which was limited to one pattern only, i.e., localist coding (Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>), to the multi-pattern case. Briefly, we derived analytically the signal-to-noise ratio (SNR) of a multi-pattern detector, and investigated the conditions for its optimality. In addition, using numerical simulations, we showed that a single neuron equipped with STDP can become selective to multiple repeating spike patterns, even without supervision and that the resulting detectors can be close to the theoretical optimum. Surprisingly, a single neuron could robustly learn up to &#x0007E;40 independent patterns (using parameters arguably in the biological range). This was not clear from previous simulations studies, in which neurons equipped with STDP only learned one pattern (localist coding) (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>; Humble et al., <xref ref-type="bibr" rid="B18">2012</xref>; Hunzinger et al., <xref ref-type="bibr" rid="B20">2012</xref>; Kasabov et al., <xref ref-type="bibr" rid="B21">2013</xref>; Klampfl and Maass, <xref ref-type="bibr" rid="B26">2013</xref>; Nessler et al., <xref ref-type="bibr" rid="B39">2013</xref>; Krunglevicius, <xref ref-type="bibr" rid="B29">2015</xref>; Sun et al., <xref ref-type="bibr" rid="B53">2016</xref>; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>), or two patterns (Yger et al., <xref ref-type="bibr" rid="B60">2015</xref>). This shows that STDP and coincidence detection are compatible with distributed coding.</p>
</sec>
<sec id="s2">
<title>2. Formal description of the problem</title>
<p>The problem we addressed is similar to the one of Masquelier (<xref ref-type="bibr" rid="B32">2017</xref>), but extended to the multi-pattern case. For the reader&#x00027;s convenience, we fully describe it below.</p>
<p>We addressed the problem of detecting one or several spatiotemporal spike patterns with a single LIF neuron. Intuitively, one should connect the neurons that are active during the patterns (or during subsections of them) to the LIF neuron. That way, the LIF will tend to be more activated by the patterns than by some other inputs. More formally, we note <italic>P</italic> the number of spike patterns, and assume that they all have the same duration <italic>L</italic>. We note <italic>N</italic> the number of neurons involved. For each pattern, we chose a subsection with duration &#x00394;<italic>t</italic> &#x02264; <italic>L</italic>, and we connect the LIF to the <italic>M</italic> neurons that emit at least one spike during at least one of these subsections (Figure <xref ref-type="fig" rid="F1">1</xref>).</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(Top)</bold> <italic>P</italic> &#x0003D; 2 repeating spike patterns (colored rectangles) with duration <italic>L</italic>, embedded in Poisson noise. The LIF is connected to the neurons that fire in some subsections of the patterns with duration &#x00394;<italic>t</italic> &#x02264; <italic>L</italic> (these emit red spikes). <bold>(Bottom)</bold> The LIF potential peaks for patterns, and the double arrow indicates the peak height.</p></caption>
<graphic xlink:href="fncom-12-00074-g0001.tif"/>
</fig>
<p>We hypothesize that all afferent neurons fire according to a homogeneous Poisson process with rate <italic>f</italic>, both inside and outside the patterns. That is the patterns correspond to some realizations of the Poisson process, which can be repeated (this is sometimes referred to a &#x0201C;frozen noise&#x0201D;). At each repetition a random time lag (jitter) is added to each spike, drawn from a uniform distribution over [&#x02212;<italic>T, T</italic>] (a normal distribution is more often used, but it would not allow analytical treatment, Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>).</p>
<p>We also assume that synapses are instantaneous, which facilitates the analytic calculations.</p>
<p>For now we ignore the LIF threshold, and we want to optimize its signal-to-noise ratio (SNR), defined as:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>V</italic><sub>max</sub> is the maximal potential reached during the pattern presentations, <inline-formula><mml:math id="M2"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> is the mean value for the potential with Poisson input (noise period), and <italic>&#x003C3;</italic><sub>noise</sub> is its standard deviation. Obviously, a higher <italic>SNR</italic> means a larger difference between the LIF membrane potential during the noise periods and its maximum value, which occurs during the selected &#x00394;<italic>t</italic> window of each pattern. Therefore, the higher the <italic>SNR</italic> the lower the probability of missing patterns, and of false alarms.</p>
<p>We consider that <italic>P</italic>, <italic>L</italic>, <italic>N</italic>, <italic>f</italic>, and <italic>T</italic> are imposed variables, and that we have the freedom to choose &#x00394;<italic>t</italic> &#x02264; <italic>L</italic> and the membrane time constant &#x003C4; in order to maximize the <italic>SNR</italic>.</p>
<p>We note that this problem is related to the synfire chain theory (Abeles, <xref ref-type="bibr" rid="B2">1991</xref>). A synfire chain consists of a series of pools of neurons linked together in a feed-forward chain, so that volleys of synchronous spikes can propagate from pool to pool in the chain. Each neuron can participate in several of such chains. The number of different chains that can coexist in a network of a given size has been termed capacity. This capacity can be optimized (Herrmann et al., <xref ref-type="bibr" rid="B16">1995</xref>). To do so, a given neuron should respond to certain spike volleys, but not to others, which is similar to our optimization of a multi-pattern <italic>SNR</italic>. Yet it is also different: we use homogeneous Poisson activity, not spike volleys, and we ignore the threshold, while synfire chains require thresholds.</p>
</sec>
<sec id="s3">
<title>3. A theoretical optimum</title>
<sec>
<title>3.1. Deriving the SNR analytically</title>
<p>Here we are to find the optimum <italic>SNR</italic> of the LIF for <italic>P</italic> patterns. To this end we should first calculate the <italic>SNR</italic> analytically. Again, the derivations are similar to the ones in Masquelier (<xref ref-type="bibr" rid="B32">2017</xref>), but extended to the multi-pattern case (which turned to mainly impact Equation 7).</p>
<p>In this section, we assume non-plastic unitary synaptic weights. That is an afferent can be either connected (<italic>w</italic> &#x0003D; 1) or disconnected (<italic>w</italic> &#x0003D; 0) [in the <xref ref-type="supplementary-material" rid="SM1">Appendix</xref> (Supplementary Material) we estimate the cost of this constraint on the <italic>SNR</italic>]. Thus the LIF obeys the following differential equation:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>&#x003C4;</mml:mi><mml:mfrac><mml:mrow><mml:mtext>d</mml:mtext><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mtext>d</mml:mtext><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>V</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>&#x003B4;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where the <italic>t</italic><sub><italic>i</italic></sub> are the presynaptic spike times of all the connected afferents.</p>
<p>Since synapses are instantaneous and firing is Poissonian, during the noise periods and outside the &#x00394;<italic>t</italic> windows we have: <inline-formula><mml:math id="M4"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mi>f</mml:mi><mml:mi>M</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M5"><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mrow><mml:mi>&#x003C4;</mml:mi><mml:mi>f</mml:mi><mml:mi>M</mml:mi><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msqrt></mml:math></inline-formula> (Burkitt, <xref ref-type="bibr" rid="B8">2006</xref>), where <italic>M</italic> is the number of connected input neurons (with unitary weights).</p>
<p>To compute <italic>V</italic><sub>max</sub>, it is convenient to introduce the reduced variable:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x0221E;</mml:mi></mml:mrow></mml:msup><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M7"><mml:msup><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x0221E;</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mi>r</mml:mi></mml:math></inline-formula> is the mean potential of the steady regime that would be reached if &#x00394;<italic>t</italic> was infinite, and <italic>r</italic> is the input spike rate during the &#x00394;<italic>t</italic> window, resulting from the total received spikes from all input neurons during this window.</p>
<p><italic>v</italic><sub>max</sub> can be calculated by exact integration of the LIF differential equation (Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>). Here we omit the derivation and present the final equation:</p>
<disp-formula id="E6"><label>(4)</label><mml:math id="M100"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>min</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x00394;</mml:mi><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mi>T</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>&#x003C4;</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>T</mml:mi></mml:mrow></mml:mfrac><mml:mi>log</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mi>max</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>&#x00394;</mml:mi><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mi>T</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>/</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mo>&#x0007C;</mml:mo><mml:mi>&#x00394;</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn><mml:mi>T</mml:mi><mml:mo>&#x0007C;</mml:mo><mml:mo>/</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Using the definition of <italic>v</italic><sub>max</sub> in Equation (3), we can rewrite the <italic>SNR</italic> equation as:</p>
<disp-formula id="E7"><label>(5)</label><mml:math id="M101"><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mfrac><mml:mrow><mml:msup><mml:mover accent='true'><mml:mi>V</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mi>&#x0221E;</mml:mi></mml:msup><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mover accent='true'><mml:mi>V</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula>
<p>Obviously, different Poisson pattern realizations will lead to different values for <italic>M</italic> and <italic>r</italic> that consequently affect each of the terms <inline-formula><mml:math id="M8"><mml:msup><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x0221E;</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, <inline-formula><mml:math id="M9"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula>, and <italic>&#x003C3;</italic><sub>noise</sub>. Here we want to compute the expected value of the <italic>SNR</italic> across different Poisson pattern realizations:</p>
<disp-formula id="E8"><label>(6)</label><mml:math id="M102"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mo>&#x02329;</mml:mo><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo>&#x02329;</mml:mo> <mml:mrow><mml:mfrac><mml:mrow><mml:msup><mml:mover accent='true'><mml:mi>V</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mi>&#x0221E;</mml:mi></mml:msup><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mover accent='true'><mml:mi>V</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mrow> <mml:mo>&#x0232A;</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:msqrt><mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x003C4;</mml:mi><mml:mo>/</mml:mo><mml:mi>f</mml:mi></mml:mrow></mml:msqrt><mml:mrow><mml:mo>&#x02329;</mml:mo> <mml:mrow><mml:mfrac><mml:mrow><mml:mi>r</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mi>f</mml:mi><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:msqrt><mml:mi>M</mml:mi></mml:msqrt></mml:mrow></mml:mfrac></mml:mrow> <mml:mo>&#x0232A;</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>&#x02248;</mml:mo><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow></mml:msub><mml:msqrt><mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x003C4;</mml:mi><mml:mo>/</mml:mo><mml:mi>f</mml:mi></mml:mrow></mml:msqrt><mml:mfrac><mml:mrow><mml:mo>&#x02329;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mi>f</mml:mi><mml:mo>&#x02329;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x0232A;</mml:mo></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mo>&#x02329;</mml:mo><mml:mi>M</mml:mi></mml:mrow></mml:msqrt><mml:mo>&#x0232A;</mml:mo></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In section 3.2 we justify this last approximation through numerical simulations, and we also show that this average <italic>SNR</italic> is not much different from the <italic>SNR</italic> of particular Poisson realizations.</p>
<p>The last step to compute &#x0232A;<italic>SNR</italic>&#x02329; in Equation 6 is to calculate &#x0232A;<italic>M</italic>&#x02329; and &#x0232A;<italic>r</italic>&#x02329;. Since firing is Poissonian with rate &#x003BB; &#x0003D; <italic>f&#x00394;t</italic>, the probability that a given afferent fires at least once in a given pattern subsection of length &#x00394;<italic>t</italic> is <italic>p</italic> &#x0003D; 1&#x02212;<italic>e</italic><sup>&#x02212;<italic>f&#x00394;t</italic></sup>. Here, we consider independent patterns, i.e., with chance-level overlap. Hence the probability that a given afferent fires at least once in at least one of the <italic>P</italic> pattern subsection is 1&#x02212;(1&#x02212;<italic>p</italic>)<sup><italic>P</italic></sup>. Thus the number of selected afferents <italic>M</italic> is on average:</p>
<disp-formula id="E9"><label>(7)</label><mml:math id="M103"><mml:mo>&#x02329;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mo>=</mml:mo><mml:mi>N</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mi>P</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>N</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mi>P</mml:mi><mml:mi>f</mml:mi><mml:mi>&#x00394;</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula>
<p>Finally, the expected effective input spike rate during the &#x00394;<italic>t</italic> window is the expected total number of spikes, <italic>fN&#x00394;t</italic>, divided by &#x00394;<italic>t</italic>, thus:</p>
<disp-formula id="E4"><label>(8)</label><mml:math id="M10"><mml:mrow><mml:mo>&#x02329;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mi>N</mml:mi><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>We note that the <italic>SNR</italic> scales with <inline-formula><mml:math id="M11"><mml:msqrt><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msqrt></mml:math></inline-formula>. In the rest of this paper we used <italic>N</italic> &#x0003D; 10<sup>4</sup> afferents, which is in the biological range.</p>
</sec>
<sec>
<title>3.2. Numerical validations</title>
<p>We first checked if the variability of the <italic>SNR</italic> across Poisson realizations is small, and also if the approximation we made to compute the average <italic>SNR</italic> in Equation 6 is reasonable. To this aim, we generated 10<sup>5</sup> Poisson patterns, and computed <italic>M</italic>, <italic>r</italic> and the reduced <italic>SNR</italic>, <inline-formula><mml:math id="M12"><mml:mi>s</mml:mi><mml:mi>n</mml:mi><mml:mi>r</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>&#x0232A;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x02329;</mml:mo><mml:mo>-</mml:mo><mml:mi>f</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x02329;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>/</mml:mo><mml:msqrt><mml:mrow><mml:mo>&#x0232A;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x02329;</mml:mo></mml:mrow></mml:msqrt></mml:math></inline-formula>, for each of them (i.e., the right factor of the <italic>SNR</italic> in Equation 6, which is the only one that depends on the Poisson realization). As can be seen on Figure <xref ref-type="fig" rid="F2">2</xref>, left, <italic>M</italic> and <italic>r</italic> are strongly correlated, and the data points lie near a line which corresponds to nearly constant <italic>snr</italic> values (see the colored background). In other words, the <italic>snr</italic> does not change much for different Poisson pattern realizations and the average <italic>snr</italic> well represents the <italic>snr</italic> distribution even for the worst and best cases.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Numerical validation of the averaging operations. <bold>(Left)</bold> <italic>M</italic>&#x000D7;<italic>r</italic> plane. The white dots correspond to different realizations of a Poisson pattern (a jitter was added to better visualize density, given that both <italic>M</italic> and <italic>r</italic> are discrete). The background color shows the corresponding <italic>snr</italic>. The red cross corresponds to the average-case scenario <italic>M</italic> &#x0003D; &#x0232A;<italic>M</italic>&#x02329; and <italic>r</italic> &#x0003D; &#x0232A;<italic>r</italic>&#x02329;. <bold>(Right)</bold> The distribution of <italic>snr</italic> values across Poisson realizations. The vertical blue solid line shows its average. The vertical red dotted line shows our approximation, <inline-formula><mml:math id="M13"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>&#x0232A;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x02329;</mml:mo><mml:mo>-</mml:mo><mml:mi>f</mml:mi><mml:mo>&#x0232A;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x02329;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>/</mml:mo><mml:msqrt><mml:mrow><mml:mo>&#x02328;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x02329;</mml:mo></mml:mrow></mml:msqrt></mml:math></inline-formula>, which matches very well the true average. Parameters: <italic>P</italic> &#x0003D; 1, &#x00394;<italic>t</italic> &#x0003D; 2 ms, <italic>f</italic> &#x0003D; 1 Hz.</p></caption>
<graphic xlink:href="fncom-12-00074-g0002.tif"/>
</fig>
<p>In addition, as can be seen on Figure <xref ref-type="fig" rid="F2">2</xref>, right, the average <italic>snr</italic> across different Poisson patterns is very close to the <italic>snr</italic> corresponding to the average-case scenario, i.e., <italic>M</italic> &#x0003D; &#x0232A;<italic>M</italic>&#x02329; and <italic>r</italic> &#x0003D; &#x0232A;<italic>r</italic>&#x02329; (as defined by Equations 7 and 8, respectively). Note that this Figure was done with relatively small values for the parameters <italic>P</italic>, &#x00394;<italic>t</italic> and <italic>f</italic> (respectively 1, 2 ms, and 1 Hz). Our simulations indicate that when increasing these parameter values, the approximation becomes even better (data not shown).</p>
<p>Next, we verified the complete <italic>SNR</italic> formula (Equation 6), which also includes <italic>v</italic><sub>max</sub>, through numerical simulations. We used a clock-based approach, and integrated the LIF equation using the forward Euler method with a 0.1 ms time bin. We used <italic>P</italic> &#x0003D; 1 and <italic>P</italic> &#x0003D; 5 patterns, and performed 100 simulations with different random Poisson patterns of duration <italic>L</italic> &#x0003D; 20 ms with rate <italic>f</italic> &#x0003D; 5 Hz. We chose &#x00394;<italic>t</italic> &#x0003D; <italic>L</italic> &#x0003D; 20 ms, i.e., the LIF was connected to all the afferents that emitted at least once during one of the patterns. In order to estimate <italic>V</italic><sub>max</sub>, each pattern was presented 1,000 times, every 400 ms. The maximal jitter was <italic>T</italic> &#x0003D; 5 ms. Between pattern presentations, the afferents fired according to a Poisson process, still with rate <italic>f</italic> &#x0003D; 5 Hz, which allowed to estimate <inline-formula><mml:math id="M14"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> and <italic>&#x003C3;</italic><sub>noise</sub>. We could thus compute the <italic>SNR</italic> from Equation 1 (and its standard deviation across the 100 simulations), which, as can be seen on Figure <xref ref-type="fig" rid="F3">3</xref>, matches very well the theoretical values, for <italic>P</italic> &#x0003D; 1 and 5. Note that the <italic>SNR</italic> standard deviation is small, which confirms that the average <italic>SNR</italic>, i.e., &#x02328;<italic>SNR</italic>&#x02329;, represents well the individual ones.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Numerical validation of the theoretical <italic>SNR</italic> values, for <italic>P</italic> &#x0003D; 1 and 5 patterns. Error bars show &#x000B1;1 s.d.</p></caption>
<graphic xlink:href="fncom-12-00074-g0003.tif"/>
</fig>
</sec>
<sec>
<title>3.3. Optimizing the SNR</title>
<p>We now want to optimize the <italic>SNR</italic> given by Equation 6, by tuning &#x003C4; and &#x00394;<italic>t</italic>. We also add the constraint &#x003C4;<italic>fM</italic> &#x02265; 10 (large number of synaptic inputs), so that the distribution of V is approximately Gaussian (Burkitt, <xref ref-type="bibr" rid="B8">2006</xref>). Otherwise, it would be positively skewed<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref>, thus a high <italic>SNR</italic> would not guarantee a low false alarm rate. We assume that <italic>L</italic> is sufficiently large so that an upper bound for &#x00394;<italic>t</italic> is not needed. We used the Matlab R2017a Optimization Toolbox (MathWorks Inc., Natick, MA, USA) to compute the optimum numerically.</p>
<p>Figure <xref ref-type="fig" rid="F4">4</xref> illustrates the results with <italic>P</italic> &#x0003D; 2. One can make the following observations (similar to our previous paper which was limited to <italic>P</italic> &#x0003D; 1; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>):</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Optimal parameters for <italic>P</italic> &#x0003D; 2, as a function of <italic>f</italic> and <italic>T</italic>. <bold>(Left)</bold> Optimal &#x00394;<italic>t</italic>, divided by &#x003C4;. <bold>(Middle)</bold> Optimal &#x003C4; (note the logarithmic colormap). <bold>(Right)</bold> Resulting <italic>SNR</italic>.</p></caption>
<graphic xlink:href="fncom-12-00074-g0004.tif"/>
</fig>
<list list-type="bullet">
<list-item><p>Unless <italic>f</italic> and <italic>T</italic> are both high, the optimal &#x003C4; and &#x00394;<italic>t</italic> have the same order of magnitude (see Figure <xref ref-type="fig" rid="F4">4</xref>, left).</p></list-item>
<list-item><p>Unless <italic>T</italic> is high (&#x0003E;10 ms), or <italic>f</italic> is low (&#x0003C;1 Hz), then these timescales should be relatively small (at most a few tens of ms; see Figure <xref ref-type="fig" rid="F4">4</xref>, middle). This means that even long patterns (hundreds of ms or more) are optimally detected by a coincidence detector working at a shorter timescale, and which thus ignores most of the patterns. One could have thought that using &#x003C4; &#x0007E; <italic>L</italic>, to integrate all the spikes from the pattern would be the best strategy. But a long &#x003C4; also decreases the detector&#x00027;s temporal resolution, thus patterns and random inputs elicit more similar responses, decreasing the <italic>SNR</italic>. Hence there is a trade-off, and it turns out that it is often more optimal to have &#x003C4; &#x0003C; <italic>L</italic>, that is to use subpatterns as signatures for the whole patterns.</p></list-item>
<list-item><p>Unsurprisingly, the optimal <italic>SNR</italic> decreases with <italic>T</italic> (see Figure <xref ref-type="fig" rid="F4">4</xref>, right). What is less trivial, is that it also decreases with <italic>f</italic>. In other words, sparse activity is preferable. We will come back to this point in the discussion.</p></list-item>
</list>
<p>What is the biological range for <italic>T</italic>, which corresponds to the spike time precision? Millisecond precision in cortex has been reported (Kayser et al., <xref ref-type="bibr" rid="B22">2010</xref>; Panzeri and Diamond, <xref ref-type="bibr" rid="B44">2010</xref>; Havenith et al., <xref ref-type="bibr" rid="B15">2011</xref>). We are aware that other studies found poorer precision, but this could be due to uncontrolled variable or the use of inappropriate reference times (Masquelier, <xref ref-type="bibr" rid="B31">2013</xref>).</p>
<p>In the rest of the paper we focus, as an example, on the point on the middle of the <italic>T</italic> &#x000D7; <italic>f</italic> plane &#x02013;<italic>T</italic> &#x0003D; 3.2 ms and <italic>f</italic> &#x0003D; 3.2 Hz. When increasing <italic>P</italic>, the optimal &#x003C4; and &#x00394;<italic>t</italic> decrease (Figure <xref ref-type="fig" rid="F5">5</xref>). Unsurprisingly, the resulting <italic>SNR</italic> also decreases, but only slowly. It thus remains acceptable for several tens of independent patterns (e.g., <italic>SNR</italic> &#x0007E; 7 for <italic>P</italic> &#x0003D; 40).</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Optimal &#x003C4; and &#x00394;<italic>t</italic> (for <italic>f</italic> &#x0003D; 3.2 Hz, <italic>T</italic> &#x0003D; 3.2 ms) and resulting <italic>SNR</italic> as a function of <italic>P</italic>.</p></caption>
<graphic xlink:href="fncom-12-00074-g0005.tif"/>
</fig>
</sec>
</sec>
<sec id="s4">
<title>4. Simulations show that STDP can be close-to-optimal</title>
<p>Next we investigated, through numerical simulations, if STDP could turn a LIF neuron into an optimal multi-pattern detector. More specifically, since STDP does not adjust the membrane time constant &#x003C4;, we set it to the optimal value and investigated whether STDP could learn all the patterns with an optimal &#x00394;<italic>t</italic><xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>. Here, unlike in the previous section, we had to introduce a threshold, in order to have postsynaptic spikes, which are required for STDP. As a result, the optimal <italic>V</italic><sub>max</sub>, computed in the previous section, was never reached. Yet a high <italic>V</italic><sub>max</sub> guarantees a low miss rate, and a low <inline-formula><mml:math id="M15"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> guarantees a low false alarm rate. Optimizing the previously defined <italic>SNR</italic> thus makes sense.</p>
<p>Again, we used a clock-based approach, and the forward Euler method with a 0.1 ms time bin. The Matlab R2017a code for these simulations has been made available in ModelDB (Hines et al., <xref ref-type="bibr" rid="B17">2004</xref>) at <ext-link ext-link-type="uri" xlink:href="https://senselab.med.yale.edu/modeldb/">https://senselab.med.yale.edu/modeldb/</ext-link>.</p>
<sec>
<title>4.1. Input spikes</title>
<p>The setup we used was similar to the one of our previous studies (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>). Between pattern presentations, the input spikes were generated randomly with a homogeneous Poisson process with rate <italic>f</italic>. The <italic>P</italic> spike patterns with duration <italic>L</italic> &#x0003D; 100 ms were generated only once using the same Poisson process (frozen noise). The pattern presentations occurred every 400 ms [in previous studies, we demonstrated that irregular intervals did not matter (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>), so here regular intervals were used for simplicity]. The <italic>P</italic> patterns were presented alternatively, over and over again. Figure <xref ref-type="fig" rid="F6">6</xref> shows an example with <italic>P</italic> &#x0003D; 2 patterns. At each pattern presentation, all the spike times were shifted independently by some random jitters uniformly distributed over [&#x02212;<italic>T, T</italic>].</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Unsupervised STDP-based pattern learning. The neuron becomes selective to <italic>P</italic> &#x0003D; 2 patterns. <bold>(Top)</bold> Initial state. On the left, we plotted the neuron&#x00027;s potential as a function of time. Colored rectangles indicate pattern presentations. Next, we plotted the two spike patterns, coloring the spikes as a function of the corresponding synaptic weights: blue for low weight (0), purple for intermediate weight, and red for high weight (1). Initial weights were uniform (here at 0.7, so the initial color is close to red). <bold>(Middle)</bold> During learning. Selectivity progressively emerges. <bold>(Bottom)</bold> After convergence. STDP has concentrated the weights on the afferents which fire at least once in at least one of the pattern subsections, located at the beginning of each pattern, and whose duration roughly matches the optimal &#x00394;<italic>t</italic> (shown in green). This results in one postsynaptic spike each time either one of the two pattern is presented. Elsewhere both <inline-formula><mml:math id="M16"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> and <italic>&#x003C3;</italic><sub>noise</sub> are low, so the <italic>SNR</italic> is high. In addition <inline-formula><mml:math id="M17"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> roughly matches the theoretical value <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow><mml:mrow><mml:mtext>opt</mml:mtext></mml:mrow></mml:msubsup></mml:math></inline-formula> (shown in green), corresponding to the optimal <italic>SNR</italic>. We also show in green <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mtext>max</mml:mtext></mml:mrow><mml:mrow><mml:mtext>opt</mml:mtext></mml:mrow></mml:msubsup></mml:math></inline-formula>, the theoretical optimal value for <italic>V</italic><sub>max</sub>. However, the potential never reaches it, because the adaptive threshold is reached before.</p></caption>
<graphic xlink:href="fncom-12-00074-g0006.tif"/>
</fig>
</sec>
<sec>
<title>4.2. A LIF neuron with adaptive threshold</title>
<p>We simulated a LIF neuron connected to all of the <italic>N</italic> afferents with plastic synaptic weights <italic>w</italic><sub><italic>i</italic></sub> &#x02208; [0, 1], thus obeying the following differential equation:</p>
<disp-formula id="E5"><label>(9)</label><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>&#x003C4;</mml:mi><mml:mfrac><mml:mrow><mml:mtext>d</mml:mtext><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mtext>d</mml:mtext><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>V</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>&#x003B4;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>t</italic><sub><italic>ij</italic></sub> is the time of the <italic>j</italic><sup><italic>th</italic></sup> spike of afferent <italic>i</italic>.</p>
<p>We used an adaptive threshold [unlike in our previous studies (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>), in which a fixed threshold was used]. This adaptive threshold was increased by a fixed amount (1.8&#x003B8;<sub>0</sub>) at each postsynaptic spike, and then exponentially decayed toward its baseline value &#x003B8;<sub>0</sub> with a time constant &#x003C4;<sub>&#x003B8;</sub> &#x0003D; 80 ms. This is a simple, yet good model of cortical cells, in the sense that it predicts very well the spikes elicited by a given input current (Gerstner and Naud, <xref ref-type="bibr" rid="B11">2009</xref>; Kobayashi et al., <xref ref-type="bibr" rid="B27">2009</xref>). Here, such an adaptive threshold is crucial to encourage the neuron to learn multiple patterns, as opposed to fire multiple successive spikes to the same pattern. Since the theory developed in the previous sections ignored the LIF threshold, using an adaptive one is not worse than a fixed one, in the sense that it does not make the theory less valid.</p>
<p>We did not know which value for &#x003B8;<sub>0</sub> could lead to the optimum. We thus performed and exhaustive search, using a geometric progression with a ratio of 2.5%.</p>
</sec>
<sec>
<title>4.3. Synaptic plasticity</title>
<p>Initial synaptic weights were all equal. Their value was computed so that <inline-formula><mml:math id="M21"><mml:msub><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mtext>noise</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> (leading to an initial firing rate of about 4 Hz, see Figure <xref ref-type="fig" rid="F6">6</xref>, top). They then evolved in [0, 1] with all-to-all spike STDP. Yet, we only modeled the Long Term Potentiation part of STDP, ignoring its Long Term Depression (LTD) term. As in Song et al. (<xref ref-type="bibr" rid="B52">2000</xref>), we used a trace of presynaptic spikes at each synapse <italic>i</italic>, <inline-formula><mml:math id="M22"><mml:msubsup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mtext>pre</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, which was incremented by &#x003B4;<italic>A</italic><sub>pre</sub> at each presynaptic spike, and then exponentially decayed toward 0 with a time constant &#x003C4;<sub>pre</sub> &#x0003D; 20ms. At each postsynaptic spike this trace was used for LTP at each synapse: <inline-formula><mml:math id="M23"><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02192;</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mtext>pre</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>.</p>
<p>Here LTD was modeled by a simple homeostatic mechanism. At each postsynaptic spike, all synapses were depressed: <inline-formula><mml:math id="M24"><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02192;</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mtext>out</mml:mtext></mml:mrow></mml:msup></mml:math></inline-formula> where <italic>w</italic><sup>out</sup> &#x0003C; 0 is a fixed parameter (Kempter et al., <xref ref-type="bibr" rid="B23">1999</xref>).</p>
<p>Note that for both LTP and LTD we used the multiplicative term <italic>w</italic><sub><italic>i</italic></sub>(1&#x02212;<italic>w</italic><sub><italic>i</italic></sub>), in contrast with additive STDP, with which the &#x00394;<italic>w</italic> is independent of the current weight value (Kempter et al., <xref ref-type="bibr" rid="B23">1999</xref>; Song et al., <xref ref-type="bibr" rid="B52">2000</xref>). This multiplicative term ensures that the weights remain in the range [0,1], and the weight dependence creates a soft bound effect: when a weight approaches a bound, weight changes tend toward zero. Here it was found to increase performance (convergence time and stability), in line with our previous studies (Masquelier and Thorpe, <xref ref-type="bibr" rid="B35">2007</xref>; Kheradpisheh et al., <xref ref-type="bibr" rid="B24">2016</xref>, <xref ref-type="bibr" rid="B25">2018</xref>; Mozafari et al., <xref ref-type="bibr" rid="B38">2018b</xref>).</p>
<p>The ratio between LTP and LTD, that is between &#x003B4;<italic>A</italic><sub>pre</sub> and <italic>w</italic><sup>out</sup> is crucial: the higher, the more synapses are maximally potentiated (<italic>w</italic> &#x0003D; 1) after convergence. Here we chose to keep &#x003B4;<italic>A</italic><sub>pre</sub> &#x0003D; 0.1 and to systematically vary <italic>w</italic><sup>out</sup>, using again a geometric progression with a ratio of 2.5%.</p>
</sec>
<sec>
<title>4.4. Results</title>
<p>For each <inline-formula><mml:math id="M25"><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mtext>out</mml:mtext></mml:mrow></mml:msup></mml:math></inline-formula> point, 100 simulations were performed with different random pattern realizations, and we computed the proportion of &#x0201C;optimal&#x0201D; ones (see below), and reported it in Table <xref ref-type="table" rid="T1">1</xref>. After 12,000 s of simulated time, the synaptic weights had all converged by saturation. That is synapses were either completely depressed (<italic>w</italic> &#x0003D; 0), or maximally potentiated (<italic>w</italic> &#x0003D; 1). A simulation was considered optimal if</p>
<list list-type="order">
<list-item><p><italic>all</italic> the patterns were learned, and</p></list-item>
<list-item><p>in an optimal way, that is if all patterns exhibited a subsection in which all spikes corresponded to maximally potentiated synapses (<italic>w</italic> &#x0003D; 1), and whose duration roughly matched the theoretical optimal &#x00394;<italic>t</italic>. In practice, we used the total number of potentiated synapses as a proxy of the mean subsection duration (since there is a non-ambiguous mapping between the two variables, given by Equation 7), and checked if this number matched the theoretical optimal <italic>M</italic> (Equation 7) with a 5% margin.</p></list-item>
</list>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Performance as a function of the number of patterns <italic>P</italic>.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold><italic>P</italic></bold></th>
<th valign="top" align="center"><bold>5</bold></th>
<th valign="top" align="center"><bold>10</bold></th>
<th valign="top" align="center"><bold>20</bold></th>
<th valign="top" align="center"><bold>40</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">&#x00394;<italic>t</italic><sup>opt</sup> (ms)</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">8.1</td>
<td valign="top" align="center">5.7</td>
<td valign="top" align="center">3.7</td>
</tr>
<tr>
<td valign="top" align="left">&#x003C4;<sup>opt</sup> (ms)</td>
<td valign="top" align="center">8.9</td>
<td valign="top" align="center">6.8</td>
<td valign="top" align="center">5.6</td>
<td valign="top" align="center">5.1</td>
</tr>
<tr>
<td valign="top" align="left"><italic>M</italic><sup>opt</sup></td>
<td valign="top" align="center">1,600</td>
<td valign="top" align="center">2,300</td>
<td valign="top" align="center">3,100</td>
<td valign="top" align="center">3,800</td>
</tr>
<tr>
<td valign="top" align="left"><italic>SNR</italic><sup>opt</sup></td>
<td valign="top" align="center">31</td>
<td valign="top" align="center">20</td>
<td valign="top" align="center">12</td>
<td valign="top" align="center">6.7</td>
</tr>
<tr style="border-top: thin solid #000000;">
<td valign="top" align="left"><italic>&#x003B8;</italic><sub>0</sub></td>
<td valign="top" align="center">190</td>
<td valign="top" align="center">140</td>
<td valign="top" align="center">110</td>
<td valign="top" align="center">92</td>
</tr>
<tr style="border-bottom: thin solid #000000;">
<td valign="top" align="left"><italic>w</italic><sup>out</sup></td>
<td valign="top" align="center">&#x02212;6.210<sup>&#x02212;3</sup></td>
<td valign="top" align="center">&#x02212;6.310<sup>&#x02212;3</sup></td>
<td valign="top" align="center">&#x02212;6.510<sup>&#x02212;3</sup></td>
<td valign="top" align="center">&#x02212;6.710<sup>&#x02212;3</sup></td>
</tr> <tr>
<td valign="top" align="left">&#x0003C;<italic>P</italic><sub>learned</sub>&#x0003E;</td>
<td valign="top" align="center">5</td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">20</td>
<td valign="top" align="center">39.5</td>
</tr>
<tr>
<td valign="top" align="left">Hit rate (%)</td>
<td valign="top" align="center">98.9</td>
<td valign="top" align="center">98.6</td>
<td valign="top" align="center">97.9</td>
<td valign="top" align="center">96.5</td>
</tr>
<tr>
<td valign="top" align="left">False alarms (Hz)</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="left">P(opt) (%)</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">58</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>The first four lines are computed from the theoretical optimum. The next two lines are the optimal values found through exhaustive search (see text). The last four lines are performance indicators, estimated during the last 100 presentations of each pattern. &#x0232A;<italic>P</italic><sub><italic>learned</italic></sub>&#x02329; is the mean number of &#x0201C;learned patterns,&#x0201D; that is by convention patterns which elicit at least one postsynaptic spike. The following line is the mean hit rate for those patterns. The subsequent line gives the false alarm rate, but we never observed any here. Finally P(opt) is the proportion of optimal cases</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Note that this second condition alone would be easy to satisfy: the total amount of potentiated synapses is determined by the LTP/LTD ratio which we adjusted by fine-tuning <italic>w</italic><sup>out</sup>. However, satisfying the two conditions is harder, especially when <italic>P</italic> increases (Table <xref ref-type="table" rid="T1">1</xref>).</p>
<p>It is worth mentioning that the learned subsections always corresponded to the beginning of the patterns, because STDP tracks back through them (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>), but this is irrelevant here since all the subsections are equivalent for the theory. Figure <xref ref-type="fig" rid="F6">6</xref> shows an optimal simulation with <italic>P</italic> &#x0003D; 2 patterns.</p>
<p>As can be seen in Table <xref ref-type="table" rid="T1">1</xref>, the proportion of optimal simulations decreases with <italic>P</italic>, as expected. But more surprisingly, several tens of patterns can be optimally learned with reasonably high probability. With <italic>P</italic> &#x0003D; 40 the probability of optimal simulations is only 58%, but the average number of learned patterns is high: 39.5! This means that nearly all patterns are learned in all simulations, yet sometimes in a suboptimal manner. Finally, Figure <xref ref-type="fig" rid="F7">7</xref> shows that convergence time increases with <italic>P</italic>.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Convergence index as a function of time and number of patterns, for an example of optimal simulation. The convergence index is defined as the mean distance between the full precision weights, and their binary quantization (0 if <italic>w</italic> &#x0003C; 0.5, and 1 otherwise).</p></caption>
<graphic xlink:href="fncom-12-00074-g0007.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<title>5. Discussion</title>
<p>The fact that STDP can generate selectivity to any repeating spike pattern in an unsupervised manner is a remarkable, yet well documented fact (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>, <xref ref-type="bibr" rid="B34">2009</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>; Humble et al., <xref ref-type="bibr" rid="B18">2012</xref>; Hunzinger et al., <xref ref-type="bibr" rid="B20">2012</xref>; Kasabov et al., <xref ref-type="bibr" rid="B21">2013</xref>; Klampfl and Maass, <xref ref-type="bibr" rid="B26">2013</xref>; Nessler et al., <xref ref-type="bibr" rid="B39">2013</xref>; Krunglevicius, <xref ref-type="bibr" rid="B29">2015</xref>; Yger et al., <xref ref-type="bibr" rid="B60">2015</xref>; Sun et al., <xref ref-type="bibr" rid="B53">2016</xref>; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>). Here we have shown that, surprisingly, a single neuron can become optimally selective to several tens of independent patterns. Hence STDP and coincidence detection are compatible with distributed coding.</p>
<p>Yet one issue with having one neuron selective to multiple patterns is stability. If one of the learned pattern does not occur for a long period during which the other patterns occur many times, causing postsynaptic spikes, the unseen pattern will tend to be forgotten. This is not an issue with localist coding: if the learned pattern does not occur, the threshold is hardly ever reached so the weights are not modified, and the pattern is retained indefinitely, even if STDP is &#x0201C;on&#x0201D; all the time.</p>
<p>Another issue with distributed coding is how the readout is done, that is how the identity of the stimulus can be inferred from multiple neuron responses, given that each response is ambiguous? This is out of the scope of the current paper, but we suspect that STDP could again help. As shown in this study, each neuron equipped with STDP can learn to fire to multiple independent stimuli. Let&#x00027;s suppose that stimuli are shown one at a time. When stimulus <italic>A</italic> is shown, all the neurons that learned this stimulus (among others) will fire synchronously. Let us call <italic>S</italic> this set of neurons. A downstream neuron equipped with STDP could easily become selective to this synchronous volley of spikes from neurons in <italic>S</italic> (Brette, <xref ref-type="bibr" rid="B5">2012</xref>). With an appropriate threshold, this neuron would fire if and only if all the neurons in <italic>S</italic> have fired. Does that necessarily mean that A is there? Yes, if the intersection of the sets of stimuli learned by neurons in <italic>S</italic> only contains <italic>A</italic>. In the general case, the intersection is likely to be much smaller than the typical sets of stimuli learned by the <italic>S</italic> neurons, so much of the ambiguity should be resolved.</p>
<p>What could determine the set of patterns to which a neuron responds? Here, we used independent, unrelated, patterns (i.e., with chance-level overlap), and yet several of these patterns could be learned by a single neuron. Of course, patterns with more overlap would be easier to group. So in the presence of multiple postsynaptic neurons, each one would tend to learn a cluster of similar patterns. Another factor is the time at which the patterns are presented: those presented at the same period are more likely to be learned by the same neuron&#x02014;a neuron which was still unselective at that period. Indeed, neurons equipped with STDP have some sort of critical period, before convergence, during which they can learn new pattern easily. Conversely, after convergence, neurons tend to fire if and only if the patterns they have learned are presented (Figure <xref ref-type="fig" rid="F6">6</xref>), and thus can hardly learn any new pattern. This is interesting, because patterns presented at the same period are likely to be somewhat related. For example, a neuron could fire to the different people you have met on your first day at work. In the presence of neurogenesis, newborn neurons could handle the learning of other patterns during the subsequent periods of your life. Finally, here we did not use any reward signal. But such a signal, if available, could modulate STDP (leading to some form of supervised learning), and encourage a given neuron to fire to a particular, meaningful, set of patterns (Mozafari et al., <xref ref-type="bibr" rid="B37">2018a</xref>,<xref ref-type="bibr" rid="B38">b</xref>), as opposed to a random set like here. For example, a single neuron could learn to fire to any animal, even if different animals cause very different sensory inputs.</p>
<p>Here the STDP rule we used always led to binary weights after learning. That is an afferent could be either selected or discarded. We thus could use our <italic>SNR</italic> calculations derived with binary weights, and checked that the selected set was optimal given the binary weight constraint. Further calculations in the <xref ref-type="supplementary-material" rid="SM1">Appendix</xref> (Supplementary Material) suggest that removing such a constraint could lead to a modest increase in <italic>SNR</italic>, of about 10%. More research is needed to see if a multiplicative STDP rule, which does not converge toward binary weights (van Rossum et al., <xref ref-type="bibr" rid="B58">2000</xref>; G&#x000FC;tig et al., <xref ref-type="bibr" rid="B13">2003</xref>), could lead to the optimal graded weights.</p>
<p>Our theoretical study suggests, together with others (G&#x000FC;tig and Sompolinsky, <xref ref-type="bibr" rid="B14">2006</xref>; Brette, <xref ref-type="bibr" rid="B5">2012</xref>), that coincidence detection is computationally powerful. In fact, it could be the main function of neurons (Abeles, <xref ref-type="bibr" rid="B1">1982</xref>; K&#x000F6;nig et al., <xref ref-type="bibr" rid="B28">1996</xref>). In line with this proposal, neurons <italic>in vivo</italic> appear to be mainly fluctuation-driven, not mean-driven (Rossant et al., <xref ref-type="bibr" rid="B49">2011</xref>; Brette, <xref ref-type="bibr" rid="B5">2012</xref>, <xref ref-type="bibr" rid="B6">2015</xref>). This is the case in particular in the balanced regime (Brette, <xref ref-type="bibr" rid="B6">2015</xref>), which appears to be the prevalent regime in the brain (Den&#x000E8;ve and Machens, <xref ref-type="bibr" rid="B9">2016</xref>). Several other points suggest that coincidence detection is the main function of neurons. Firstly, strong feedforward inhibitory circuits throughout the central nervous system often shorten the neurons&#x00027; effective integration windows (Bruno, <xref ref-type="bibr" rid="B7">2011</xref>). Secondly, the effective integration time constant in dendrites might be one order of magnitude shorter than the soma&#x00027;s one (K&#x000F6;nig et al., <xref ref-type="bibr" rid="B28">1996</xref>). Finally, recent experiments indicate that a neuron&#x00027;s threshold quickly adapts to recent potential values (Platkiewicz and Brette, <xref ref-type="bibr" rid="B45">2011</xref>; Fontaine et al., <xref ref-type="bibr" rid="B10">2014</xref>; Mensi et al., <xref ref-type="bibr" rid="B36">2016</xref>), so that only a sudden potential increase can trigger a postsynaptic spike. This enhances coincidence detection. It remains unclear if other spike time aspects such as ranks (Thorpe and Gautrais, <xref ref-type="bibr" rid="B57">1998</xref>) also matter.</p>
<p>Our results show that lower firing rates lead to better signal-to-ratio. It is worth mentioning that mean firing rates are probably largely overestimated in the electrophysiological literature, because extracellular recordings&#x02014;by far the most popular technique&#x02014; are totally blind to cells that do not fire at all (Thorpe, <xref ref-type="bibr" rid="B56">2011</xref>). Even a cell that fire only a handful of spikes will be ignored, because spike sorting algorithms need tens of spikes from a given cell before they can create a new cluster corresponding to that cell. Furthermore, experimentalists tend to search for stimuli that elicit strong responses, and, when they can move the electrode(s), tend to look for most responsive cells, introducing strong selection biases. Mean firing rates, averaged across time and cells, are largely unknown, but they could be smaller than 1 Hz (Shoham et al., <xref ref-type="bibr" rid="B51">2006</xref>). It seems like coding is sparse: neurons only fire when they need to signal an important event, and that every spike matters (Wolfe et al., <xref ref-type="bibr" rid="B59">2010</xref>).</p>
<p>Finally, we see an analogy between our theory, and the one of neural associative memory (NAM), in which an output (data) vector is produced by multiplying an input (address) vector by a weight matrix. Unlike NAM, our framework is dynamic, yet after learning, to a first approximation, our STDP neurons count the number of input spikes arriving through reinforced synapses in a short integration window, and each one outputs a 1 (i.e., a spike) if this count exceeds a threshold, and a 0 otherwise, leading to a binary output vector, much like in a binary NAM. It is thus unsurprising that sparsity is desirable both in our theory, and in NAMs (Palm, <xref ref-type="bibr" rid="B43">2013</xref>).</p>
</sec>
<sec id="s6">
<title>Author contributions</title>
<p>SK and TM did the analytical derivations. TM performed the numerical simulations. SK and TM analyzed the results. TM wrote the first version of the manuscript. SK reviewed it.</p>
<sec>
<title>Conflict of interest statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack><p>This research received funding from the European Research Council under the European Union&#x00027;s 7th Framework Program (FP/2007-2013) / ERC Grant Agreement n.323711 (M4 project). We thank Milad Mozafari for smart implementation hints, and Jean Pierre Jaffr&#x000E9;zou for his excellent copy editing.</p>
</ack>
<sec sec-type="supplementary-material" id="s7">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fncom.2018.00074/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fncom.2018.00074/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Abeles</surname> <given-names>M.</given-names></name></person-group> (<year>1982</year>). <article-title>Role of the cortical neuron: integrator or coincidence detector?</article-title> <source>Isr. J. Med. Sci.</source> <volume>18</volume>, <fpage>83</fpage>&#x02013;<lpage>92</lpage>.</citation>
</ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Abeles</surname> <given-names>M.</given-names></name></person-group> (<year>1991</year>). <source>Corticonics : Neural Circuits of the Cerebral Cortex</source>. <publisher-loc>Cambridge; New York, NY</publisher-loc>: <publisher-name>Cambridge University Press</publisher-name>.</citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bowers</surname> <given-names>J. S.</given-names></name></person-group> (<year>2009</year>). <article-title>On the biological plausibility of grandmother cells: implications for neural network theories in psychology and neuroscience</article-title>. <source>Psychol. Rev.</source> <volume>116</volume>, <fpage>220</fpage>&#x02013;<lpage>251</lpage>. <pub-id pub-id-type="doi">10.1037/a0014462</pub-id><pub-id pub-id-type="pmid">19159155</pub-id></citation>
</ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bowers</surname> <given-names>J. S.</given-names></name></person-group> (<year>2017</year>). <article-title>Parallel distributed processing theory in the age of deep networks</article-title>. <source>Trends Cogn. Sci.</source> <volume>21</volume>, <fpage>950</fpage>&#x02013;<lpage>961</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2017.09.013</pub-id><pub-id pub-id-type="pmid">29100738</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>Computing with neural synchrony</article-title>. <source>PLOS Comput. Biol.</source> <volume>8</volume>:<fpage>e1002561</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002561</pub-id><pub-id pub-id-type="pmid">22719243</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). <article-title>Philosophy of the spike: rate-based vs. spike-based theories of the brain</article-title>. <source>Front. Syst. Neurosci.</source> <volume>9</volume>:<fpage>151</fpage>. <pub-id pub-id-type="doi">10.3389/fnsys.2015.00151</pub-id><pub-id pub-id-type="pmid">26617496</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bruno</surname> <given-names>R. M.</given-names></name></person-group> (<year>2011</year>). <article-title>Synchrony in sensation</article-title>. <source>Curr. Opin. Neurobiol.</source> <volume>21</volume>, <fpage>701</fpage>&#x02013;<lpage>708</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2011.06.003</pub-id><pub-id pub-id-type="pmid">21723114</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Burkitt</surname> <given-names>A. N.</given-names></name></person-group> (<year>2006</year>). <article-title>A review of the integrate-and-fire neuron model: I. Homogeneous synaptic input</article-title>. <source>Biol. Cybern.</source> <volume>95</volume>, <fpage>1</fpage>&#x02013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1007/s00422-006-0068-6</pub-id><pub-id pub-id-type="pmid">16622699</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Den&#x000E8;ve</surname> <given-names>S.</given-names></name> <name><surname>Machens</surname> <given-names>C. K.</given-names></name></person-group> (<year>2016</year>). <article-title>Efficient codes and balanced networks</article-title>. <source>Nat. Neurosci.</source> <volume>19</volume>, <fpage>375</fpage>&#x02013;<lpage>382</lpage>. <pub-id pub-id-type="doi">10.1038/nn.4243</pub-id><pub-id pub-id-type="pmid">26906504</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fontaine</surname> <given-names>B.</given-names></name> <name><surname>Pe&#x000F1;a</surname> <given-names>J. L.</given-names></name> <name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2014</year>). <article-title>Spike-threshold adaptation predicted by membrane potential dynamics <italic>in vivo</italic></article-title>. <source>PLoS Comput. Biol.</source> <volume>10</volume>:<fpage>e1003560</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1003560</pub-id><pub-id pub-id-type="pmid">24722397</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gerstner</surname> <given-names>W.</given-names></name> <name><surname>Naud</surname> <given-names>R.</given-names></name></person-group> (<year>2009</year>). <article-title>How good are neuron models?</article-title> <source>Science</source> <volume>326</volume>, <fpage>379</fpage>&#x02013;<lpage>380</lpage>. <pub-id pub-id-type="doi">10.1126/science.1181936</pub-id><pub-id pub-id-type="pmid">19833951</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gilson</surname> <given-names>M.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name> <name><surname>Hugues</surname> <given-names>E.</given-names></name></person-group> (<year>2011</year>). <article-title>STDP allows fast rate-modulated coding with Poisson-like spike trains</article-title>. <source>PLoS Comput. Biol.</source> <volume>7</volume>:<fpage>e1002231</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002231</pub-id><pub-id pub-id-type="pmid">22046113</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>G&#x000FC;tig</surname> <given-names>R.</given-names></name> <name><surname>Aharonov</surname> <given-names>R.</given-names></name> <name><surname>Rotter</surname> <given-names>S.</given-names></name> <name><surname>Sompolinsky</surname> <given-names>H.</given-names></name></person-group> (<year>2003</year>). <article-title>Learning input correlations through nonlinear temporally asymmetric Hebbian plasticity</article-title>. <source>J. Neurosci.</source> <volume>23</volume>, <fpage>3697</fpage>&#x02013;<lpage>3714</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.23-09-03697.2003</pub-id><pub-id pub-id-type="pmid">12736341</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>G&#x000FC;tig</surname> <given-names>R.</given-names></name> <name><surname>Sompolinsky</surname> <given-names>H.</given-names></name></person-group> (<year>2006</year>). <article-title>The tempotron: a neuron that learns spike timing-based decisions</article-title>. <source>Nat. Neurosci.</source> <volume>9</volume>, <fpage>420</fpage>&#x02013;<lpage>428</lpage>. <pub-id pub-id-type="doi">10.1038/nn1643</pub-id><pub-id pub-id-type="pmid">16474393</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Havenith</surname> <given-names>M. N.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name> <name><surname>Biederlack</surname> <given-names>J.</given-names></name> <name><surname>Chen</surname> <given-names>N.-H.</given-names></name> <name><surname>Singer</surname> <given-names>W.</given-names></name> <name><surname>Nikolic</surname> <given-names>D.</given-names></name></person-group> (<year>2011</year>). <article-title>Synchrony makes neurons fire in sequence, and stimulus properties determine who is ahead</article-title>. <source>J. Neurosci.</source> <volume>31</volume>, <fpage>8570</fpage>&#x02013;<lpage>8584</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2817-10.2011</pub-id><pub-id pub-id-type="pmid">21653861</pub-id></citation>
</ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Herrmann</surname> <given-names>M.</given-names></name> <name><surname>Hertz</surname> <given-names>J. A.</given-names></name> <name><surname>Pr&#x000FC;gel-Bennett</surname> <given-names>A.</given-names></name></person-group> (<year>1995</year>). <article-title>Analysis of synfire chains</article-title>. <source>Netw. Comput. Neural Syst.</source> <volume>6</volume>, <fpage>403</fpage>&#x02013;<lpage>414</lpage>.</citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hines</surname> <given-names>M. L.</given-names></name> <name><surname>Morse</surname> <given-names>T.</given-names></name> <name><surname>Migliore</surname> <given-names>M.</given-names></name> <name><surname>Carnevale</surname> <given-names>N. T.</given-names></name> <name><surname>Shepherd</surname> <given-names>G. M.</given-names></name></person-group> (<year>2004</year>). <article-title>ModelDB: a database to support computational neuroscience</article-title>. <source>J. Comput. Neurosci.</source> <volume>17</volume>, <fpage>7</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1023/B:JCNS.0000023869.22017.2e</pub-id><pub-id pub-id-type="pmid">15218350</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Humble</surname> <given-names>J.</given-names></name> <name><surname>Denham</surname> <given-names>S.</given-names></name> <name><surname>Wennekers</surname> <given-names>T.</given-names></name></person-group> (<year>2012</year>). <article-title>Spatio-temporal pattern recognizers using spiking neurons and spike-timing-dependent plasticity</article-title>. <source>Front. Comput. Neurosci.</source> <volume>6</volume>:<fpage>84</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2012.00084</pub-id><pub-id pub-id-type="pmid">23087641</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hung</surname> <given-names>C. P.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name> <name><surname>Poggio</surname> <given-names>T.</given-names></name> <name><surname>DiCarlo</surname> <given-names>J. J.</given-names></name></person-group> (<year>2005</year>). <article-title>Fast readout of object identity from macaque inferior temporal cortex</article-title>. <source>Science</source> <volume>310</volume>, <fpage>863</fpage>&#x02013;<lpage>866</lpage>. <pub-id pub-id-type="doi">10.1126/science.1117593</pub-id><pub-id pub-id-type="pmid">16272124</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hunzinger</surname> <given-names>J. F.</given-names></name> <name><surname>Chan</surname> <given-names>V. H.</given-names></name> <name><surname>Froemke</surname> <given-names>R. C.</given-names></name></person-group> (<year>2012</year>). <article-title>Learning complex temporal patterns with resource-dependent spike timing-dependent plasticity</article-title>. <source>J. Neurophysiol.</source> <volume>108</volume>, <fpage>551</fpage>&#x02013;<lpage>566</lpage>. <pub-id pub-id-type="doi">10.1152/jn.01150.2011</pub-id><pub-id pub-id-type="pmid">22496526</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kasabov</surname> <given-names>N.</given-names></name> <name><surname>Dhoble</surname> <given-names>K.</given-names></name> <name><surname>Nuntalid</surname> <given-names>N.</given-names></name> <name><surname>Indiveri</surname> <given-names>G.</given-names></name></person-group> (<year>2013</year>). <article-title>Dynamic evolving spiking neural networks for on-line spatio- and spectro-temporal pattern recognition</article-title>. <source>Neural Netw.</source> <volume>41</volume>, <fpage>188</fpage>&#x02013;<lpage>201</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2012.11.014</pub-id><pub-id pub-id-type="pmid">23340243</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kayser</surname> <given-names>C.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name></person-group> (<year>2010</year>). <article-title>Millisecond encoding precision of auditory cortex neurons</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A.</source> <volume>107</volume>, <fpage>16976</fpage>&#x02013;<lpage>16981</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1012656107</pub-id><pub-id pub-id-type="pmid">20837521</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kempter</surname> <given-names>R.</given-names></name> <name><surname>Gerstner</surname> <given-names>W.</given-names></name> <name><surname>van Hemmen</surname> <given-names>J. L.</given-names></name></person-group> (<year>1999</year>). <article-title>Hebbian learning and spiking neurons</article-title>. <source>Phys. Rev. E</source> <volume>59</volume>, <fpage>4498</fpage>&#x02013;<lpage>4514</lpage>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kheradpisheh</surname> <given-names>S. R.</given-names></name> <name><surname>Ganjtabesh</surname> <given-names>M.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2016</year>). <article-title>Bio-inspired unsupervised learning of visual features leads to robust invariant object recognition</article-title>. <source>Neurocomputing</source> <volume>205</volume>, <fpage>382</fpage>&#x02013;<lpage>392</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2016.04.029</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kheradpisheh</surname> <given-names>S. R.</given-names></name> <name><surname>Ganjtabesh</surname> <given-names>M.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2018</year>). <article-title>STDP-based spiking deep convolutional neural networks for object recognition</article-title>. <source>Neural Netw.</source> <volume>99</volume>, <fpage>56</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2017.12.005</pub-id><pub-id pub-id-type="pmid">29328958</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klampfl</surname> <given-names>S.</given-names></name> <name><surname>Maass</surname> <given-names>W.</given-names></name></person-group> (<year>2013</year>). <article-title>Emergence of dynamic memory traces in cortical microcircuit models through STDP</article-title>. <source>J. Neurosci.</source> <volume>33</volume>, <fpage>11515</fpage>&#x02013;<lpage>11529</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.5044-12.2013</pub-id><pub-id pub-id-type="pmid">23843522</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kobayashi</surname> <given-names>R.</given-names></name> <name><surname>Tsubo</surname> <given-names>Y.</given-names></name> <name><surname>Shinomoto</surname> <given-names>S.</given-names></name></person-group> (<year>2009</year>). <article-title>Made-to-order spiking neuron model equipped with a multi-timescale adaptive threshold</article-title>. <source>Front. Comput. Neurosci.</source> <volume>3</volume>:<fpage>9</fpage>. <pub-id pub-id-type="doi">10.3389/neuro.10.009.2009</pub-id><pub-id pub-id-type="pmid">19668702</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>K&#x000F6;nig</surname> <given-names>P.</given-names></name> <name><surname>Engel</surname> <given-names>A. K.</given-names></name> <name><surname>Singer</surname> <given-names>W.</given-names></name></person-group> (<year>1996</year>). <article-title>Integrator or coincidence detector? The role of the cortical neuron revisited</article-title>. <source>Trends Neurosci.</source> <volume>19</volume>, <fpage>130</fpage>&#x02013;<lpage>137</lpage>. <pub-id pub-id-type="pmid">8658595</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krunglevicius</surname> <given-names>D.</given-names></name></person-group> (<year>2015</year>). <article-title>Competitive STDP learning of overlapping spatial patterns</article-title>. <source>Neural Comput.</source> <volume>27</volume>, <fpage>1673</fpage>&#x02013;<lpage>1685</lpage>. <pub-id pub-id-type="doi">10.1162/NECO_a_00753</pub-id><pub-id pub-id-type="pmid">26079753</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>LeCun</surname> <given-names>Y.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>436</fpage>&#x02013;<lpage>444</lpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id><pub-id pub-id-type="pmid">26017442</pub-id></citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2013</year>). <article-title>Neural variability, or lack thereof</article-title>. <source>Front. Comput. Neurosci.</source> <volume>7</volume>, <fpage>1</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.3389/fncom.2013.00007</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2017</year>). <article-title>STDP allows close-to-optimal spatiotemporal spike pattern detection by single coincidence detector neurons</article-title>. <source>Neuroscience</source> <pub-id pub-id-type="doi">10.1016/j.neuroscience.2017.06.032.</pub-id> [Epub ahead of print]. <pub-id pub-id-type="pmid">28668487</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masquelier</surname> <given-names>T.</given-names></name> <name><surname>Guyonneau</surname> <given-names>R.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name></person-group> (<year>2008</year>). <article-title>Spike timing dependent plasticity finds the start of repeating patterns in continuous spike trains</article-title>. <source>PLoS ONE</source> 3:e1377. <pub-id pub-id-type="doi">10.1371/journal.pone.0001377</pub-id><pub-id pub-id-type="pmid">18167538</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masquelier</surname> <given-names>T.</given-names></name> <name><surname>Guyonneau</surname> <given-names>R.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name></person-group> (<year>2009</year>). <article-title>Competitive STDP-based spike pattern learning</article-title>. <source>Neural Comput.</source> <volume>21</volume>, <fpage>1259</fpage>&#x02013;<lpage>1276</lpage>. <pub-id pub-id-type="doi">10.1162/neco.2008.06-08-804</pub-id><pub-id pub-id-type="pmid">19718815</pub-id></citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masquelier</surname> <given-names>T.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name></person-group> (<year>2007</year>). <article-title>Unsupervised learning of visual features through spike timing dependent plasticity</article-title>. <source>PLoS Comput. Biol.</source> <volume>3</volume>:<fpage>e31</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.0030031</pub-id><pub-id pub-id-type="pmid">17305422</pub-id></citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mensi</surname> <given-names>S.</given-names></name> <name><surname>Hagens</surname> <given-names>O.</given-names></name> <name><surname>Gerstner</surname> <given-names>W.</given-names></name> <name><surname>Pozzorini</surname> <given-names>C.</given-names></name></person-group> (<year>2016</year>). <article-title>Enhanced sensitivity to rapid input fluctuations by nonlinear threshold dynamics in neocortical pyramidal neurons</article-title>. <source>PLOS Comput. Biol.</source> <volume>12</volume>:<fpage>e1004761</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1004761</pub-id><pub-id pub-id-type="pmid">26907675</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mozafari</surname> <given-names>M.</given-names></name> <name><surname>Ganjtabesh</surname> <given-names>M.</given-names></name> <name><surname>Nowzari-Dalini</surname> <given-names>A.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2018a</year>). <article-title>Combining STDP and reward-modulated STDP in deep convolutional spiking neural networks for digit recognition</article-title>. <source>arXiv[preprint]</source> <volume>arXiv</volume>:<fpage>1804.00227</fpage>.</citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mozafari</surname> <given-names>M.</given-names></name> <name><surname>Kheradpisheh</surname> <given-names>S.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name> <name><surname>Nowzari-Dalini</surname> <given-names>A.</given-names></name> <name><surname>Ganjtabesh</surname> <given-names>M.</given-names></name></person-group> (<year>2018b</year>). <article-title>First-spike-based visual categorization using reward-modulated STDP</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst</source>. <pub-id pub-id-type="doi">10.1109/TNNLS.2018.2826721.</pub-id> [Epub ahead of print]. <pub-id pub-id-type="pmid">29993898</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nessler</surname> <given-names>B.</given-names></name> <name><surname>Pfeiffer</surname> <given-names>M.</given-names></name> <name><surname>Buesing</surname> <given-names>L.</given-names></name> <name><surname>Maass</surname> <given-names>W.</given-names></name></person-group> (<year>2013</year>). <article-title>Bayesian computation emerges in generic cortical microcircuits through spike-timing-dependent plasticity</article-title>. <source>PLoS Comput. Biol.</source> <volume>9</volume>:<fpage>e1003037</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1003037</pub-id><pub-id pub-id-type="pmid">23633941</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nguyen</surname> <given-names>A.</given-names></name> <name><surname>Dosovitskiy</surname> <given-names>A.</given-names></name> <name><surname>Yosinski</surname> <given-names>J.</given-names></name> <name><surname>Brox</surname> <given-names>T.</given-names></name> <name><surname>Clune</surname> <given-names>J.</given-names></name></person-group> (<year>2016</year>). <article-title>Synthesizing the preferred inputs for neurons in neural networks via deep generator networks</article-title> in <source>Advances in Neural Information Processing Systems</source>.</citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Olah</surname> <given-names>C.</given-names></name> <name><surname>Mordvintsev</surname> <given-names>A.</given-names></name> <name><surname>Schubert</surname> <given-names>L.</given-names></name></person-group> (<year>2017</year>). <article-title>Feature visualization</article-title>. <source>Distill</source>. <pub-id pub-id-type="doi">10.23915/distill.00007</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>O&#x00027;Reilly</surname> <given-names>R. C.</given-names></name></person-group> (<year>1998</year>). <article-title>Six principles for biologically based computational models of cortical cognition</article-title>. <source>Trends Cogn. Sci.</source> <volume>2</volume>, <fpage>455</fpage>&#x02013;<lpage>462</lpage>. <pub-id pub-id-type="pmid">21227277</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Palm</surname> <given-names>G.</given-names></name></person-group> (<year>2013</year>). <article-title>Neural associative memories and sparse coding</article-title>. <source>Neural Netw.</source> <volume>37</volume>, <fpage>165</fpage>&#x02013;<lpage>171</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2012.08.013</pub-id><pub-id pub-id-type="pmid">23043727</pub-id></citation>
</ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Panzeri</surname> <given-names>S.</given-names></name> <name><surname>Diamond</surname> <given-names>M. E.</given-names></name></person-group> (<year>2010</year>). <article-title>Information carried by population spike times in the whisker sensory cortex can be decoded without knowledge of stimulus time</article-title>. <source>Front. Synapt. Neurosci.</source> <volume>2</volume>:<fpage>17</fpage>. <pub-id pub-id-type="doi">10.3389/fnsyn.2010.00017</pub-id><pub-id pub-id-type="pmid">21423503</pub-id></citation>
</ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Platkiewicz</surname> <given-names>J.</given-names></name> <name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2011</year>). <article-title>Impact of fast sodium channel inactivation on spike threshold dynamics and synaptic integration</article-title>. <source>PLoS Comput. Biol.</source> <volume>7</volume>:<fpage>e1001129</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1001129</pub-id><pub-id pub-id-type="pmid">21573200</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Quian Quiroga</surname> <given-names>R.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name></person-group> (<year>2010</year>). <article-title>Measuring sparseness in the brain: comment on bowers (2009)</article-title>. <source>Psychol. Rev.</source> <volume>117</volume>, <fpage>291</fpage>&#x02013;<lpage>297</lpage>. <pub-id pub-id-type="doi">10.1037/a0016917</pub-id><pub-id pub-id-type="pmid">20063978</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Quiroga</surname> <given-names>R. Q.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name> <name><surname>Koch</surname> <given-names>C.</given-names></name> <name><surname>Fried</surname> <given-names>I.</given-names></name></person-group> (<year>2008</year>). <article-title>Sparse but not &#x0201C;Grandmother-cell&#x0201D; coding in the medial temporal lobe</article-title>. <source>Trends Cogn. Sci</source>. <volume>12</volume>, <fpage>87</fpage>&#x02013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2007.12.003</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rolls</surname> <given-names>E. T.</given-names></name> <name><surname>Treves</surname> <given-names>A.</given-names></name> <name><surname>Tovee</surname> <given-names>M. J.</given-names></name></person-group> (<year>1997</year>). <article-title>The representational capacity of the distributed encoding of information provided by populations of neurons in primate temporal visual cortex</article-title>. <source>Exp. Brain Res</source>. <volume>114</volume>, <fpage>149</fpage>&#x02013;<lpage>162</lpage>. <pub-id pub-id-type="pmid">9125461</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rossant</surname> <given-names>C.</given-names></name> <name><surname>Leijon</surname> <given-names>S.</given-names></name> <name><surname>Magnusson</surname> <given-names>A. K.</given-names></name> <name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2011</year>). <article-title>Sensitivity of noisy neurons to coincident inputs</article-title>. <source>J. Neurosci.</source> <volume>31</volume>, <fpage>17193</fpage>&#x02013;<lpage>17206</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2482-11.2011</pub-id><pub-id pub-id-type="pmid">22114286</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roy</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>The theory of localist representation and of a purely abstract cognitive system: the evidence from cortical columns, category cells, and multisensory neurons</article-title>. <source>Front. Psychol</source>. <volume>8</volume>:<fpage>186</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.00186</pub-id><pub-id pub-id-type="pmid">28261127</pub-id></citation>
</ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shoham</surname> <given-names>S.</given-names></name> <name><surname>O&#x00027;Connor</surname> <given-names>D. H.</given-names></name> <name><surname>Segev</surname> <given-names>R.</given-names></name></person-group> (<year>2006</year>). <article-title>How silent is the brain: is there a "dark matter" problem in neuroscience?</article-title> <source>J. Compar. Physiol.</source> <volume>192</volume>, <fpage>777</fpage>&#x02013;<lpage>784</lpage>. <pub-id pub-id-type="doi">10.1007/s00359-006-0117-6</pub-id><pub-id pub-id-type="pmid">16550391</pub-id></citation>
</ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>S.</given-names></name> <name><surname>Miller</surname> <given-names>K. D.</given-names></name> <name><surname>Abbott</surname> <given-names>L. F.</given-names></name></person-group> (<year>2000</year>). <article-title>Competitive hebbian learning through spike-timing-dependent synaptic plasticity</article-title>. <source>Nat. Neurosci.</source> <volume>3</volume>, <fpage>919</fpage>&#x02013;<lpage>926</lpage>. <pub-id pub-id-type="doi">10.1038/78829</pub-id><pub-id pub-id-type="pmid">10966623</pub-id></citation>
</ref>
<ref id="B53">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>H.</given-names></name> <name><surname>Sourina</surname> <given-names>O.</given-names></name> <name><surname>Huang</surname> <given-names>G.-B.</given-names></name></person-group> (<year>2016</year>). <article-title>Learning polychronous neuronal groups using joint weight-delay spike-timing-dependent plasticity</article-title>. <source>Neural Comput.</source> <volume>28</volume>, <fpage>2181</fpage>&#x02013;<lpage>2212</lpage>. <pub-id pub-id-type="doi">10.1162/NECO_a_00879</pub-id><pub-id pub-id-type="pmid">27557107</pub-id></citation>
</ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thorpe</surname> <given-names>S.</given-names></name></person-group> (<year>1989</year>). <article-title>Local vs. distributed coding</article-title>. <source>Intelletica</source> <volume>8</volume>, <fpage>3</fpage>&#x02013;<lpage>40</lpage>.</citation>
</ref>
<ref id="B55">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thorpe</surname> <given-names>S. J.</given-names></name></person-group> (<year>2009</year>). <article-title>Single units and sensation: still just as relevant today</article-title>. <source>Perception</source> <volume>38</volume>, <fpage>802</fpage>&#x02013;<lpage>803</lpage>; discussion: <fpage>804</fpage>&#x02013;<lpage>807</lpage>. <pub-id pub-id-type="pmid">19806958</pub-id></citation>
</ref>
<ref id="B56">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Thorpe</surname> <given-names>S. J.</given-names></name></person-group> (<year>2011</year>). <article-title>Grandmother cells and distributed representations</article-title> in <source>Visual Population Codes-Toward a Common Multivariate Framework for Cell Recording and Functional Imaging</source>, eds <person-group person-group-type="editor"><name><surname>Kriegeskorte</surname> <given-names>N.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name></person-group> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>), <fpage>23</fpage>&#x02013;<lpage>51</lpage>.</citation>
</ref>
<ref id="B57">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Thorpe</surname> <given-names>S. J.</given-names></name> <name><surname>Gautrais</surname> <given-names>J.</given-names></name></person-group> (<year>1998</year>). <article-title>Rank order coding</article-title> in <source>Computational Neuroscience : Trends in Research</source>, ed <person-group person-group-type="editor"><name><surname>Bower</surname> <given-names>J. M.</given-names></name></person-group> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Plenum Press</publisher-name>), <fpage>113</fpage>&#x02013;<lpage>118</lpage>.</citation>
</ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Rossum</surname> <given-names>M. C.</given-names></name> <name><surname>Bi</surname> <given-names>G. Q.</given-names></name> <name><surname>Turrigiano</surname> <given-names>G. G.</given-names></name></person-group> (<year>2000</year>). <article-title>Stable Hebbian learning from spike timing-dependent plasticity</article-title>. <source>J. Neurosci.</source> <volume>20</volume>, <fpage>8812</fpage>&#x02013;<lpage>8821</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.20-23-08812.2000</pub-id><pub-id pub-id-type="pmid">11102489</pub-id></citation>
</ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wolfe</surname> <given-names>J.</given-names></name> <name><surname>Houweling</surname> <given-names>A. R.</given-names></name> <name><surname>Brecht</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Sparse and powerful cortical spikes</article-title>. <source>Curr. Opin. Neurobiol.</source> <volume>20</volume>, <fpage>306</fpage>&#x02013;<lpage>312</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2010.03.006</pub-id><pub-id pub-id-type="pmid">20400290</pub-id></citation>
</ref>
<ref id="B60">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yger</surname> <given-names>P.</given-names></name> <name><surname>Stimberg</surname> <given-names>M.</given-names></name> <name><surname>Brette</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). <article-title>Fast learning with weak synaptic plasticity</article-title>. <source>J. Neurosci.</source> <volume>35</volume>, <fpage>13351</fpage>&#x02013;<lpage>13362</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0607-15.2015</pub-id><pub-id pub-id-type="pmid">26424883</pub-id></citation>
</ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>B.</given-names></name> <name><surname>Khosla</surname> <given-names>A.</given-names></name> <name><surname>Lapedriza</surname> <given-names>A.</given-names></name> <name><surname>Oliva</surname> <given-names>A.</given-names></name> <name><surname>Torralba</surname> <given-names>A.</given-names></name></person-group> (<year>2015</year>). <article-title>Object detectors emerge in deep scene CNNs</article-title>. <source>arXiv[preprint]</source> <volume>arXiv</volume>:<fpage>1412.6856</fpage>.</citation>
</ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>With a low number of synaptic inputs, the mean V is close to zero. Since V is non-negative, its distribution is not symmetric anymore, but positively skewed.</p></fn>
<fn id="fn0002"><p><sup>2</sup>When <italic>L</italic> is large (say tens of ms), STDP will typically not select all the afferents that fire in a full pattern, but only those that fire in a subsection of it, typically located at the beginning (Masquelier et al., <xref ref-type="bibr" rid="B33">2008</xref>; Gilson et al., <xref ref-type="bibr" rid="B12">2011</xref>; Masquelier, <xref ref-type="bibr" rid="B32">2017</xref>), unless competition forces the neurons to learn subsequent subsections (Masquelier et al., <xref ref-type="bibr" rid="B34">2009</xref>). The subsection duration depends on the parameters, and here we investigate the conditions under which this duration is optimal.</p></fn>
</fn-group>
</back>
</article> 