<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Neurosci.</journal-id>
<journal-title>Frontiers in Computational Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5188</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fncom.2021.743537</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Multiple-Timescale Neural Networks: Generation of History-Dependent Sequences and Inference Through Autonomous Bifurcations</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Kurikawa</surname> <given-names>Tomoki</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1411744/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Kaneko</surname> <given-names>Kunihiko</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/12724/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Physics, Kansai Medical University</institution>, <addr-line>Hirakata</addr-line>, <country>Japan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Basic Science, Graduate School of Arts and Sciences, University of Tokyo</institution>, <addr-line>Tokyo</addr-line>, <country>Japan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Center for Complex Systems Biology, Universal Biology Institute, University of Tokyo</institution>, <addr-line>Tokyo</addr-line>, <country>Japan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Spase Petkoski, INSERM U1106 Institut de Neurosciences des Syst&#x000E8;mes, France</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Adam Ponzi, Okinawa Institute of Science and Technology Graduate University, Japan; Claudius Gros, Goethe University Frankfurt, Germany</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Tomoki Kurikawa <email>kurikawt&#x00040;hirakata.kmu.ac.jp</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>10</day>
<month>12</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>15</volume>
<elocation-id>743537</elocation-id>
<history>
<date date-type="received">
<day>18</day>
<month>07</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>11</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Kurikawa and Kaneko.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Kurikawa and Kaneko</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract><p>Sequential transitions between metastable states are ubiquitously observed in the neural system and underlying various cognitive functions such as perception and decision making. Although a number of studies with asymmetric Hebbian connectivity have investigated how such sequences are generated, the focused sequences are simple Markov ones. On the other hand, fine recurrent neural networks trained with supervised machine learning methods can generate complex non-Markov sequences, but these sequences are vulnerable against perturbations and such learning methods are biologically implausible. How stable and complex sequences are generated in the neural system still remains unclear. We have developed a neural network with fast and slow dynamics, which are inspired by the hierarchy of timescales on neural activities in the cortex. The slow dynamics store the history of inputs and outputs and affect the fast dynamics depending on the stored history. We show that the learning rule that requires only local information can form the network generating the complex and robust sequences in the fast dynamics. The slow dynamics work as bifurcation parameters for the fast one, wherein they stabilize the next pattern of the sequence before the current pattern is destabilized depending on the previous patterns. This co-existence period leads to the stable transition between the current and the next pattern in the non-Markov sequence. We further find that timescale balance is critical to the co-existence period. Our study provides a novel mechanism generating robust complex sequences with multiple timescales. Considering the multiple timescales are widely observed, the mechanism advances our understanding of temporal processing in the neural system.</p></abstract>
<kwd-group>
<kwd>slow-fast systems</kwd>
<kwd>recurrent neural networks</kwd>
<kwd>bifurcations</kwd>
<kwd>sequential patterns</kwd>
<kwd>non-Markov sequences</kwd>
</kwd-group>
<contract-sponsor id="cn001">Japan Society for the Promotion of Science<named-content content-type="fundref-id">10.13039/501100001691</named-content></contract-sponsor>
<counts>
<fig-count count="8"/>
<table-count count="0"/>
<equation-count count="4"/>
<ref-count count="66"/>
<page-count count="13"/>
<word-count count="10228"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Sequentially activated patterns are widely observed in neural systems, for instance, the cerebral cortex (Jones et al., <xref ref-type="bibr" rid="B24">2007</xref>; Ponce-Alvarez et al., <xref ref-type="bibr" rid="B46">2012</xref>; Stokes et al., <xref ref-type="bibr" rid="B56">2013</xref>; Mazzucato et al., <xref ref-type="bibr" rid="B39">2015</xref>; Kurikawa et al., <xref ref-type="bibr" rid="B30">2018</xref>; Taghia et al., <xref ref-type="bibr" rid="B58">2018</xref>), hippocampus (HPC) (Gupta et al., <xref ref-type="bibr" rid="B16">2010</xref>; Maboudi et al., <xref ref-type="bibr" rid="B37">2018</xref>; Schuck and Niv, <xref ref-type="bibr" rid="B52">2019</xref>; Wimmer et al., <xref ref-type="bibr" rid="B65">2020</xref>), and the striatum (Akhlaghpour et al., <xref ref-type="bibr" rid="B1">2016</xref>). These patterns underlie a range of cognitive functions: perception (Jones et al., <xref ref-type="bibr" rid="B24">2007</xref>; Miller and Katz, <xref ref-type="bibr" rid="B42">2010</xref>), decision making (Ponce-Alvarez et al., <xref ref-type="bibr" rid="B46">2012</xref>), working memory (Stokes et al., <xref ref-type="bibr" rid="B56">2013</xref>; Taghia et al., <xref ref-type="bibr" rid="B58">2018</xref>), and recall of long-term memory (Wimmer et al., <xref ref-type="bibr" rid="B65">2020</xref>). They process temporal information by concatenating shorter sequences (Gupta et al., <xref ref-type="bibr" rid="B16">2010</xref>), reorganizing the order in sequential patterns (Wimmer et al., <xref ref-type="bibr" rid="B65">2020</xref>), and chunking sequences (Jin et al., <xref ref-type="bibr" rid="B22">2014</xref>), which lead to inference and recall, based on previous experiences.</p>
<p>Several models have been proposed to understand how such sequential patterns are shaped in the neural systems to perform complex tasks (Kleinfeld, <xref ref-type="bibr" rid="B27">1986</xref>; Sompolinsky and Kanter, <xref ref-type="bibr" rid="B55">1986</xref>; Seliger et al., <xref ref-type="bibr" rid="B53">2003</xref>; Gros, <xref ref-type="bibr" rid="B13">2007</xref>; Sussillo and Abbott, <xref ref-type="bibr" rid="B57">2009</xref>; Russo and Treves, <xref ref-type="bibr" rid="B51">2012</xref>; Laje and Buonomano, <xref ref-type="bibr" rid="B34">2013</xref>; Recanatesi et al., <xref ref-type="bibr" rid="B48">2015</xref>; Chaisangmongkon et al., <xref ref-type="bibr" rid="B8">2017</xref>; Haga and Fukai, <xref ref-type="bibr" rid="B17">2019</xref>). Popular Hebbian models provide a simple framework in which each pattern in the sequence is represented as a metastable state, which is formed through Hebbian learning. An asymmetric connection from the current to the successive pattern (Amari, <xref ref-type="bibr" rid="B2">1972</xref>; Kleinfeld, <xref ref-type="bibr" rid="B27">1986</xref>; Sompolinsky and Kanter, <xref ref-type="bibr" rid="B55">1986</xref>; Nishimori et al., <xref ref-type="bibr" rid="B44">1990</xref>; Seliger et al., <xref ref-type="bibr" rid="B53">2003</xref>; Gros, <xref ref-type="bibr" rid="B13">2007</xref>; Russo and Treves, <xref ref-type="bibr" rid="B51">2012</xref>; Recanatesi et al., <xref ref-type="bibr" rid="B48">2015</xref>; Haga and Fukai, <xref ref-type="bibr" rid="B17">2019</xref>) causes transition between patterns. Such transitions are also induced by slower destabilization terms (Gros, <xref ref-type="bibr" rid="B13">2007</xref>; Russo and Treves, <xref ref-type="bibr" rid="B51">2012</xref>; Recanatesi et al., <xref ref-type="bibr" rid="B48">2015</xref>). Note that these sequences are widely observed in neural systems (Miller, <xref ref-type="bibr" rid="B41">2016</xref>). In other studies (Sussillo and Abbott, <xref ref-type="bibr" rid="B57">2009</xref>; Laje and Buonomano, <xref ref-type="bibr" rid="B34">2013</xref>; Mante et al., <xref ref-type="bibr" rid="B38">2013</xref>; Chaisangmongkon et al., <xref ref-type="bibr" rid="B8">2017</xref>), recurrent neural networks (RNN) are trained by using machine learning methods so that experimentally observed neural dynamics are generated.</p>
<p>Despite the great success of these studies, however, some fundamental questions remain unanswered. In models that generate sequential metastable states, a transition between these states is embedded rigidly into the connectivity (i.e., the correlation between the current to the next pattern), resulting in successive patterns being determined by the immediately preceding pattern. Hence, the generation of sequences depending on the long history of the previous patterns is not possible. On the other hand, RNNs trained with machine learning methods allow for generating complex sequences dependent on history. The training methods require non-local information and have to retain the information until the sequence finishes, which is not biologically plausible. In addition, the formed sequences are vulnerable to noise or perturbation to the initial state (Laje and Buonomano, <xref ref-type="bibr" rid="B34">2013</xref>).</p>
<p>To resolve these unanswered questions, we introduce a neural network model with slow and fast neurons that can learn the history-dependent sequences and connect the sequences. The fast neural dynamics generate patterns in response to an external input with the feedback from the slow dynamics. The slow dynamics store the history of the inputs <italic>via</italic> the fast dynamics, and feed the stored information back to the fast, as shown in <xref ref-type="fig" rid="F1">Figure 1A</xref>. By this model, we provide a novel framework in temporal processing in the neural system in which the slow dynamics control successive bifurcations of fixed points of fast dynamics, based on the stored history of previous patterns and inputs. By adopting a biologically plausible learning rule based solely on the correlation between the pre- and post-synaptic neural activities as introduced previously (Kurikawa and Kaneko, <xref ref-type="bibr" rid="B31">2013</xref>, <xref ref-type="bibr" rid="B33">2016</xref>; Kurikawa et al., <xref ref-type="bibr" rid="B29">2020</xref>), we demonstrate that our model with the fast and slow neural dynamics memorizes the history-dependent sequences and enables inference based on them.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(A)</bold> Schematic diagram of the proposed model for two sequences (<italic>K</italic> &#x0003D; 2) and three patterns (<italic>M</italic> &#x0003D; 3). <bold>(B)</bold> Neural dynamics during the learning process of three targets. Top: the time series of one of the fast variables <italic>x</italic><sub>0</sub> (solid line) and the corresponding slow variable <italic>y</italic><sub>0</sub> (broken line) during the learning process. Bottom: <inline-formula><mml:math id="M1"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, overlaps of <bold><italic>x</italic></bold> with <inline-formula><mml:math id="M2"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> (blue), <inline-formula><mml:math id="M3"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> (orange), and <inline-formula><mml:math id="M4"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> (green). The black line represents the overlap between <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold> denoted as <italic>m</italic><sup><italic>xy</italic></sup>. The bars above the panels indicate the targeted patterns given to the network in corresponding periods. <bold>(C)</bold> The fraction of successful recalls is plotted as a function of <italic>M</italic> for <italic>K</italic> &#x0003D; 1, 2. It is averaged over 50 realizations (10 networks and five pairs of the target and input patterns for each network). Here, a successful recall is defined as the case in which all <italic>K</italic> &#x000D7; <italic>M</italic> targets are sequentially generated in the correct order in the presence of the corresponding inputs.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0001.tif"/>
</fig>
<p>Multiple-timescale neural dynamics are observed across cortical areas (Honey et al., <xref ref-type="bibr" rid="B19">2012</xref>; Murray et al., <xref ref-type="bibr" rid="B43">2014</xref>; Chaudhuri et al., <xref ref-type="bibr" rid="B10">2015</xref>; Hasson et al., <xref ref-type="bibr" rid="B18">2015</xref>). Neural activities in lower sensory cortices change in a faster timescale and respond instantaneously to stimuli, whereas those in higher association cortices change in a slower timescale and integrate information over longer periods. Cooperations between the higher and lower cortices are necessary to process the temporal information.</p>
<p>Some model studies focused on the multiple-timescale dynamics (Kiebel et al., <xref ref-type="bibr" rid="B25">2008</xref>; Yamashita and Tani, <xref ref-type="bibr" rid="B66">2008</xref>) and showed that their models generate history-dependent sequences. These models, however, adopted machine learning methods, and thus, biological plausibility is hard to be assured. In contrast, our model proposes a biologically plausible mechanism in which the higher cortices regulate the lower ones to generate complex sequences.</p>
<p>In the following, we focus on two basic aspects of neural sequences in temporal information processing; context (history)-dependent sequences and inference. In the context-dependent working memory task (Mante et al., <xref ref-type="bibr" rid="B38">2013</xref>; Stokes et al., <xref ref-type="bibr" rid="B56">2013</xref>), distinct sequences of neural patterns are evoked by identical stimuli depending on the preceding context signals. Second, in this study, inference is defined as the ability to make appropriate responses against a new environment by using previously learned examples. For instance (Jones et al., <xref ref-type="bibr" rid="B23">2012</xref>; Wikenheiser and Schoenbaum, <xref ref-type="bibr" rid="B64">2016</xref>), consider a rat learning successive stimuli, A followed by B, and then reward C. After changing the environment, the rat is required to learn a new combination of stimuli, A&#x00027; followed by B. In this situation, the rat is able to infer that stimuli A&#x00027; causes the reward C <italic>via</italic> B. Neural activities reflecting this cognitive function should show sequential patterns A&#x00027;BC even after learning only A&#x00027;B. After showing basic behaviors in our model, we demonstrate that how such a context-dependent sequence is generated and how inference is executed.</p></sec>
<sec sec-type="materials and methods" id="s2">
<title>2. Materials and Methods</title>
<sec>
<title>2.1. Neural Model</title>
<p>We consider learning of <italic>K</italic> sequences, each of which contains <italic>M</italic> patterns, with <italic>K</italic> input patterns. We denote the &#x003BC;-th targeted pattern in the &#x003B1;-th sequence as <inline-formula><mml:math id="M5"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, and the corresponding input as <bold><italic>&#x003B7;</italic></bold><sup>&#x003B1;</sup> for &#x003BC; &#x0003D; 1, 2, &#x022EF;&#x02009;, <italic>M</italic> over the inputs &#x003B1; &#x0003D; 1, &#x022EF;&#x02009;, <italic>K</italic>. <xref ref-type="fig" rid="F1">Figure 1A</xref> illustrates the case with <italic>K</italic> &#x0003D; 2 and <italic>M</italic> &#x0003D; 3: In this case, a given sequence <inline-formula><mml:math id="M6"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>,<inline-formula><mml:math id="M7"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>,<inline-formula><mml:math id="M8"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> (&#x003B1; &#x0003D; 1, 2) should be generated upon a given corresponding input <bold><italic>&#x003B7;</italic></bold><sup>&#x003B1;</sup>. Generally, a pattern to be generated next is determined not only by the current pattern but also by earlier patterns. Thus, a network has to retain the history of previous patterns to generate a sequence correctly.</p>
<p>To achieve this, we built a two-population model with different timescales, one with <italic>N</italic> fast neurons and one with <italic>N</italic> slow neurons, denoted as <italic>X</italic> and <italic>Y</italic>, respectively. <italic>X</italic> receives an external input, and <italic>Y</italic> receives the output from <italic>X</italic> and provides input to <italic>X</italic>, as shown in <xref ref-type="fig" rid="F1">Figure 1A</xref>. The neural activities <italic>x</italic><sub><italic>i</italic></sub> in <italic>X</italic> and <italic>y</italic><sub><italic>i</italic></sub> in <italic>Y</italic> evolve according to the following equation:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x02219;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mo class="qopname">tanh</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E2"><label>(2)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x02219;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mo class="qopname">tanh</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E3"><label>(3)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mo class="qopname">tanh</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M12"><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02260;</mml:mo><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>; <inline-formula><mml:math id="M13"><mml:msub><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mi>Y</mml:mi></mml:mrow></mml:msubsup><mml:mo class="qopname">tanh</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. <inline-formula><mml:math id="M14"><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is a recurrent connection from the <italic>j</italic>-th to the <italic>i</italic>-th neuron in <italic>X</italic>, and <inline-formula><mml:math id="M15"><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mi>Y</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is a connection from the <italic>j</italic>-th neuron in <italic>Y</italic> to the <italic>i</italic>-th neuron in <italic>X</italic>. <italic>J</italic><sup><italic>X</italic></sup> is a fully connected network without self connections. It is modified during the learning process as described in the following subsection &#x0201C;Learning model&#x0201D; and initialized with the binary values <inline-formula><mml:math id="M16"><mml:mi>P</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>. The diagonal entries of <italic>J</italic><sup><italic>X</italic></sup> are kept at zero during the entire learning process. <italic>J</italic><sup><italic>XY</italic></sup>, in contrast, is a non-plastic sparse network; <inline-formula><mml:math id="M17"><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mi>Y</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:mi>c</mml:mi><mml:msup><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>&#x003C1;</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M18"><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mi>Y</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mn>2</mml:mn><mml:mi>&#x003C1;</mml:mi></mml:math></inline-formula>. <italic>X</italic> is required to generate the pattern <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> in the presence of <bold><italic>&#x003B7;</italic></bold><sup>&#x003B1;</sup>, i.e., an attractor that matches <inline-formula><mml:math id="M20"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is generated under <bold><italic>&#x003B7;</italic></bold><sup>&#x003B1;</sup>. The <italic>i</italic>-th element of a targeted pattern denoted as <inline-formula><mml:math id="M21"><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, is assigned to the <italic>i</italic>-th neuron in <italic>X</italic>, and randomly sampled according to the probability <inline-formula><mml:math id="M22"><mml:mi>P</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>. The input <inline-formula><mml:math id="M23"><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is injected to the <italic>i</italic>-th neuron in <italic>X</italic>, randomly sampled according to <inline-formula><mml:math id="M24"><mml:mi>P</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>. <bold><italic>&#x003BE;</italic></bold> and <bold><italic>&#x003B7;</italic></bold> are the same dimensional vectors as the fast dynamics, i.e., <italic>N</italic>-dimensional vectors. We set <italic>N</italic> &#x0003D; 100, &#x003B2;<sub><italic>x</italic></sub> &#x0003D; 2, &#x003B2;<sub><italic>y</italic></sub> &#x0003D; 20, &#x003C4;<sub><italic>x</italic></sub> &#x0003D; 1, &#x003C4;<sub><italic>y</italic></sub> &#x0003D; 100, &#x003C1; &#x0003D; 0.05, and <italic>c</italic> &#x0003D; 7. The dependence of the performance on these parameters are shown in the <xref ref-type="supplementary-material" rid="SM1">Supplementary Materials</xref>, while the details for the parameter setting are described.</p>
<p>In Equation (3), we implement the non-linear activation at the apical dendrites (Larkum et al., <xref ref-type="bibr" rid="B36">2009</xref>). We assumed that the input from <italic>Y</italic> to <italic>X</italic> innervated on the apical dendrites of neurons in <italic>X</italic>, which is consistent with observations that the feedback inputs from the higher cortical areas innervate on the apical dendrites of layer 5 pyramidal neurons (Larkum, <xref ref-type="bibr" rid="B35">2013</xref>), whereas the recurrent input from <italic>X</italic> to <italic>X</italic> was assumed to innervate the proximal dendrites. The synaptic inputs to the apical dendrites are integrated and evoke the calcium spike when the integrated input exceeds the threshold of spikes (refer to the detail of this type of spike in Larkum et al., <xref ref-type="bibr" rid="B36">2009</xref>). To reproduce this information processing at the apical dendrites, we used two nonlinear filters by the hyperbolic tangent function for the input from <italic>Y</italic> to <italic>X</italic>. First, by adopting tanh(<italic>y</italic><sub><italic>j</italic></sub>), the activity of the <italic>j</italic>-th neuron in <italic>Y</italic> is amplified in a nonlinear way at a synapse onto the neuron <italic>x</italic><sub><italic>i</italic></sub>. Second, tanh(<italic>r</italic><sub><italic>i</italic></sub>) represents calcium spike at the branching point of the tuft dendrite. Even if these hyperbolic tangent functions are not leveraged in Equation (3), the behavior of the model is not changed qualitatively, although the performance of the model is reduced.</p></sec>
<sec>
<title>2.2. Learning Model</title>
<p>Only <italic>J</italic><sup><italic>X</italic></sup> changes to generate the target according to the following equation:</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mi>y</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mover accent="true"><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>&#x02219;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mi>N</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003C4;<sub><italic>syn</italic></sub> is the learning speed (set to 100). This learning rule comprises a combination of a Hebbian term between the target and the presynaptic neuron, and an anti-Hebbian term between the pre- and post-synaptic neurons with a decay term <inline-formula><mml:math id="M26"><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> for normalization <xref ref-type="fn" rid="fn0001"><sup>1</sup></xref>. This form satisfies locality across connections and is biologically plausible (Kurikawa et al., <xref ref-type="bibr" rid="B29">2020</xref>). We previously applied this learning rule to a single network of <italic>X</italic> and demonstrated that the network learns <italic>K</italic> maps between inputs and targets, i.e., <italic>M</italic> &#x0003D; 1 (Kurikawa and Kaneko, <xref ref-type="bibr" rid="B31">2013</xref>, <xref ref-type="bibr" rid="B33">2016</xref>; Kurikawa et al., <xref ref-type="bibr" rid="B29">2020</xref>). However, in that case, generating a sequence (<italic>M</italic>&#x02265;2) was not possible. In the present study, there are two inputs for <italic>X</italic>, one from input <bold><italic>&#x003B7;</italic></bold> and one from <italic>Y</italic> that stores previous information. Thus, the network can generate a pattern depending not only on the present input pattern, but also on the previous patterns.</p></sec>
<sec>
<title>2.3. Learning Procedure</title>
<p>In our model, the patterns in the sequence are learned sequentially. A learning step of a single pattern is accomplished when the neural dynamics satisfy the following two criteria: <bold><italic>x</italic></bold> sufficiently approaches the target pattern, i.e., <inline-formula><mml:math id="M28"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x02261;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003A3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>/</mml:mo><mml:mi>N</mml:mi><mml:mo>&#x0003E;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>85</mml:mn></mml:math></inline-formula>, and <bold><italic>y</italic></bold> is sufficiently close to <bold><italic>x</italic></bold>, i.e., &#x003A3;<sub><italic>i</italic></sub><italic>x</italic><sub><italic>i</italic></sub><italic>y</italic><sub><italic>i</italic></sub>/<italic>N</italic>&#x0003E;0.5. After the completion of one learning step, a new pattern <inline-formula><mml:math id="M29"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> is presented instead of <inline-formula><mml:math id="M30"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> with a perturbation of fast variables <italic>x</italic><sub><italic>i</italic></sub>, by multiplying a random number uniformly sampled from zero to one. We execute these steps sequentially from &#x003BC; &#x0003D; 1 to <italic>M</italic> to learn a sequence once, denoted as one epoch of the learning. Before finishing the learning process, this procedure is repeated 20 times (i.e., 20 epochs). The second criterion for terminating the learning step is introduced for memorizing the sequences, especially the history-dependent sequences. Further, the value 0.5 of this criterion must take an intermediate value. If this criterion is not adopted or this criterion value is small, the target pattern is switched as soon as <bold><italic>x</italic></bold> is close to the target during the learning process. At this time, <bold><italic>y</italic></bold> is far from <bold><italic>x</italic></bold> because <bold><italic>y</italic></bold> is much slower than <bold><italic>x</italic></bold>. In this case, <bold><italic>y</italic></bold> cannot store any information about <bold><italic>x</italic></bold>. On the other side, when the value is close to unity, <bold><italic>y</italic></bold> matches <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold> can store only the present <bold><italic>x</italic></bold>. In both cases, <bold><italic>y</italic></bold> cannot store the history of <bold><italic>x</italic></bold>.</p>
<sec>
<title>2.3.1. Inference Task</title>
<p>In the inference task, we present sequentially different inputs in a sequence, whereas a single input is applied for a sequence in other tasks. We include the super- and sub-scripts in the notation of <bold><italic>&#x003B7;</italic></bold> as <inline-formula><mml:math id="M31"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> that represents the &#x003BC;-th input pattern in the &#x003B1;-th sequence. In this task, a network learns three sequences:(<italic>S, A, B, C</italic>), (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>), and (<italic>D</italic>). The former two sequences are used for inference, whereas the last one is a distractor to prevent the over-stability of the other two sequences. First, in the learning process, the network learns (<italic>S, A, B, C</italic>) and (<italic>D</italic>). Second, it learns (<italic>S, A</italic>&#x02032;, <italic>B</italic>) after training is completed. Then, we examine if the network generates (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>), implying inference from <italic>B</italic> to <italic>C</italic>.</p>
<p>For the first sequence (<italic>S, A, B, C</italic>), we apply <inline-formula><mml:math id="M32"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>s</mml:mi></mml:math></inline-formula> for the target <italic>S</italic>, <inline-formula><mml:math id="M33"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>a</mml:mi></mml:math></inline-formula> for <italic>A</italic>, and <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>b</mml:mi></mml:math></inline-formula> for <italic>B</italic> and <italic>C</italic>. For the second sequence (<italic>A</italic>&#x02032;, <italic>B, C</italic>), we apply <inline-formula><mml:math id="M35"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>s</mml:mi></mml:math></inline-formula> for the target <italic>S</italic>, <inline-formula><mml:math id="M36"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> for <italic>A</italic>&#x02032;, and <inline-formula><mml:math id="M37"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003B7;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>b</mml:mi></mml:math></inline-formula> for <italic>B</italic>. All the targets and inputs are randomly sampled according to the probability <inline-formula><mml:math id="M38"><mml:mi>P</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x000B1;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>.</p>
<p>Because, in this task, the input pattern is changed in a single sequence, we apply the input and target over 100 unit times and change them to the next in the sequence through the learning process. In the recall process, we regularly change the input pattern every 100 unit times, independently of the value of the neural activities.</p></sec></sec>
<sec>
<title>2.4. Data Analysis</title>
<sec>
<title>2.4.1. Principal Component Analysis (PCA)</title>
<p>We analyzed neural trajectories by using mainly PCA in <xref ref-type="fig" rid="F2">Figures 2</xref>, <bold>5</bold>&#x02013;<bold>7</bold> and <xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S5</xref>. The <italic>N</italic> &#x000D7; <italic>T</italic> dimensional neural data of <bold><italic>x</italic></bold> is used for the PCA. In this study, <italic>T</italic> is the duration time to analyze neural dynamics multiplied by the sampling number of <bold><italic>x</italic></bold> per unit time (In this case, 20). For analysis of <italic>y</italic>, we also used PCs obtained by x.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Bifurcation of <bold><italic>x</italic></bold> with quenched <bold><italic>y</italic></bold>. <bold>(A)</bold> Neural dynamics during the recall process of the three learned patterns. Overlaps of neural activities <inline-formula><mml:math id="M39"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, &#x003BC; &#x0003D; 1, 2, 3 in <bold><italic>x</italic></bold> (top) and <bold><italic>y</italic></bold> (bottom) for <italic>M</italic> &#x0003D; 3 are plotted in the same color as shown in <xref ref-type="fig" rid="F1">Figure 1B</xref>. <bold><italic>y</italic></bold> is sampled from the trajectory at 200 &#x0003C; <italic>t</italic> &#x0003C;500 for the bifurcation diagram of <bold><italic>x</italic></bold> shown in <bold>(B)</bold>. <bold>(B)</bold> Bifurcation diagram of <bold><italic>x</italic></bold> as quenched <bold><italic>y</italic></bold> is updated with the sampling time. Fixed points of <bold><italic>x</italic></bold> are shown by projecting to the first principal component (PC1) of principal component analysis (PCA). Small circles indicate fixed points with small basins: neural activity beginning only from the vicinity of the target converges to these points. Large circles represent fixed points with large basins: neural activities from the initial states converge to these points. To identify fixed points, the neural states are plotted after the transient period. Colored lines indicate the locations of the targets (<inline-formula><mml:math id="M40"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> in blue, orange, and green, respectively). Vertical arrows show the transitions of <bold><italic>x</italic></bold> to different targets in the recall process. <bold>(C)</bold> The neural dynamics for a given <bold><italic>y</italic></bold> at <italic>t</italic> &#x0003D; 225, 285, 335, 375 shadowed in <bold>(B)</bold> are depicted by projecting <bold><italic>x</italic></bold> to the 2-dimensional principal component (PC) space [PC1 is same as that in <bold>(B)</bold>]. Fifteen trajectories (three from the vicinity of the target, and others from random initial states) are plotted. Large and small circles represent fixed points given in <bold>(B)</bold>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0002.tif"/>
</fig>
</sec>
<sec>
<title>2.4.2. Calculation of Success Rate in the Inference Task</title>
<p>To compute the success rate of the generation of the sub-sequence (<italic>B, C</italic>) under input <italic>b</italic>, we first identify the sequence of the patterns from the continuous neural activity by setting a threshold for the overlap value at 0.7. In the recall process, since the overlap with either (or a few) of the targets <italic>S, A, B, C, D, A</italic>&#x02032; is selectively high most of the time, the sequences composed of some of the patterns <italic>S, A, B, C, D, A</italic>&#x02032; are obtained for each of different input sequences (<italic>s, a, b, b</italic>), (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>), and (<italic>s, v, b, b</italic>). In this study, <italic>v</italic> is a random pattern that is not used for learning. We generate these sequential patterns starting from 20 initial states for each of 100 network realizations and use the subpart of them in the presence of <italic>b</italic> to calculate the success rate of the sub-sequence (<italic>B, C</italic>). In this study, 100 network realizations are obtained by generating <italic>J</italic><sup><italic>X</italic></sup>, <italic>J</italic><sup><italic>XY</italic></sup>, <italic><bold>&#x003BE;</bold></italic>, and <bold>&#x003B7;</bold> 100 times according to independent and identically probability distributions as described in the &#x0201C;Neural model&#x0201D; in this section.</p>
<p>We test significant differences in the success rates of the generation of the sub-sequence (<italic>B, C</italic>) for different input sequences by Wilcoxon&#x00027;s signed-rank test. The success rates for (<italic>s, a</italic>&#x02032;, <italic>b</italic>) and (<italic>s, v, b</italic>) are calculated from 20 sequences for each network realization. Hundred samples of two related paired success rates are obtained and used in Wilcoxon&#x00027;s signed-rank test.</p></sec></sec></sec>
<sec sec-type="results" id="s3">
<title>3. Results</title>
<p>Before exploring the history-dependent sequence, we analyzed if our learning rule generates simple sequences, namely, sequences in which the successive pattern is determined solely by the current pattern. <xref ref-type="fig" rid="F1">Figure 1B</xref> shows a sample learning process for <italic>K</italic> &#x0003D; 1. We applied <bold><italic>&#x003B7;</italic></bold><sup>1</sup> to a network and presented <inline-formula><mml:math id="M44"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> as the first pattern of a target sequence. After the transient time, <bold><italic>x</italic></bold> converges to <inline-formula><mml:math id="M45"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> due to synaptic change. <bold><italic>y</italic></bold> follows <bold><italic>x</italic></bold> according to Equation 2 and, consequently, moves to the target.</p>
<p>We present an example of a recall process after the learning process for (<italic>K, M</italic>) &#x0003D; (1, 3) in <xref ref-type="fig" rid="F2">Figure 2A</xref>. In recall, the connectivity is not changed. The initial states of the fast variables are set at random values sampled from a uniform distribution of &#x02013;1 to 1. The slow variables are set at values of their final states in the learning process in order for the network to generate the sequence starting from <inline-formula><mml:math id="M46"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>. (If the slow variables are set randomly in a similar manner as the fast variables, the network can still generate the sequence, but the first pattern of the sequence is not <inline-formula><mml:math id="M47"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>). The targets appear sequentially in <italic>X</italic> in order. Note that in the recall process, the transition occurs spontaneously without any external operation.</p>
<p>We explored the success rate of the learning and found that increasing <italic>M</italic> and <italic>K</italic> generally leads to a decrease in the success rate of recalls. For <italic>N</italic> &#x0003D; 100 and <italic>K</italic> &#x0003D; 1, the success rate is over 80% for <italic>M</italic> &#x0003D; 1 up to 11, and decreases beyond <italic>M</italic> &#x0003D; 12. For <italic>K</italic> &#x0003D; 2, the success rate is approximately 80% for <italic>M</italic> &#x0003D; 3 and decreases gradually as <italic>M</italic> increases (<xref ref-type="fig" rid="F1">Figure 1C</xref>, refer to the <xref ref-type="supplementary-material" rid="SM1">Supplementary Material</xref> for detailed results). Furthermore, we investigated how the balance between the timescale of the slow variables &#x003C4;<sub><italic>y</italic></sub> and that of learning &#x003C4;<sub><italic>syn</italic></sub> affect the success rate.</p>
<p>Next, the spontaneous activities without the input are analyzed. <xref ref-type="fig" rid="F3">Figure 3A</xref> exemplifies the characteristic behavior of the complex spontaneous dynamics: fast oscillating activities and slowly varying ones appear alternatively. In the period of the oscillating activities, the memorized patterns are activated sequentially, which are not all memorized patterns, but their subsets, as shown in <xref ref-type="fig" rid="F3">Figures 3B&#x02013;D</xref>. Different subsets of patterns appear intermittently. For instance, (<bold><italic>&#x003BE;</italic><sup>1</sup></bold>, <bold><italic>&#x003BE;</italic><sup>2</sup></bold>, <bold><italic>&#x003BE;</italic><sup>5</sup></bold>), (<bold><italic>&#x003BE;</italic><sup>3</sup></bold>, <bold><italic>&#x003BE;</italic><sup>4</sup></bold>, <bold><italic>&#x003BE;</italic><sup>5</sup></bold>), and (<bold><italic>&#x003BE;</italic><sup>1</sup></bold>, <bold><italic>&#x003BE;</italic><sup>4</sup></bold>, <bold><italic>&#x003BE;</italic><sup>5</sup></bold>) are observed in <xref ref-type="fig" rid="F3">Figures 3B&#x02013;D</xref>, respectively. In contrast, in the period of the slowly varying activities, one or few patterns are stable for a while and then are collapsed.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>The neural dynamics without input after the learning process for (<italic>K, M</italic>)=(1,5). <bold>(A)</bold> (Upper) Overlaps of the spontaneous fast dynamics with the targets <inline-formula><mml:math id="M41"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, &#x003BC; &#x0003D; 1, 2, 3, 4, 5 are plotted in different colors indicated at bottom of the panels. (Lower) Overlaps of the spontaneous slow dynamics <inline-formula><mml:math id="M42"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, &#x003BC; &#x0003D; 1, 2, 3, 4, 5 are plotted in the same colors as the top panel. <bold>(B&#x02013;D)</bold> The enlarged view of spontaneous dynamics <inline-formula><mml:math id="M43"><mml:msubsup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> in <bold>(A)</bold> are shown. The enlarged time span adopted in each panel is indicated by the corresponding black bar at the top of <bold>(A)</bold>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0003.tif"/>
</fig>
<sec>
<title>3.1. Bifurcations of Fast Neural Dynamics</title>
<p>To elucidate how such a recall is possible, we analyzed the phase space of <bold><italic>x</italic></bold> with <bold><italic>y</italic></bold> quenched. In other words, <bold><italic>y</italic></bold> is regarded as bifurcation parameters for the fast dynamics. Specifically, we focused on the neural dynamics for 200 &#x02264; <italic>t</italic> &#x02264; 500, as shown in <xref ref-type="fig" rid="F2">Figure 2A</xref>. In this period, the fast dynamics show transitions from <inline-formula><mml:math id="M48"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M49"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> at <italic>t</italic> &#x0003D; 290, from <inline-formula><mml:math id="M50"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M51"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> at <italic>t</italic> &#x0003D; 375, and from <inline-formula><mml:math id="M52"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M53"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> at <italic>t</italic> &#x0003D; 220, 460. We sampled the slow variables every five units of time from <italic>t</italic> &#x0003D; 200 to 500, <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 200</sub>, <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 205</sub>, &#x022EF;&#x02009;, <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 500</sub>, along the trajectory, and analyzed the dynamics of <bold><italic>x</italic></bold> with the slow variables quenched at each sampled <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 200, 205, &#x022EF;&#x02009;, 500</sub>. <xref ref-type="fig" rid="F2">Figure 2B</xref> shows the bifurcation diagram of <bold><italic>x</italic></bold> against the change in <bold><italic>y</italic></bold>, and <xref ref-type="fig" rid="F2">Figure 2C</xref> shows the trajectories of <bold><italic>x</italic></bold> for specific <bold><italic>y</italic></bold>.</p>
<p>We now consider the neural dynamics for <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 225</sub>, just after the transition from <inline-formula><mml:math id="M54"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M55"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> [<xref ref-type="fig" rid="F2">Figure 2C</xref> (i)]. For this <bold><italic>y</italic></bold>, a single fixed point corresponding to the present pattern (<inline-formula><mml:math id="M56"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>) exists, leading to its stability against noise. As <bold><italic>y</italic></bold> is changed, the basin of <inline-formula><mml:math id="M57"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> shrinks, while a fixed point corresponding to the next target <inline-formula><mml:math id="M58"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> appears, and its basin expands <xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>, as shown in <xref ref-type="fig" rid="F2">Figure 2C</xref> (ii). At <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 290</sub>, the fixed point <inline-formula><mml:math id="M60"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> becomes unstable. Thus, the neural state <bold><italic>x</italic></bold> at <inline-formula><mml:math id="M61"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> goes out of there, and falls on <inline-formula><mml:math id="M62"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>, i.e., a transition occurs.</p>
<p>With a further shift of <bold><italic>y</italic></bold>, <bold><italic>y</italic></bold><sub><italic>t</italic> &#x0003D; 295, 300, &#x022EF;</sub>, a regime of coexistence of <inline-formula><mml:math id="M63"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M64"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> with large basins appears [<xref ref-type="fig" rid="F2">Figure 2C</xref> (iii)]. The basin of the attractor <inline-formula><mml:math id="M65"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> shrinks and vanishes [<xref ref-type="fig" rid="F2">Figure 2C</xref> (iv)], and the transition from <inline-formula><mml:math id="M66"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M67"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> occurs at <italic>t</italic> &#x0003D; 375. The next transition from <inline-formula><mml:math id="M68"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M69"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> occurs in the same manner at <italic>t</italic> &#x0003D; 460. These processes provide the mechanism for robust sequential recall: fixed points <bold><italic>x</italic></bold> of the current and successive targets coexist, and then, the current target becomes unstable when the slow variables change.</p>
<p>To examine the robustness of the recall, we explored trajectories from different initial conditions with Gaussian white noise with strength <italic>s</italic> (refer to <xref ref-type="supplementary-material" rid="SM1">Supplementary Material</xref> for details). All of these trajectories converge correctly to a target sequence after some transient period for weak noise. By increasing the noise strength, the recall performance of noisy dynamics is made equal to that of the noiseless dynamics up to noise strength <italic>s</italic> &#x0003D; 0.3. For stronger noise, the duration of residence at the target is decreased, because the neural state of <bold><italic>x</italic></bold> is kicked out of the target earlier than in the noiseless case. Even upon applying a strong and instantaneous perturbation to both <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold>, the trajectory recovers the correct sequence. The sequence is represented as a limit cycle containing <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold> and, thus, is recalled robustly.</p></sec>
<sec>
<title>3.2. Inference by Concatenation</title>
<p>Next, we test if our model flexibly infers new sequences based on the previously learned sequence. To this end, we consider the following task (Refer to Materials and methods for details). First, a network learns a sequence <inline-formula><mml:math id="M70"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> in response to the associated input sequences <inline-formula><mml:math id="M71"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. In addition, we provide <inline-formula><mml:math id="M72"><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>=</mml:mo><mml:mi>s</mml:mi></mml:math></inline-formula> associated with <inline-formula><mml:math id="M73"><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>=</mml:mo><mml:mi>S</mml:mi></mml:math></inline-formula> preceding the sequence as a fixation cue and the response to it (e. g., the subject&#x00027;s gaze to the fixation point), respectively. After the learning is completed, the network should generate the sequence (<italic>S, A, B, C</italic>) in response to the input sequence (<italic>s, a, b, b</italic>). Then, the network learns a new sequence <inline-formula><mml:math id="M74"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>S</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>B</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, which is associated with <inline-formula><mml:math id="M75"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. In this study, we intend to examine if the inference of <italic>C</italic> is achieved selectively from (<italic>S, A</italic>&#x02032;, <italic>B</italic>) in the presence of <italic>b</italic>. For it, we need to prevent the tight and trivial association between input <italic>b</italic> and the sub-sequence (<italic>B, C</italic>). For this purpose, the network is also postulated to learn the association between <inline-formula><mml:math id="M76"><mml:mstyle mathvariant="bold-italic"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>=</mml:mo><mml:mi>b</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M77"><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>=</mml:mo><mml:mi>D</mml:mi></mml:math></inline-formula> as a distractor. We explore if the network generates the sequence (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>) in response to the input sequence (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>) after learning the association between (<italic>S, A</italic>&#x02032;, <italic>B</italic>) and (<italic>s, a</italic>&#x02032;, <italic>b</italic>).</p>
<p>During the learning of the first sequence, the overlaps with all of the targets reach more than 0.9 after 20 epochs of learning, as shown in <xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S4A</xref>. Actually, <xref ref-type="fig" rid="F4">Figure 4A</xref> shows that the first sequence is successfully generated. Next, the network learns the new sequential patterns (<italic>S, A</italic>&#x02032;, <italic>B</italic>). If the network infers (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>) by using the already learned sub-sequence (<italic>B, C</italic>), it generates the sequence (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>) after learning only (<italic>S, A</italic>&#x02032;, <italic>B</italic>) without (<italic>S, A</italic>&#x02032;, <italic>B, C</italic>). As expected, the average overlaps with all of the targets in the second sequence (not only <italic>A</italic>&#x02032;,<italic>B</italic>,but also <italic>C</italic>) are increased through learning (<xref ref-type="fig" rid="F4">Figure 4B</xref>). After the fifth epoch of learning, the overlap with <italic>C</italic> declines, whereas those with <italic>A</italic>&#x02032; and <italic>B</italic> continuously increase. As an example, we plot the recall dynamics after learning <italic>A</italic>&#x02032; and <italic>B</italic> in <xref ref-type="fig" rid="F4">Figure 4C</xref>. <italic>A</italic>&#x02032; evokes <italic>B</italic> and <italic>C</italic>, although the overlap with the first target <italic>A</italic>&#x02032; is not large. Simultaneously, the network generates the first sequence (<xref ref-type="supplementary-material" rid="SM1">Supplementary Figure</xref> <xref ref-type="supplementary-material" rid="SM1">S4C</xref>). Note that the sub-sequence (<italic>B, C</italic>) is generated selectively by the input either (<italic>s, a, b, b</italic>) or (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>). In contrast, when a random input <italic>v</italic> is given instead of <italic>a</italic> or <italic>a</italic>&#x02032; in an input sequence, the sub-sequence (<italic>B, C</italic>)is not evoked, as shown in <xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S4B</xref>. To examine the difference among the recall dynamics in response to (<italic>s, a, b</italic>), (<italic>s, a</italic>&#x02032;, <italic>b</italic>), and (<italic>s, v, b</italic>), the success rate of generating the sub-sequence (refer to its definition in Materials and methods) is analyzed statistically in <xref ref-type="fig" rid="F4">Figure 4D</xref>. We found that the input sequences (<italic>s, a, b</italic>), (<italic>s, a</italic>&#x02032;, <italic>b</italic>) evoke the sub-sequence (<italic>B, C</italic>) with a significantly higher rate than the input sequence (<italic>s, v, b</italic>). Thus, our model is able to infer a new sequence based on the previously learned sequence.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Neural activities in the inference task. <bold>(A)</bold> The neural dynamics in the recall of the first sequence (<italic>S, A, B, C</italic>) in response to (<italic>s, a, b, b</italic>) are plotted after the learning is completed. Color lines show the overlaps with <italic>S</italic>, <italic>A</italic>, <italic>B</italic>, and <italic>C</italic>. The bar at the top indicates the inputs (<italic>s, a, b, b</italic>) and their applied periods. <bold>(B)</bold> The average overlaps with the patterns <italic>A</italic>&#x02032;, <italic>B</italic>, and <italic>C</italic> are plotted in cyan, green, and red, respectively, during the learning of the second sequence. Each overlap is obtained by averaging over 20 realizations of networks. <bold>(C)</bold> The neural dynamics in recall in response to (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>) are plotted by using the overlaps with <italic>S</italic>, <italic>A</italic>&#x02032;, <italic>B</italic>, and <italic>C</italic> after three epochs of learning (<italic>S, A</italic>&#x02032;, <italic>B</italic>). <bold>(D)</bold> Success rate of generation of the sub-sequence (<italic>B, C</italic>) is plotted for different input sequences (<italic>s, a, b, b</italic>), (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>), and (<italic>s, v, b, b</italic>) as a function of the learning epoch for the second sequence. The black bar indicates the region in which the success rates between (<italic>s, a</italic>&#x02032;, <italic>b, b</italic>) and (<italic>s, v, b, b</italic>) are significantly different (Wilcoxon&#x00027;s signed-rank test <italic>p</italic> &#x0003C; 0.05). Refer to Materials and methods for details.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0004.tif"/>
</fig></sec>
<sec>
<title>3.3. Learning of History-Dependent Sequences</title>
<p>We examined if the proposed model learns the history-dependent sequence (<italic>M</italic> &#x0003D; 6), in which the same patterns exist in a sequence such as <inline-formula><mml:math id="M78"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mspace width="0.3em" class="thinspace"/><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>6</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>E</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. The patterns succeeding <italic>B</italic> are <italic>C</italic> or <italic>E</italic>, depending on whether the previous pattern is <italic>A</italic> or <italic>D</italic>. Then, the neural dynamics have to retain the information of the target <italic>A</italic> or <italic>D</italic>, to recall the target <italic>C</italic> or <italic>E</italic> correctly. Our model succeeded in recalling this sequence, as shown in <xref ref-type="fig" rid="F5">Figure 5A</xref>. Just before the target <italic>C</italic> and <italic>E</italic> are recalled, there is no clear difference in the values of fast variables <bold><italic>x</italic></bold>, as indicated by the circles in <xref ref-type="fig" rid="F5">Figure 5B</xref>. However, the values of slow variables <bold><italic>y</italic></bold> are different, depending on the previous targets shown in <xref ref-type="fig" rid="F5">Figure 5C</xref>, which stabilize different patterns of <bold><italic>x</italic></bold>. Furthermore, we demonstrate that our model succeeded in recalling more complex sequences (<italic>M</italic> &#x0003D; 8) such as <inline-formula><mml:math id="M79"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mspace width="0.3em" class="thinspace"/><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>8</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi><mml:mo>,</mml:mo><mml:mi>E</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, as shown in <xref ref-type="fig" rid="F5">Figures 5D&#x02013;F</xref>. In this case, the neural dynamics have to keep three previous targets in memory to recall the target <italic>D</italic> or <italic>F</italic> after <italic>B</italic> and <italic>C</italic>. As expected, generating the sequence with <italic>M</italic> &#x0003D; 8 is a harder task than that with <italic>M</italic> &#x0003D; 6. However, some networks still can generate the sequence.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Recall processes for history-dependent sequences for <italic>K</italic> &#x0003D; 1, <italic>M</italic> &#x0003D; 6 <bold>(A&#x02013;C)</bold> and for <italic>K</italic> &#x0003D; 1, <italic>M</italic> &#x0003D; 8 <bold>(D&#x02013;F)</bold>. <bold>(A</bold> and <bold>D)</bold> The neural activities of <bold><italic>x</italic></bold> upon <bold><italic>&#x003B7;</italic></bold><sup>1</sup> are plotted by using their overlaps with the targets. Colors and alphabets indicate which targets overlapped. <bold>(B,C,E</bold> and <bold>F)</bold> The neural dynamics plotted in <bold>(A</bold> and <bold>D)</bold> are shown by projecting the fast dynamics in <bold>(B</bold> and <bold>E)</bold> and the slow dynamics in <bold>(C</bold> and <bold>F)</bold> onto a 2-dimensional PC space. X-shaped marks represent the locations of the targets. Magenta and cyan circles in B indicate the locations of <bold><italic>x</italic></bold>, respectively, just before targets <italic>C</italic> and <italic>E</italic> are recalled (as indicated by the arrows in <bold>(A)</bold>, whereas the circles in <bold>(C)</bold> indicate the locations of <bold><italic>y</italic></bold>. <bold>(G)</bold> The success rate in generating these sequences is shown for different lengths. For generating the sequence with <italic>M</italic>, the network is required to store the information about <italic>M</italic>/2&#x02212;1 preceding patterns. We measured the success rate over 50 network realizations and plotted it as a function of the number of the preceding patterns to be stored.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0005.tif"/>
</fig>
<p>To understand the performance comprehensively, we measured the success rate in generating these sequences for different lengths. For the sequence with the length of <italic>M</italic>, the network is required to retain the information about <italic>M</italic>/2&#x02212;1 preceding patterns. We examined the success rate for <italic>M</italic> &#x0003D; 6, 8, and 10 over 50 networks realizations in the same manner as that in the inference task. The success rate is reduced for the longer <italic>M</italic> and nearly zero for <italic>M</italic> &#x0003D; 10. This result indicates that our model can store three preceding patterns at a maximum, but is difficult to memorize four preceding patterns.</p>
<p>We, next, explored whether the model can memorize another type of the history-dependent sequence such as <inline-formula><mml:math id="M80"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mspace width="0.3em" class="thinspace"/><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>6</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, as shown in <xref ref-type="fig" rid="F6">Figure 6</xref>. The network is required to discriminate three neural states in the slow dynamics just before <bold><italic>x</italic></bold> approaches <italic>B</italic>, <italic>C</italic>, and <italic>D</italic>, as shown by circles in <xref ref-type="fig" rid="F6">Figure 6C</xref>. When the network discriminates these states successfully, it generates the sequence adequately, as shown in <xref ref-type="fig" rid="F6">Figures 6A&#x02013;C</xref>. We measured the success rate in generating these sequences for different numbers of the states to be discriminated (namely, <italic>M</italic>/2 states for an <italic>M</italic>-pattern sequence) in <xref ref-type="fig" rid="F6">Figure 6D</xref>. For the shortest case, the success rate takes less than 0.4.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Recall processes for history-dependent sequences such as (<italic>A, B, A, C, A, D</italic>). <bold>(A)</bold> The neural activities of <bold><italic>x</italic></bold> upon <bold><italic>&#x003B7;</italic></bold><sup>1</sup> are plotted by using their overlaps with the targets. <bold>(B,C)</bold> The neural dynamics plotted in <bold>(A)</bold> are shown by projecting the fast dynamics in <bold>(B)</bold> and the slow dynamics in <bold>(C)</bold> onto a 2-dimensional PC space. X-shaped marks represent the locations of the targets. Gray circles in <bold>(C)</bold> indicate the locations of <bold><italic>y</italic></bold> just before targets <italic>B</italic>, <italic>C</italic>, and <italic>D</italic> are recalled. <bold>(D)</bold> The success rate in generating these sequences is shown for different lengths. For generating the sequence with <italic>M</italic>, the network is required to discriminate <italic>M</italic>/2 neural states just before targets <italic>B, C, D</italic>, &#x02026; are recalled. We measured the success rate over 50 network realizations and plotted it as a function of the number of the states to be discriminated.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0006.tif"/>
</fig>
<p>As a final example of the complex sequences, we explored learning two history-dependent sequences (<xref ref-type="fig" rid="F7">Figure 7</xref>), namely, (<inline-formula><mml:math id="M81"><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:math></inline-formula>)=(<italic>A, B, C</italic>) upon <bold><italic>&#x003B7;</italic></bold><sup>1</sup>, and (<inline-formula><mml:math id="M82"><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:msubsup><mml:mrow><mml:mi>&#x003BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mstyle></mml:math></inline-formula>)=(<italic>C, B, A</italic>) upon <bold><italic>&#x003B7;</italic></bold><sup>2</sup>. In these sequences, the flow <italic>A</italic>&#x02192;<italic>B</italic>&#x02192;<italic>C</italic> on the state space under <bold><italic>&#x003B7;</italic></bold><sup>1</sup> should be reversed under <bold><italic>&#x003B7;</italic></bold><sup>2</sup>. The learned network succeeds in generating these sequences. Although orbits of <bold><italic>x</italic></bold> under inputs almost overlap in the 2-dimensional space, those of <bold><italic>y</italic></bold> does not. This difference in <bold><italic>y</italic></bold>, in addition to inputs, allows the orbits of <bold><italic>x</italic></bold> in the reverse order of patterns. Generally, <bold><italic>y</italic></bold> is different depending on the history of the previous patterns and inputs even when <bold><italic>x</italic></bold> is the same. Different <bold><italic>y</italic></bold> stabilizes different fixed point of <bold><italic>x</italic></bold>, to generate the history-dependent sequence.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Recall processes for history-dependent sequences for <italic>K</italic> &#x0003D; 2, <italic>M</italic> &#x0003D; 3. <bold>(A</bold> and <bold>B)</bold> The neural activities of <bold><italic>x</italic></bold> upon <bold><italic>&#x003B7;</italic></bold><sup>1</sup> and <bold><italic>&#x003B7;</italic></bold><sup>2</sup> are plotted by using their overlaps with the targets in <bold>(A</bold> and <bold>B)</bold>, respectively. Colors and alphabets indicate the targets overlapped. <bold>(C</bold> and <bold>D)</bold> The neural dynamics shown in <bold>(A</bold> and <bold>B)</bold> are shown by projecting them onto a 2-dimensional PC space. The fast dynamics are shown in <bold>(C)</bold> and the slow dynamics are shown in <bold>(D)</bold>. The neural trajectories upon <bold><italic>&#x003B7;</italic></bold><sup>1</sup> and <bold><italic>&#x003B7;</italic></bold><sup>2</sup> are plotted in gray and black, respectively.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0007.tif"/>
</fig>
<p>Generating these sequences is rather hard. Actually, the success rate is 0.14 for three-pattern sequences [(<italic>A, B, C</italic>) under one input, (<italic>C, B, A</italic>) under the other input] and 0.08 for four-pattern sequences. The success rate is low even for the shortest sequences. The difficulty of this task could be attributed to the request that networks have to memorize the bidirectional transitions between the target patterns (namely, the transition from <italic>A</italic> to <italic>B</italic> and its reverse transition) dependent on the external input.</p></sec>
<sec>
<title>3.4. Timescale Dependence</title>
<p>Recall performance is highly dependent on the relation between &#x003C4;<sub><italic>x</italic></sub>, &#x003C4;<sub><italic>y</italic></sub>, and &#x003C4;<sub><italic>syn</italic></sub>. To investigate the dependence of the performance on the timescales, we trained fifty realizations of networks for various values of timescales and calculated the success rate of training as a function of &#x003C4;<sub><italic>syn</italic></sub> for different &#x003C4;<sub><italic>y</italic></sub> by fixing &#x003C4;<sub><italic>x</italic></sub> at 1, as are plotted after rescaling &#x003C4;<sub><italic>syn</italic></sub> by &#x003C4;<sub><italic>y</italic></sub> in <xref ref-type="fig" rid="F8">Figure 8A</xref>. The ratios yield a common curve that shows an optimal value &#x0007E;1 at &#x003C4;<sub><italic>syn</italic></sub>, approximately equal to &#x003C4;<sub><italic>y</italic></sub>. As an exceptional case, the success rate for &#x003C4;<sub><italic>y</italic></sub> &#x0003D; 10 yields a lower value for the optimal &#x003C4;<sub><italic>syn</italic></sub> because &#x003C4;<sub><italic>y</italic></sub> is too close to &#x003C4;<sub><italic>x</italic></sub> to store the information about <bold><italic>x</italic></bold>. The balance between &#x003C4;<sub><italic>syn</italic></sub> and &#x003C4;<sub><italic>y</italic></sub> is important to regulate the success rate when they are sufficiently smaller than &#x003C4;<sub><italic>x</italic></sub>.</p>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Time scale dependence of neural dynamics. <bold>(A)</bold> The success rate of recalls as functions of &#x003C4;<sub><italic>syn</italic></sub> for given &#x003C4;<sub><italic>y</italic></sub>. The curves of the success rates are rescaled by &#x003C4;<sub><italic>y</italic></sub>. Different colors represent different &#x003C4;<sub><italic>y</italic></sub> indicated by bars below panels. The success rate is calculated across fifty realizations for <italic>K</italic> &#x0003D; 1, <italic>M</italic> &#x0003D; 7. <bold>(B</bold> and <bold>C)</bold> The neural activities of <bold><italic>x</italic></bold> (upper) and <bold><italic>y</italic></bold> (lower) in the recall process are shown by using the overlaps with the targets in the same color as shown in <xref ref-type="fig" rid="F2">Figure 2A</xref>. The neural dynamics for &#x003C4;<sub><italic>y</italic></sub> &#x0003D; 100, &#x003C4;<sub><italic>syn</italic></sub> &#x0003D; 10 are shown in B, while those for &#x003C4;<sub><italic>y</italic></sub> &#x0003D; 100, &#x003C4;<sub><italic>syn</italic></sub> &#x0003D; 1, 000 are shown in <bold>(C)</bold>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-15-743537-g0008.tif"/>
</fig>
<p>To unveil the significance of the timescale balance, we, first, present how the recall is failed for &#x003C4;<sub><italic>y</italic></sub>&#x0003E;&#x0003E;&#x003C4;<sub><italic>syn</italic></sub>, (&#x003C4;<sub><italic>y</italic></sub> &#x0003D; 100, &#x003C4;<sub><italic>syn</italic></sub> &#x0003D; 10 in <xref ref-type="fig" rid="F8">Figure 8B</xref>). Some of the targets are recalled sequentially in the wrong order, whereas other targets do not appear in the recall process. To uncover the underlying mechanism of the failed recall, we analyze the neural dynamics of fast variables with slow variables quenched in a manner similar to that shown in <xref ref-type="fig" rid="F2">Figure 2</xref> (refer to <xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S5A</xref>). Here, all the targets are stable for certain <bold><italic>y</italic></bold>, although <inline-formula><mml:math id="M83"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> does not appear in the recall process. We also found that fixed points corresponding to <inline-formula><mml:math id="M84"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M85"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> do not coexist for any <bold><italic>y</italic></bold>: the fixed point corresponding to <inline-formula><mml:math id="M86"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> has a large basin across all <bold><italic>y</italic></bold>. This leads to a transition from <inline-formula><mml:math id="M87"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> to <inline-formula><mml:math id="M88"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> by skipping <inline-formula><mml:math id="M89"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>, and thus, the recall is failed.</p>
<p>Interestingly, how a recall is failed for &#x003C4;<sub><italic>y</italic></sub> &#x0003C; &#x0003C; &#x003C4;<sub><italic>syn</italic></sub> are distinct from that for &#x003C4;<sub><italic>y</italic></sub>&#x0003E;&#x0003E;&#x003C4;<sub><italic>syn</italic></sub>. For &#x003C4;<sub><italic>y</italic></sub> &#x0003D; 100, &#x003C4;<sub><italic>syn</italic></sub> &#x0003D; 1, 000, only the most recently learned target is stable for almost all <bold><italic>y</italic></bold>, and thus, only this target is recalled, as shown in <xref ref-type="fig" rid="F8">Figure 8C</xref>. We sampled the slow variables from the last learning step of the sequence (<xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S5B</xref>), and analyzed the bifurcation of the fast variables against change in slow variables, in the same way as above. In this study, only the latest target (here, <inline-formula><mml:math id="M90"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>) is a fixed point, whereas the other targets are not. Thus, transitions between targets are missed, except the transition to the latest target.</p>
<p>Why does the network fail in generating the sequences when &#x003C4;<sub><italic>y</italic></sub> and &#x003C4;<sub><italic>syn</italic></sub> are not matched? The reason is as follows: the learning time for the non-optimal timescales takes a longer time than that for the optimal ones. In this study, we consider the time that is normalized by &#x003C4;<sub><italic>syn</italic></sub> because it characterizes the timescale regulating the synaptic plasticity and, consequently, the neural dynamics. For &#x003C4;<sub><italic>y</italic></sub>&#x0003E;&#x0003E;&#x003C4;<sub><italic>syn</italic></sub> (<xref ref-type="supplementary-material" rid="SM1">Supplementary Figure S5C</xref>), the trajectories of neural activities are similar to those for &#x003C4;<sub><italic>y</italic></sub>&#x0007E;&#x003C4;<sub><italic>syn</italic></sub>, but it takes a longer time for <bold><italic>y</italic></bold> to approach <bold><italic>x</italic></bold> after <bold><italic>x</italic></bold> converges to the target. As the approach to <bold><italic>x</italic></bold> is so slow, the learning process stabilizes the present target during the approach by modifying the connectivity, so that the target is too stable over a wide range of <bold><italic>y</italic></bold>. On the other hand, for &#x003C4;<sub><italic>y</italic></sub> &#x0003C; &#x0003C; &#x003C4;<sub><italic>syn</italic></sub> (<xref ref-type="supplementary-material" rid="SM1">Supplementary Figure</xref> <xref ref-type="supplementary-material" rid="SM1">S5E</xref>), the neural activities of <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold> wander before <bold><italic>x</italic></bold> converges to the target, resulting in disruption of information about the previous targets stored in <bold><italic>y</italic></bold>. Thus, the networks for both cases of &#x003C4;<sub><italic>y</italic></sub>&#x0003E;&#x0003E;&#x003C4;<sub><italic>syn</italic></sub> and &#x003C4;<sub><italic>y</italic></sub> &#x0003C; &#x0003C; &#x003C4;<sub><italic>syn</italic></sub> fail to generate the sequence. These results indicate that the relative timescale &#x003C4;<sub><italic>y</italic></sub> to &#x003C4;<sub><italic>syn</italic></sub> changes the bifurcation of the fast dynamics and the memory capacity.</p></sec></sec>	
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>Sequential transitions between metastable patterns are ubiquitously observed in the neural system (Miller, <xref ref-type="bibr" rid="B41">2016</xref>) during various tasks, such as perception (Jones et al., <xref ref-type="bibr" rid="B24">2007</xref>; Miller and Katz, <xref ref-type="bibr" rid="B42">2010</xref>), decision making (Ponce-Alvarez et al., <xref ref-type="bibr" rid="B46">2012</xref>), working memory (Stokes et al., <xref ref-type="bibr" rid="B56">2013</xref>; Taghia et al., <xref ref-type="bibr" rid="B58">2018</xref>),and recall of long-term memory (Wimmer et al., <xref ref-type="bibr" rid="B65">2020</xref>). We have developed a novel neural network model with fast and slow dynamics to generate sequences with non-Markov property and concatenate sequences, which are based on these cognitive functions.</p>
<p>In a standard model for generating sequential patterns (Kleinfeld, <xref ref-type="bibr" rid="B27">1986</xref>; Sompolinsky and Kanter, <xref ref-type="bibr" rid="B55">1986</xref>; Nishimori et al., <xref ref-type="bibr" rid="B44">1990</xref>; Russo and Treves, <xref ref-type="bibr" rid="B51">2012</xref>; Recanatesi et al., <xref ref-type="bibr" rid="B48">2015</xref>; Haga and Fukai, <xref ref-type="bibr" rid="B17">2019</xref>), asymmetric Hebbian learning between a pattern &#x003BC; and the next &#x003BC;&#x0002B;1, i.e., <bold>&#x003BE;</bold><italic><sup>&#x003BC;&#x0002B;1</sup></italic>(<bold>&#x003BE;</bold><italic><sup>&#x003BC;</sup></italic>)<italic>t</italic>, is used to create the transition from <bold>&#x003BE;</bold><italic><sup>&#x003BC;</sup></italic> to <bold>&#x003BE;</bold><sup>&#x003BC;&#x0002B;1</sup> (Kleinfeld, <xref ref-type="bibr" rid="B27">1986</xref>; Sompolinsky and Kanter, <xref ref-type="bibr" rid="B55">1986</xref>; Russo and Treves, <xref ref-type="bibr" rid="B51">2012</xref>; Recanatesi et al., <xref ref-type="bibr" rid="B48">2015</xref>; Haga and Fukai, <xref ref-type="bibr" rid="B17">2019</xref>). In these studies, however, only the connections between the current and immediately preceding patterns are embedded in the connectivity, resulting in that the prolonged history of the patterns cannot be embedded. In other studies, Rabinovich&#x00027;s group (Seliger et al., <xref ref-type="bibr" rid="B53">2003</xref>; Rabinovich and Varona, <xref ref-type="bibr" rid="B47">2018</xref>) proposed the model generating sequential activities by heteroclinic orbits between patterns. As the above standard model, asymmetric Hebbian learning forms the connectivity for generating the sequences (Seliger et al., <xref ref-type="bibr" rid="B53">2003</xref>). The information about the history of patterns is not stored in the model. Thus, non-Markov sequences are not generated in contrast to our model<xref ref-type="fn" rid="fn0003"><sup>3</sup></xref>.</p>
<p>In some models, a term that changes slower than the neural dynamics (e.g., an adaptation term) is introduced to lead to the transition. In Gros (<xref ref-type="bibr" rid="B13">2007</xref>); Russo and Treves (<xref ref-type="bibr" rid="B51">2012</xref>), and Recanatesi et al. (<xref ref-type="bibr" rid="B48">2015</xref>), the slow term is introduced to destabilize the current pattern. These methods imply non-Markov dynamics because the slow term needs prolonged times to recover, leading to change in the transition probabilities among the patterns. However, this term does not determine the next pattern and, thus, some additional mechanism is necessary for the transition to the desired pattern. The feedback from the slow population in our model, in contrast, not only destabilizes the current pattern but also simultaneously stabilizes the next targeted pattern. As the current and next patterns coexist for some time, the robust transition between them is achieved.</p>
<p>Alternatively, supervised learning methods used in machine learning fields, such as Back-Propagation Through Time (BPTT) (Werbos, <xref ref-type="bibr" rid="B62">1990</xref>), are investigated to reproduce sequential neural activities observed experimentally (Mante et al., <xref ref-type="bibr" rid="B38">2013</xref>; Carnevale et al., <xref ref-type="bibr" rid="B7">2015</xref>; Chaisangmongkon et al., <xref ref-type="bibr" rid="B8">2017</xref>), including non-Markov trajectories (Sussillo and Abbott, <xref ref-type="bibr" rid="B57">2009</xref>; Laje and Buonomano, <xref ref-type="bibr" rid="B34">2013</xref>). The BPTT, however, requires non-local information and the network has to retain a large amount of information until the trajectory terminates, which is biologically implausible. Furthermore, the trajectories shaped by this method are vulnerable to noise (Laje and Buonomano, <xref ref-type="bibr" rid="B34">2013</xref>). Our model is free from these deficiencies.</p>
<p>In our model, the recurrent connections in the fast population (i.e., connections within a cortical area) are modified to shape the transitions between memorized states whereas the connections between the fast and slow populations and those from the input to the fast population are fixed (i.e., connections across cortical areas). In another approach, Gros and Kaczor (<xref ref-type="bibr" rid="B14">2010</xref>) demonstrated that the plasticity in the afferent connections with the fixed recurrent connections is useful for semantic learning, by connecting appropriately external stimuli with already established neural patterns in the recurrent network. In the neural system, generally, both connections across and within cortical areas are plastic. The existence of both dual plasticity possibly leads to interference between these connections, potentially resulting in reducing the learning performance. Future studies are needed to clarify how such dual plasticity cooperatively builds neural activities to perform cognitive functions.</p>
<p>Timescales in the neural activities are hierarchically distributed across several cortical areas (Honey et al., <xref ref-type="bibr" rid="B19">2012</xref>; Murray et al., <xref ref-type="bibr" rid="B43">2014</xref>; Hasson et al., <xref ref-type="bibr" rid="B18">2015</xref>; Runyan et al., <xref ref-type="bibr" rid="B50">2017</xref>). For instance, consider the hippocampus (HPC) and the prefrontal cortex (PFC), which are coupled by mono-synaptic and di-synaptic connections (Ito et al., <xref ref-type="bibr" rid="B21">2015</xref>). HPC neurons respond to the location of animals (Kumaran et al., <xref ref-type="bibr" rid="B28">2016</xref>) with faster timescales than those in PFC, which has the slowest timescale among cortical areas (Murray et al., <xref ref-type="bibr" rid="B43">2014</xref>). Experimental studies (Ito et al., <xref ref-type="bibr" rid="B21">2015</xref>; Guise and Shapiro, <xref ref-type="bibr" rid="B15">2017</xref>) revealed that PFC neurons are necessary to differentiate HPC dynamics depending on the context and previous experience. Similarly, neurons in the orbitofrontal cortex (OFC), whose timescales are considered to be slower than those in HPC, are necessary for concatenating the sequences in the stimulus-reward response (Jones et al., <xref ref-type="bibr" rid="B23">2012</xref>; Wikenheiser and Schoenbaum, <xref ref-type="bibr" rid="B64">2016</xref>). Accordingly, it is suggested that the area with the slow dynamics is necessary to generate and concatenate the sequences.</p>
<p>Neural networks with multiple timescales are investigated theoretically in several studies. In some studies (Yamashita and Tani, <xref ref-type="bibr" rid="B66">2008</xref>; Perdikis et al., <xref ref-type="bibr" rid="B45">2011</xref>), the slow dynamics are introduced to concatenate primitive movements and produce a complex movement, while hidden states of the hierarchical external stimuli are inferred by the multiple timescales in the neural dynamics in another study (Kiebel et al., <xref ref-type="bibr" rid="B26">2009</xref>). In Kiebel et al. (<xref ref-type="bibr" rid="B26">2009</xref>) and Perdikis et al. (<xref ref-type="bibr" rid="B45">2011</xref>), the relationship between the slow and fast dynamics are fixed a priori to perform their tasks, whereas, in our model, such a relationship is shaped through the learning process. In Yamashita and Tani (<xref ref-type="bibr" rid="B66">2008</xref>), the BPTT method is adopted for training the network; thus, it faced the same drawbacks as already mentioned. In the studies of the multiple timescales system, analytical methods such that singular perturbation methods are adopted, which are commonly used to elucidate the transition between the states on the different slow manifolds (Ermentrout, <xref ref-type="bibr" rid="B11">1998</xref>; Rubin et al., <xref ref-type="bibr" rid="B49">2013</xref>; Bertram and Rubin, <xref ref-type="bibr" rid="B5">2017</xref>; Wernecke et al., <xref ref-type="bibr" rid="B63">2018</xref>) and the stability of fixed points (Meyer-B&#x000E4;se et al., <xref ref-type="bibr" rid="B40">1996</xref>; Hongtao and Amari, <xref ref-type="bibr" rid="B20">2006</xref>). Our model provides how these transitions are formed through learning and it generates and concatenates the history-dependent sequences, while the application of these methods will be useful in the future.</p>
<p>As for the timescales, we need further studies to fill a gap between our model and experimental observations. The ratio of the timescale in the slow dynamics to that in the fast dynamics is less than 10 times across cortical areas (Wang and Kennedy, <xref ref-type="bibr" rid="B60">2016</xref>), which are smaller than the optimal ratio in our model. Further, the difference between the timescales in the slow dynamics (on the order of a second) and in the synaptic plasticity (on the order of a minute Bliss and Lomo, <xref ref-type="bibr" rid="B6">1973</xref>; Bayazitov et al., <xref ref-type="bibr" rid="B3">2007</xref>) is larger than that adopted in our model.</p>
<p>Diversity in the timescales of individual neurons and the calcium dynamics possibly resolve this discrepancy. The timescale of individual neurons in the same area is distributed over two digits (Bernacchia et al., <xref ref-type="bibr" rid="B4">2011</xref>; Wasmuht et al., <xref ref-type="bibr" rid="B61">2018</xref>). The calcium dynamics in the synapses can modify the synaptic efficacy on the order of a second (Shouval et al., <xref ref-type="bibr" rid="B54">2010</xref>; Graupner and Brunel, <xref ref-type="bibr" rid="B12">2012</xref>). By taking these effects into account, our model may be consistent with the experimental observations, although further studies will be important, including those with spiking neurons (Kurikawa and Kaneko, <xref ref-type="bibr" rid="B32">2015</xref>) and spike-timing-dependent potentiation.</p>
<p>Finally, we discuss the biological plausibility of the learning rule in our model. The fast network receives two inputs; an external input (<bold>&#x003B7;</bold>) and the input from the slow network. In the neural system, the external input is conveyed through afferent connections from a lower cortical area (or sensory input) and the feedback input comes from a higher cortical area. In addition to these inputs, another input for the learning is introduced in our model, which provides information to generate sequential patterns to be learned. Thus, our network is trained to map between sensory cues and sequential patterns in the output area by using a Hebbian rule (correlation between &#x003BE;<sub><italic>j</italic></sub> and <italic>x</italic><sub><italic>i</italic></sub>) and an anti-Hebbian rule (correlation between <italic>x</italic><sub><italic>i</italic></sub> and <italic>x</italic><sub><italic>j</italic></sub>). After training, the network evokes the sequential patterns under the sensory input.</p></sec>
<sec sec-type="data-availability" id="s5">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="SM1">Supplementary Material</xref>, further inquiries can be directed to the corresponding author.</p></sec>
<sec id="s6">
<title>Author Contributions</title>
<p>TK and KK: conceptualization and writing. TK: formal analysis. Both authors contributed to the article and approved the submitted version.</p></sec>
<sec sec-type="funding-information" id="s7">
<title>Funding</title>
<p>This study was partly supported by JSPS KAKENHI (Nos. 18K15343 and 20H00123).</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
<sec sec-type="disclaimer" id="s8">
<title>Publisher&#x00027;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec> </body>
<back>
<ack><p>We thank Omri Barak and Rei Akaishi for fruitful discussion and Tatsuya Haga for useful comments on our manuscript.</p>
</ack>
<sec sec-type="supplementary-material" id="s9">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fncom.2021.743537/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fncom.2021.743537/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/></sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Akhlaghpour</surname> <given-names>H.</given-names></name> <name><surname>Wiskerke</surname> <given-names>J.</given-names></name> <name><surname>Choi</surname> <given-names>J. Y.</given-names></name> <name><surname>Taliaferro</surname> <given-names>J. P.</given-names></name> <name><surname>Au</surname> <given-names>J.</given-names></name> <name><surname>Witten</surname> <given-names>I. B.</given-names></name></person-group> (<year>2016</year>). <article-title>Dissociated sequential activity and stimulus encoding in the dorsomedial striatum during spatial working memory</article-title>. <source>Elife</source> <volume>5</volume>, <fpage>1</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.7554/eLife.19507</pub-id><pub-id pub-id-type="pmid">27636864</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Amari</surname> <given-names>S.-I..</given-names></name></person-group> (<year>1972</year>). <article-title>Learning patterns and pattern sequences by self-organizing nets of threshold elements</article-title>. <source>IEEE Trans. Comput</source>. <volume>100</volume>, <fpage>1197</fpage>&#x02013;<lpage>1206</lpage>. <pub-id pub-id-type="doi">10.1109/T-C.1972.223477</pub-id><pub-id pub-id-type="pmid">27295638</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bayazitov</surname> <given-names>I. T.</given-names></name> <name><surname>Richardson</surname> <given-names>R. J.</given-names></name> <name><surname>Fricke</surname> <given-names>R. G.</given-names></name> <name><surname>Zakharenko</surname> <given-names>S. S.</given-names></name></person-group> (<year>2007</year>). <article-title>Slow presynaptic and fast postsynaptic components of compound long-term potentiation</article-title>. <source>J. Neurosci</source>. <volume>27</volume>, <fpage>11510</fpage>&#x02013;<lpage>11521</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.3077-07.2007</pub-id><pub-id pub-id-type="pmid">17959794</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bernacchia</surname> <given-names>A.</given-names></name> <name><surname>Seo</surname> <given-names>H.</given-names></name> <name><surname>Lee</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>X.-J. J.</given-names></name></person-group> (<year>2011</year>). <article-title>A reservoir of time constants for memory traces in cortical neurons</article-title>. <source>Nat. Neurosci</source>. <volume>14</volume>, <fpage>366</fpage>&#x02013;<lpage>372</lpage>. <pub-id pub-id-type="doi">10.1038/nn.2752</pub-id><pub-id pub-id-type="pmid">21317906</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bertram</surname> <given-names>R.</given-names></name> <name><surname>Rubin</surname> <given-names>J. E.</given-names></name></person-group> (<year>2017</year>). <article-title>Multi-timescale systems and fast-slow analysis</article-title>. <source>Math. Biosci</source>. <volume>287</volume>, <fpage>105</fpage>&#x02013;<lpage>121</lpage>. <pub-id pub-id-type="doi">10.1016/j.mbs.2016.07.003</pub-id><pub-id pub-id-type="pmid">27424950</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bliss</surname> <given-names>T. V.</given-names></name> <name><surname>Lomo</surname> <given-names>T.</given-names></name></person-group> (<year>1973</year>). <article-title>Long-lasting potentiation of synaptic transmission in the dentate area of the anaesthetized rabbit following stimulation of the perforant path</article-title>. <source>J. Physiol</source>. <volume>232</volume>, <fpage>331</fpage>&#x02013;<lpage>356</lpage>. <pub-id pub-id-type="doi">10.1113/jphysiol.1973.sp010273</pub-id><pub-id pub-id-type="pmid">4727084</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Carnevale</surname> <given-names>F.</given-names></name> <name><surname>DeLafuente</surname> <given-names>V.</given-names></name> <name><surname>Romo</surname> <given-names>R.</given-names></name> <name><surname>Barak</surname> <given-names>O.</given-names></name> <name><surname>Parga</surname> <given-names>N.</given-names></name> <name><surname>de Lafuente</surname> <given-names>V.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Dynamic control of response criterion in premotor cortex during perceptual detection under temporal uncertainty</article-title>. <source>Neuron</source> <volume>86</volume>, <fpage>1067</fpage>&#x02013;<lpage>1077</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2015.04.014</pub-id><pub-id pub-id-type="pmid">25959731</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chaisangmongkon</surname> <given-names>W.</given-names></name> <name><surname>Swaminathan</surname> <given-names>S. K.</given-names></name> <name><surname>Freedman</surname> <given-names>D. J.</given-names></name> <name><surname>Wang</surname> <given-names>X.-J. J.</given-names></name></person-group> (<year>2017</year>). <article-title>Computing by robust transience: how the fronto-parietal network performs sequential, category-based decisions</article-title>. <source>Neuron</source> <volume>93</volume>, <fpage>1504.e4</fpage>&#x02013;<lpage>1517.e4</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2017.03.002</pub-id><pub-id pub-id-type="pmid">28334612</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chartier</surname> <given-names>S.</given-names></name> <name><surname>Boukadoum</surname> <given-names>M.</given-names></name></person-group> (<year>2006</year>). <article-title>A sequential dynamic heteroassociative memory for multistep pattern recognition and one-to-many association</article-title>. <source>IEEE Trans. Neural Netw</source>. <volume>17</volume>, <fpage>59</fpage>&#x02013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2005.860855</pub-id><pub-id pub-id-type="pmid">16526476</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chaudhuri</surname> <given-names>R.</given-names></name> <name><surname>Knoblauch</surname> <given-names>K.</given-names></name> <name><surname>Gariel</surname> <given-names>M. A.</given-names></name> <name><surname>Kennedy</surname> <given-names>H.</given-names></name> <name><surname>Wang</surname> <given-names>X. J.</given-names></name></person-group> (<year>2015</year>). <article-title>A large-scale circuit mechanism for hierarchical dynamical processing in the primate cortex</article-title>. <source>Neuron</source> <volume>88</volume>, <fpage>419</fpage>&#x02013;<lpage>431</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2015.09.008</pub-id><pub-id pub-id-type="pmid">26439530</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ermentrout</surname> <given-names>B..</given-names></name></person-group> (<year>1998</year>). <article-title>Neural networks as spatio-temporal pattern-forming systems</article-title>. <source>Rep. Progr. Phys</source>. <volume>61</volume>, <fpage>353</fpage>&#x02013;<lpage>430</lpage>. <pub-id pub-id-type="doi">10.1088/0034-4885/61/4/002</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Graupner</surname> <given-names>M.</given-names></name> <name><surname>Brunel</surname> <given-names>N.</given-names></name></person-group> (<year>2012</year>). <article-title>Calcium-based plasticity model explains sensitivity of synaptic changes to spike pattern, rate, and dendritic location</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A</source>. <volume>109</volume>, <fpage>21551</fpage>&#x02013;<lpage>21551</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1109359109</pub-id><pub-id pub-id-type="pmid">22357758</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gros</surname> <given-names>C..</given-names></name></person-group> (<year>2007</year>). <article-title>Neural networks with transient state dynamics</article-title>. <source>New J. Phys</source>. <volume>9</volume>:<fpage>109</fpage>. <pub-id pub-id-type="doi">10.1088/1367-2630/9/4/109</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gros</surname> <given-names>C.</given-names></name> <name><surname>Kaczor</surname> <given-names>G.</given-names></name></person-group> (<year>2010</year>). <article-title>Semantic learning in autonomously active recurrent neural networks</article-title>. <source>Logic J. IGPL</source> <volume>18</volume>, <fpage>686</fpage>&#x02013;<lpage>704</lpage>. <pub-id pub-id-type="doi">10.1093/jigpal/jzp045</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guise</surname> <given-names>K. G.</given-names></name> <name><surname>Shapiro</surname> <given-names>M. L.</given-names></name></person-group> (<year>2017</year>). <article-title>Medial prefrontal cortex reduces memory interference by modifying hippocampal encoding</article-title>. <source>Neuron</source>. <volume>94</volume>, <fpage>183.e8</fpage>&#x02013;<lpage>192.e8</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2017.03.011</pub-id><pub-id pub-id-type="pmid">28343868</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>A. S.</given-names></name> <name><surname>van der Meer</surname> <given-names>M. A.</given-names></name> <name><surname>Touretzky</surname> <given-names>D. S.</given-names></name> <name><surname>Redish</surname> <given-names>A. D.</given-names></name></person-group> (<year>2010</year>). <article-title>hippocampal replay is not a simple function of experience</article-title>. <source>Neuron</source> <volume>65</volume>, <fpage>695</fpage>&#x02013;<lpage>705</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2010.01.034</pub-id><pub-id pub-id-type="pmid">20223204</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haga</surname> <given-names>T.</given-names></name> <name><surname>Fukai</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Extended temporal association memory by modulations of inhibitory circuits</article-title>. <source>Phys. Rev. Lett</source>. <volume>123</volume>, <fpage>78101</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.123.078101</pub-id><pub-id pub-id-type="pmid">31491118</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hasson</surname> <given-names>U.</given-names></name> <name><surname>Chen</surname> <given-names>J.</given-names></name> <name><surname>Honey</surname> <given-names>C. J.</given-names></name></person-group> (<year>2015</year>). <article-title>Hierarchical process memory: memory as an integral component of information processing</article-title>. <source>Trends Cogn. Sci</source>. <volume>19</volume>, <fpage>304</fpage>&#x02013;<lpage>313</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2015.04.006</pub-id><pub-id pub-id-type="pmid">25980649</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Honey</surname> <given-names>C. J.</given-names></name> <name><surname>Thesen</surname> <given-names>T.</given-names></name> <name><surname>Donner</surname> <given-names>T. H.</given-names></name> <name><surname>Silbert</surname> <given-names>L. J.</given-names></name> <name><surname>Carlson</surname> <given-names>C. E.</given-names></name> <name><surname>Devinsky</surname> <given-names>O.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Slow cortical dynamics and the accumulation of information over long timescales</article-title>. <source>Neuron</source> <volume>76</volume>, <fpage>423</fpage>&#x02013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2012.08.011</pub-id><pub-id pub-id-type="pmid">23083743</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hongtao</surname> <given-names>L.</given-names></name> <name><surname>Amari</surname> <given-names>S. I.</given-names></name></person-group> (<year>2006</year>). <article-title>Global exponential stability of multitime scale competitive neural networks with nonsmooth functions</article-title>. <source>IEEE Trans. Neural Netw</source>. <volume>17</volume>, <fpage>1152</fpage>&#x02013;<lpage>1164</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2006.875995</pub-id><pub-id pub-id-type="pmid">17001977</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ito</surname> <given-names>H. T.</given-names></name> <name><surname>Zhang</surname> <given-names>S.-,j.</given-names></name> <name><surname>Witter</surname> <given-names>M. P.</given-names></name> <name><surname>Moser</surname> <given-names>E. I.</given-names></name> <name><surname>Moser</surname> <given-names>M.-B.</given-names></name></person-group> (<year>2015</year>). <article-title>A prefrontal-thalamo-hippocampal circuit for goal-directed spatial navigation</article-title>. <source>Nature</source> <volume>522</volume>, <fpage>50</fpage>&#x02013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1038/nature14396</pub-id><pub-id pub-id-type="pmid">26017312</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>X.</given-names></name> <name><surname>Tecuapetla</surname> <given-names>F.</given-names></name> <name><surname>Costa</surname> <given-names>R. M.</given-names></name></person-group> (<year>2014</year>). <article-title>Basal ganglia subcircuits distinctively encode the parsing and concatenation of action sequences</article-title>. <source>Nat. Neurosci</source>. <volume>17</volume>, <fpage>423</fpage>&#x02013;<lpage>430</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3632</pub-id><pub-id pub-id-type="pmid">24464039</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jones</surname> <given-names>J. L.</given-names></name> <name><surname>Esber</surname> <given-names>G. R.</given-names></name> <name><surname>McDannald</surname> <given-names>M. A.</given-names></name> <name><surname>Gruber</surname> <given-names>A. J.</given-names></name> <name><surname>Hernandez</surname> <given-names>A.</given-names></name> <name><surname>Mirenzi</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Orbitofrontal cortex supports behavior and learning using inferred but not cached values</article-title>. <source>Science</source> <volume>338</volume>, <fpage>953</fpage>&#x02013;<lpage>956</lpage>. <pub-id pub-id-type="doi">10.1126/science.1227489</pub-id><pub-id pub-id-type="pmid">23162000</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jones</surname> <given-names>L. M.</given-names></name> <name><surname>Fontanini</surname> <given-names>A.</given-names></name> <name><surname>Sadacca</surname> <given-names>B. F.</given-names></name> <name><surname>Miller</surname> <given-names>P.</given-names></name> <name><surname>Katz</surname> <given-names>D. B.</given-names></name></person-group> (<year>2007</year>). <article-title>Natural stimuli evoke dynamic sequences of states in sensory cortical ensembles</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A</source>. <volume>104</volume>, <fpage>18772</fpage>&#x02013;<lpage>18777</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.0705546104</pub-id><pub-id pub-id-type="pmid">18000059</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kiebel</surname> <given-names>S. J.</given-names></name> <name><surname>Daunizeau</surname> <given-names>J.</given-names></name> <name><surname>Friston</surname> <given-names>K. J.</given-names></name></person-group> (<year>2008</year>). <article-title>A hierarchy of time-scales and the brain</article-title>. <source>PLoS Comput. Biol</source>. <volume>4</volume>:<fpage>e1000209</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1000209</pub-id><pub-id pub-id-type="pmid">19008936</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kiebel</surname> <given-names>S. J.</given-names></name> <name><surname>von Kriegstein</surname> <given-names>K.</given-names></name> <name><surname>Daunizeau</surname> <given-names>J.</given-names></name> <name><surname>Friston</surname> <given-names>K. J.</given-names></name></person-group> (<year>2009</year>). <article-title>Recognizing sequences of sequences</article-title>. <source>PLoS Comput. Biol</source>. <volume>5</volume>:<fpage>e1000464</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1000464</pub-id><pub-id pub-id-type="pmid">19680429</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kleinfeld</surname> <given-names>D..</given-names></name></person-group> (<year>1986</year>). <article-title>Sequential state generation by model neural networks</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A</source>. <volume>83</volume>, <fpage>9469</fpage>&#x02013;<lpage>9473</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.83.24.9469</pub-id><pub-id pub-id-type="pmid">3467316</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kumaran</surname> <given-names>D.</given-names></name> <name><surname>Hassabis</surname> <given-names>D.</given-names></name> <name><surname>McClelland</surname> <given-names>J. L.</given-names></name></person-group> (<year>2016</year>). <article-title>What learning systems do intelligent agents need? complementary learning systems theory updated</article-title>. <source>Trends Cogn. Sci</source> <volume>20</volume>, <fpage>512</fpage>&#x02013;<lpage>534</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2016.05.004</pub-id><pub-id pub-id-type="pmid">27315762</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kurikawa</surname> <given-names>T.</given-names></name> <name><surname>Barak</surname> <given-names>O.</given-names></name> <name><surname>Kaneko</surname> <given-names>K.</given-names></name></person-group> (<year>2020</year>). <article-title>Repeated sequential learning increases memory capacity via effective decorrelation in a recurrent neural network</article-title>. <source>Phys. Rev. Res</source>. <volume>2</volume>, <fpage>023307</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevResearch.2.023307</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kurikawa</surname> <given-names>T.</given-names></name> <name><surname>Haga</surname> <given-names>T.</given-names></name> <name><surname>Handa</surname> <given-names>T.</given-names></name> <name><surname>Harukuni</surname> <given-names>R.</given-names></name> <name><surname>Fukai</surname> <given-names>T.</given-names></name></person-group> (<year>2018</year>). <article-title>Neuronal stability in medial frontal cortex sets individual variability in decision-making</article-title>. <source>Nat. Neurosci</source>. <volume>21</volume>, <fpage>1764</fpage>&#x02013;<lpage>1773</lpage>. <pub-id pub-id-type="doi">10.1038/s41593-018-0263-5</pub-id><pub-id pub-id-type="pmid">30420732</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kurikawa</surname> <given-names>T.</given-names></name> <name><surname>Kaneko</surname> <given-names>K.</given-names></name></person-group> (<year>2013</year>). <article-title>Embedding responses in spontaneous neural activity shaped through sequential learning</article-title>. <source>PLoS Comput. Biol</source>. <volume>9</volume>:<fpage>e1002943</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002943</pub-id><pub-id pub-id-type="pmid">23505355</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kurikawa</surname> <given-names>T.</given-names></name> <name><surname>Kaneko</surname> <given-names>K.</given-names></name></person-group> (<year>2015</year>). <article-title>Memories as bifurcations: Realization by collective dynamics of spiking neurons under stochastic inputs</article-title>. <source>Neural Netw</source>. <volume>62</volume>, <fpage>25</fpage>&#x02013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2014.07.005</pub-id><pub-id pub-id-type="pmid">25124069</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kurikawa</surname> <given-names>T.</given-names></name> <name><surname>Kaneko</surname> <given-names>K.</given-names></name></person-group> (<year>2016</year>). <article-title>Dynamic organization of hierarchical memories</article-title>. <source>PLoS ONE</source> <volume>11</volume>:<fpage>e0162640</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0162640</pub-id><pub-id pub-id-type="pmid">27618549</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Laje</surname> <given-names>R.</given-names></name> <name><surname>Buonomano</surname> <given-names>D. V.</given-names></name></person-group> (<year>2013</year>). <article-title>Robust timing and motor patterns by taming chaos in recurrent neural networks</article-title>. <source>Nat. Neurosci</source>. <volume>16</volume>, <fpage>925</fpage>&#x02013;<lpage>933</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3405</pub-id><pub-id pub-id-type="pmid">23708144</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Larkum</surname> <given-names>M..</given-names></name></person-group> (<year>2013</year>). <article-title>A cellular mechanism for cortical associations: an organizing principle for the cerebral cortex</article-title>. <source>Trends Neurosci</source>. <volume>36</volume>, <fpage>141</fpage>&#x02013;<lpage>151</lpage>. <pub-id pub-id-type="doi">10.1016/j.tins.2012.11.006</pub-id><pub-id pub-id-type="pmid">23273272</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Larkum</surname> <given-names>M. E.</given-names></name> <name><surname>Nevian</surname> <given-names>T.</given-names></name> <name><surname>Sandler</surname> <given-names>M.</given-names></name> <name><surname>Polsky</surname> <given-names>A.</given-names></name> <name><surname>Schiller</surname> <given-names>J.</given-names></name></person-group> (<year>2009</year>). <article-title>Synaptic integration in tuft dendrites of layer 5 pyramidal neurons: a new unifying principle</article-title>. <source>Science</source> <volume>325</volume>, <fpage>756</fpage>&#x02013;<lpage>760</lpage>. <pub-id pub-id-type="doi">10.1126/science.1171958</pub-id><pub-id pub-id-type="pmid">19661433</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Maboudi</surname> <given-names>K.</given-names></name> <name><surname>Ackermann</surname> <given-names>E.</given-names></name> <name><surname>de Jong</surname> <given-names>L. W.</given-names></name> <name><surname>Pfeiffer</surname> <given-names>B. E.</given-names></name> <name><surname>Foster</surname> <given-names>D.</given-names></name> <name><surname>Diba</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Uncovering temporal structure in hippocampal output patterns</article-title>. <source>Elife</source> <volume>7</volume>, <fpage>1</fpage>&#x02013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.7554/eLife.34467</pub-id><pub-id pub-id-type="pmid">29869611</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mante</surname> <given-names>V.</given-names></name> <name><surname>Sussillo</surname> <given-names>D.</given-names></name> <name><surname>Shenoy</surname> <given-names>K. V.</given-names></name> <name><surname>Newsome</surname> <given-names>W. T.</given-names></name></person-group> (<year>2013</year>). <article-title>Context-dependent computation by recurrent dynamics in prefrontal cortex</article-title>. <source>Nature</source> <volume>503</volume>, <fpage>78</fpage>&#x02013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1038/nature12742</pub-id><pub-id pub-id-type="pmid">24201281</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mazzucato</surname> <given-names>L.</given-names></name> <name><surname>Fontanini</surname> <given-names>A.</given-names></name> <name><surname>La Camera</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Dynamics of multistable states during ongoing and evoked cortical activity</article-title>. <source>J. Neurosci</source>. <volume>35</volume>, <fpage>8214</fpage>&#x02013;<lpage>8231</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.4819-14.2015</pub-id><pub-id pub-id-type="pmid">26019337</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Meyer-B&#x000E4;se</surname> <given-names>A.</given-names></name> <name><surname>Ohl</surname> <given-names>F.</given-names></name> <name><surname>Scheich</surname> <given-names>H.</given-names></name></person-group> (<year>1996</year>). <article-title>Singular perturbation analysis of competitive neural networks with different time scales</article-title>. <source>Neural Comput</source>. <volume>8</volume>, <fpage>1731</fpage>&#x02013;<lpage>1742</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1996.8.8.1731</pub-id><pub-id pub-id-type="pmid">8888615</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>P..</given-names></name></person-group> (<year>2016</year>). <article-title>Itinerancy between attractor states in neural systems</article-title>. <source>Curr. Opin Neurobiol</source>. <volume>40</volume>, <fpage>14</fpage>&#x02013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2016.05.005</pub-id><pub-id pub-id-type="pmid">27318972</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>P.</given-names></name> <name><surname>Katz</surname> <given-names>D. B.</given-names></name></person-group> (<year>2010</year>). <article-title>Stochastic transitions between neural states in taste processing and decision-making</article-title>. <source>J. Neurosci</source>. <volume>30</volume>, <fpage>2559</fpage>&#x02013;<lpage>2570</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.3047-09.2010</pub-id><pub-id pub-id-type="pmid">20164341</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murray</surname> <given-names>J. D.</given-names></name> <name><surname>Bernacchia</surname> <given-names>A.</given-names></name> <name><surname>Freedman</surname> <given-names>D. J.</given-names></name> <name><surname>Romo</surname> <given-names>R.</given-names></name> <name><surname>Wallis</surname> <given-names>J. D.</given-names></name> <name><surname>Cai</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>A hierarchy of intrinsic timescales across primate cortex</article-title>. <source>Nat. Neurosci</source>. <volume>17</volume>, <fpage>1661</fpage>&#x02013;<lpage>1663</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3862</pub-id><pub-id pub-id-type="pmid">25383900</pub-id></citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nishimori</surname> <given-names>H.</given-names></name> <name><surname>Nakamura</surname> <given-names>T.</given-names></name> <name><surname>Shiino</surname> <given-names>M.</given-names></name></person-group> (<year>1990</year>). <article-title>Retrieval of spatio-temporal sequence in asynchronous neural network</article-title>. <source>Phys. Rev. A</source> <volume>41</volume>, <fpage>3346</fpage>&#x02013;<lpage>3354</lpage>. <pub-id pub-id-type="doi">10.1103/PhysRevA.41.3346</pub-id><pub-id pub-id-type="pmid">9903491</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perdikis</surname> <given-names>D.</given-names></name> <name><surname>Huys</surname> <given-names>R.</given-names></name> <name><surname>Jirsa</surname> <given-names>V. K.</given-names></name></person-group> (<year>2011</year>). <article-title>Time scale hierarchies in the functional organization of complex behaviors</article-title>. <source>PLoS Comput. Biol</source>. <volume>7</volume>:<fpage>e1002198</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002198</pub-id><pub-id pub-id-type="pmid">21980278</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ponce-Alvarez</surname> <given-names>A.</given-names></name> <name><surname>N&#x000E1;cher</surname> <given-names>V.</given-names></name> <name><surname>Luna</surname> <given-names>R.</given-names></name> <name><surname>Riehle</surname> <given-names>A.</given-names></name> <name><surname>Romo</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>Dynamics of cortical neuronal ensembles transit from decision making to storage for later report</article-title>. <source>J. Neurosci</source>. <volume>32</volume>, <fpage>11956</fpage>&#x02013;<lpage>11969</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.6176-11.2012</pub-id><pub-id pub-id-type="pmid">22933781</pub-id></citation></ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rabinovich</surname> <given-names>M. I.</given-names></name> <name><surname>Varona</surname> <given-names>P.</given-names></name></person-group> (<year>2018</year>). <article-title>Discrete sequential information coding: heteroclinic cognitive dynamics</article-title>. <source>Front. Comput. Neurosci</source>. <volume>12</volume>:<fpage>73</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2018.00073</pub-id><pub-id pub-id-type="pmid">30245621</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Recanatesi</surname> <given-names>S.</given-names></name> <name><surname>Katkov</surname> <given-names>M.</given-names></name> <name><surname>Romani</surname> <given-names>S.</given-names></name> <name><surname>Tsodyks</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Neural network model of memory retrieval</article-title>. <source>Front. Comput. Neurosci</source>. <volume>9</volume>:<fpage>149</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2015.00149</pub-id><pub-id pub-id-type="pmid">26732491</pub-id></citation></ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rubin</surname> <given-names>J. J.</given-names></name> <name><surname>Rubin</surname> <given-names>J. E.</given-names></name> <name><surname>Ermentrout</surname> <given-names>G. B.</given-names></name></person-group> (<year>2013</year>). <article-title>Analysis of synchronization in a slowly changing environment: how slow coupling becomes fast weak coupling</article-title>. <source>Phys. Rev. Lett</source>. <volume>110</volume>, <fpage>204101</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.110.204101</pub-id><pub-id pub-id-type="pmid">25167415</pub-id></citation></ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Runyan</surname> <given-names>C. A.</given-names></name> <name><surname>Piasini</surname> <given-names>E.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name> <name><surname>Harvey</surname> <given-names>C. D.</given-names></name></person-group> (<year>2017</year>). <article-title>Distinct timescales of population coding across cortex</article-title>. <source>Nature</source> <volume>548</volume>, <fpage>92</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1038/nature23020</pub-id><pub-id pub-id-type="pmid">28723889</pub-id></citation></ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Russo</surname> <given-names>E.</given-names></name> <name><surname>Treves</surname> <given-names>A.</given-names></name></person-group> (<year>2012</year>). <article-title>Cortical free-association dynamics: distinct phases of a latching network</article-title>. <source>Phys. Rev. E</source> <volume>85</volume>, <fpage>1</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1103/PhysRevE.85.051920</pub-id><pub-id pub-id-type="pmid">23004800</pub-id></citation></ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schuck</surname> <given-names>N. W.</given-names></name> <name><surname>Niv</surname> <given-names>Y.</given-names></name></person-group> (<year>2019</year>). <article-title>Sequential replay of nonspatial task states in the human hippocampus</article-title>. <source>Science</source> <volume>364</volume>. <pub-id pub-id-type="doi">10.1126/science.aaw5181</pub-id><pub-id pub-id-type="pmid">31249030</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Seliger</surname> <given-names>P.</given-names></name> <name><surname>Tsimring</surname> <given-names>L. S.</given-names></name> <name><surname>Rabinovich</surname> <given-names>M. I.</given-names></name></person-group> (<year>2003</year>). <article-title>Dynamics-based sequential memory: winnerless competition of patterns</article-title>. <source>Phys. Rev. E</source> <volume>67</volume>, <fpage>4</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevE.67.011905</pub-id><pub-id pub-id-type="pmid">12636530</pub-id></citation></ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shouval</surname> <given-names>H. Z.</given-names></name> <name><surname>Wang</surname> <given-names>S. S.</given-names></name> <name><surname>Wittenberg</surname> <given-names>G. M.</given-names></name></person-group> (<year>2010</year>). <article-title>Spike timing dependent plasticity: a consequence of more fundamental learning rules</article-title>. <source>Front. Comput. Neurosci</source>. <volume>4</volume>:<fpage>19</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2010.00019</pub-id><pub-id pub-id-type="pmid">20725599</pub-id></citation></ref>
<ref id="B55">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sompolinsky</surname> <given-names>H.</given-names></name> <name><surname>Kanter</surname> <given-names>I.</given-names></name></person-group> (<year>1986</year>). <article-title>Temporal association in asymmetric neural networks</article-title>. <source>Phys. Rev. Lett</source>. <volume>57</volume>, <fpage>2861</fpage>&#x02013;<lpage>2864</lpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.57.2861</pub-id><pub-id pub-id-type="pmid">10033885</pub-id></citation></ref>
<ref id="B56">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stokes</surname> <given-names>M. G.</given-names></name> <name><surname>Kusunoki</surname> <given-names>M.</given-names></name> <name><surname>Sigala</surname> <given-names>N.</given-names></name> <name><surname>Nili</surname> <given-names>H.</given-names></name> <name><surname>Gaffan</surname> <given-names>D.</given-names></name> <name><surname>Duncan</surname> <given-names>J.</given-names></name></person-group> (<year>2013</year>). <article-title>Dynamic coding for cognitive control in prefrontal cortex</article-title>. <source>Neuron</source> <volume>78</volume>, <fpage>364</fpage>&#x02013;<lpage>375</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2013.01.039</pub-id><pub-id pub-id-type="pmid">23562541</pub-id></citation></ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sussillo</surname> <given-names>D.</given-names></name> <name><surname>Abbott</surname> <given-names>L. F.</given-names></name></person-group> (<year>2009</year>). <article-title>Generating coherent patterns of activity from chaotic neural networks</article-title>. <source>Neuron</source> <volume>63</volume>, <fpage>544</fpage>&#x02013;<lpage>557</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2009.07.018</pub-id><pub-id pub-id-type="pmid">19709635</pub-id></citation></ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taghia</surname> <given-names>J.</given-names></name> <name><surname>Cai</surname> <given-names>W.</given-names></name> <name><surname>Ryali</surname> <given-names>S.</given-names></name> <name><surname>Kochalka</surname> <given-names>J.</given-names></name> <name><surname>Nicholas</surname> <given-names>J.</given-names></name> <name><surname>Chen</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Uncovering hidden brain state dynamics that regulate performance and decision-making during cognition</article-title>. <source>Nat. Commun</source>. <volume>9</volume>. <pub-id pub-id-type="doi">10.1038/s41467-018-04723-6</pub-id><pub-id pub-id-type="pmid">29950686</pub-id></citation></ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Verduzco-Flores</surname> <given-names>S. O.</given-names></name> <name><surname>Bodner</surname> <given-names>M.</given-names></name> <name><surname>Ermentrout</surname> <given-names>B.</given-names></name> <name><surname>Oscar</surname> <given-names>S.</given-names></name> <name><surname>Bodner</surname> <given-names>V.-F. M.</given-names></name></person-group> (<year>2012</year>). <article-title>A model for complex sequence learning and reproduction in neural populations</article-title>. <source>J. Comput. Neurosci</source>. <volume>32</volume>, <fpage>403</fpage>&#x02013;<lpage>423</lpage>. <pub-id pub-id-type="doi">10.1007/s10827-011-0360-x</pub-id><pub-id pub-id-type="pmid">21887499</pub-id></citation></ref>
<ref id="B60">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>X.-,j.</given-names></name> <name><surname>Kennedy</surname> <given-names>H</given-names></name></person-group> (<year>2016</year>). <article-title>Brain structure and dynamics across scales : in search of rules</article-title>. <source>Curr. Opin. Neurobiol</source>. <volume>37</volume>, <fpage>92</fpage>&#x02013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2015.12.010</pub-id><pub-id pub-id-type="pmid">26868043</pub-id></citation></ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wasmuht</surname> <given-names>D. F.</given-names></name> <name><surname>Spaak</surname> <given-names>E.</given-names></name> <name><surname>Buschman</surname> <given-names>T. J.</given-names></name> <name><surname>Miller</surname> <given-names>E. K.</given-names></name> <name><surname>Stokes</surname> <given-names>M. G.</given-names></name></person-group> (<year>2018</year>). <article-title>Intrinsic neuronal dynamics predict distinct functional roles during working memory</article-title>. <source>Nat. Commun</source>. <volume>9</volume>, <fpage>3499</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-018-05961-4</pub-id><pub-id pub-id-type="pmid">30158572</pub-id></citation></ref>
<ref id="B62">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Werbos</surname> <given-names>P. J..</given-names></name></person-group> (<year>1990</year>). <article-title>Backpropagation through time: what it does and how to do it</article-title>. <source>Proc. IEEE</source> <volume>78</volume>, <fpage>1550</fpage>&#x02013;<lpage>1560</lpage>. <pub-id pub-id-type="doi">10.1109/5.58337</pub-id><pub-id pub-id-type="pmid">27295638</pub-id></citation></ref>
<ref id="B63">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wernecke</surname> <given-names>H.</given-names></name> <name><surname>S&#x000E1;ndor</surname> <given-names>B.</given-names></name> <name><surname>Gros</surname> <given-names>C.</given-names></name></person-group> (<year>2018</year>). <article-title>Attractor metadynamics in terms of target points in slow-fast systems: adiabatic versus symmetry protected flow in a recurrent neural network</article-title>. <source>J. Phys. Commun</source>. <volume>2</volume>, <fpage>095008</fpage>. <pub-id pub-id-type="doi">10.1088/2399-6528/aac33c</pub-id></citation>
</ref>
<ref id="B64">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wikenheiser</surname> <given-names>A. M.</given-names></name> <name><surname>Schoenbaum</surname> <given-names>G.</given-names></name></person-group> (<year>2016</year>). <article-title>Over the river, through the woods: cognitive maps in the hippocampus and orbitofrontal cortex</article-title>. <source>Nat. Rev. Neurosci</source>. <volume>17</volume>, <fpage>513</fpage>&#x02013;<lpage>523</lpage>. <pub-id pub-id-type="doi">10.1038/nrn.2016.56</pub-id><pub-id pub-id-type="pmid">27256552</pub-id></citation></ref>
<ref id="B65">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wimmer</surname> <given-names>G. E.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Vehar</surname> <given-names>N.</given-names></name> <name><surname>Behrens</surname> <given-names>T. E.</given-names></name> <name><surname>Dolan</surname> <given-names>R. J.</given-names></name></person-group> (<year>2020</year>). <article-title>Episodic memory retrieval success is associated with rapid replay of episode content</article-title>. <source>Nat. Neurosci</source>. <volume>23</volume>, <fpage>1025</fpage>&#x02013;<lpage>1033</lpage>. <pub-id pub-id-type="doi">10.1038/s41593-020-0649-z</pub-id><pub-id pub-id-type="pmid">32514135</pub-id></citation></ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yamashita</surname> <given-names>Y.</given-names></name> <name><surname>Tani</surname> <given-names>J.</given-names></name></person-group> (<year>2008</year>). <article-title>Emergence of functional hierarchy in a multiple timescale neural network model: a humanoid robot experiment</article-title>. <source>PLoS Comput. Biol</source>. <volume>4</volume>:<fpage>e1000220</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1000220</pub-id><pub-id pub-id-type="pmid">18989398</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>According to Eq. (4), <inline-formula><mml:math id="M27"><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02260;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:munder><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>/</mml:mo><mml:mi>d</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x0221D;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02260;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:munder><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>.</p></fn>
<fn id="fn0002"><p><sup>2</sup>The other fixed point corresponding to <inline-formula><mml:math id="M59"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>&#x003BE;</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> also appears, but its basin is quite small. Thus, we can neglect this fixed point.</p></fn>
<fn id="fn0003"><p><sup>3</sup>It is shown that finely designed networks realize the history-dependent sequences (Chartier and Boukadoum, <xref ref-type="bibr" rid="B9">2006</xref>; Verduzco-Flores et al., <xref ref-type="bibr" rid="B59">2012</xref>). However, it is shown that the additional neurons or sub-networks as many as the number of memories or sequences are required in these studies.</p></fn>
</fn-group>
</back>
</article> 