<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="review-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Commun.</journal-id>
<journal-title>Frontiers in Communication</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Commun.</abbrev-journal-title>
<issn pub-type="epub">2297-900X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fcomm.2017.00011</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Communication</subject>
<subj-group>
<subject>Review</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Toward Cognitively Constrained Models of Language Processing: A Review</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Vogelzang</surname> <given-names>Margreet</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x0002A;</xref>
<uri xlink:href="http://frontiersin.org/people/u/388290"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Mills</surname> <given-names>Anne C.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/442792"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Reitter</surname> <given-names>David</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/223520"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Van Rij</surname> <given-names>Jacolien</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/290310"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Hendriks</surname> <given-names>Petra</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/118583"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Van Rijn</surname> <given-names>Hedderik</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/25165"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Center for Language and Cognition Groningen, University of Groningen</institution>, <addr-line>Groningen</addr-line>, <country>Netherlands</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Experimental Psychology, University of Groningen</institution>, <addr-line>Groningen</addr-line>, <country>Netherlands</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Experimental Psychology, University of Oxford</institution>, <addr-line>Oxford</addr-line>, <country>United Kingdom</country></aff>
<aff id="aff4"><sup>4</sup><institution>College of Information Sciences and Technology, The Pennsylvania State University</institution>, <addr-line>State College, PA</addr-line>, <country>United States</country></aff>
<aff id="aff5"><sup>5</sup><institution>Department of Statistical Methods and Psychometrics, University of Groningen</institution>, <addr-line>Groningen</addr-line>, <country>Netherlands</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: &#x000C1;ngel J. Gallego, Universitat Aut&#x000F2;noma de Barcelona, Spain</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Mireille Besson, Cognitive Neuroscience Laboratory, CNRS and Aix-Marseille University, France; Cristiano Chesi, Istituto Universitario di Studi Superiori di Pavia (IUSS), Italy</p></fn>
<corresp content-type="corresp" id="cor1">&#x0002A;Correspondence: Margreet Vogelzang, <email>margreetvogelzang&#x00040;gmail.com</email></corresp>
<fn fn-type="other" id="fn001"><p>Specialty section: This article was submitted to Language Sciences, a section of the journal Frontiers in Communication</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>08</day>
<month>09</month>
<year>2017</year>
</pub-date>
<pub-date pub-type="collection">
<year>2017</year>
</pub-date>
<volume>2</volume>
<elocation-id>11</elocation-id>
<history>
<date date-type="received">
<day>03</day>
<month>04</month>
<year>2017</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>08</month>
<year>2017</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2017 Vogelzang, Mills, Reitter, Van Rij, Hendriks and Van Rijn.</copyright-statement>
<copyright-year>2017</copyright-year>
<copyright-holder>Vogelzang, Mills, Reitter, Van Rij, Hendriks and Van Rijn</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Language processing is not an isolated capacity, but is embedded in other aspects of our cognition. However, it is still largely unexplored to what extent and how language processing interacts with general cognitive resources. This question can be investigated with cognitively constrained computational models, which simulate the cognitive processes involved in language processing. The theoretical claims implemented in cognitive models interact with general architectural constraints such as memory limitations. This way, it generates new predictions that can be tested in experiments, thus generating new data that can give rise to new theoretical insights. This theory-model-experiment cycle is a promising method for investigating aspects of language processing that are difficult to investigate with more traditional experimental techniques. This review specifically examines the language processing models of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>), Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>), and Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>), all implemented in the cognitive architecture Adaptive Control of Thought&#x02014;Rational (Anderson et al., <xref ref-type="bibr" rid="B5">2004</xref>). These models are all limited by the assumptions about cognitive capacities provided by the cognitive architecture, but use different linguistic approaches. Because of this, their comparison provides insight into the extent to which assumptions about general cognitive resources influence concretely implemented models of linguistic competence. For example, the sheer speed and accuracy of human language processing is a current challenge in the field of cognitive modeling, as it does not seem to adhere to the same memory and processing capacities that have been found in other cognitive processes. Architecture-based cognitive models of language processing may be able to make explicit which language-specific resources are needed to acquire and process natural language. The review sheds light on cognitively constrained models of language processing from two angles: we discuss (1) whether currently adopted cognitive assumptions meet the requirements for language processing, and (2) how validated cognitive architectures can constrain linguistically motivated models, which, all other things being equal, will increase the cognitive plausibility of these models. Overall, the evaluation of cognitively constrained models of language processing will allow for a better understanding of the relation between data, linguistic theory, cognitive assumptions, and explanation.</p>
</abstract>
<kwd-group>
<kwd>language processing</kwd>
<kwd>sentence processing</kwd>
<kwd>linguistic theory</kwd>
<kwd>cognitive modeling</kwd>
<kwd>Adaptive Control of Thought&#x02014;Rational</kwd>
<kwd>cognitive resources</kwd>
<kwd>computational simulations</kwd>
</kwd-group>
<counts>
<fig-count count="7"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="114"/>
<page-count count="18"/>
<word-count count="15510"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="introduction">
<title>Introduction</title>
<p>Language is one of the most remarkable capacities of the human mind. Arguably, language is not an isolated capacity of the mind but is embedded in other aspects of cognition. This can be seen in, for example, linguistic recursion. Although linguistic recursion (e.g., &#x0201C;the sister of the father of the cousin of&#x02026;&#x0201D;) could in principle be applied infinitely many times, if the construction becomes too complex we will lose track of its meaning due to memory constraints (Gibson, <xref ref-type="bibr" rid="B47">2000</xref>; Fedorenko et al., <xref ref-type="bibr" rid="B41">2013</xref>). Even though there are ample examples of cognitive resources like memory playing a role in language processing (e.g., King and Just, <xref ref-type="bibr" rid="B69">1991</xref>; Christiansen and Chater, <xref ref-type="bibr" rid="B34">2016</xref>; Huettig and Janse, <xref ref-type="bibr" rid="B58">2016</xref>), it is still largely unexplored to what extent language processing and general cognitive resources interact. That is, which general cognitive resources and which language processing-specific resources are used for language processing? For example, is language processing supported by the same memory system that is used in other cognitive processes? In this review, we will investigate to what extent general cognitive resources limit and influence models of linguistic competence. To this end, we will review cognitively constrained computational models of language processing implemented in the cognitive architecture Adaptive Control of Thought&#x02014;Rational (ACT-R) and evaluate how general cognitive limitations influence linguistic processing in these models. These computational cognitive models explicitly implement theoretical claims, for example about language, based on empirical observations or experimental data. The evaluation of these models will generate new insights about the interplay between language and other aspects of cognition.</p>
<p>Memory is one of the most important general cognitive principles for language processing. In sentence processing, words have to be processed rapidly, because otherwise the memory of the preceding context, necessary for understanding the complete sentence, will be lost (Christiansen and Chater, <xref ref-type="bibr" rid="B34">2016</xref>). Evidence that language processing shares a memory system with other cognitive processes can be found in the relation between general working memory tests and linguistic tests. For example, individual differences in working memory capacity have been found to play a role in syntactic processing (King and Just, <xref ref-type="bibr" rid="B69">1991</xref>), predictive language processing (Huettig and Janse, <xref ref-type="bibr" rid="B58">2016</xref>), and discourse production (Kuijper et al., <xref ref-type="bibr" rid="B73">2015</xref>). Besides memory, other factors like attentional focus (Lewis et al., <xref ref-type="bibr" rid="B77">2006</xref>) and processing speed (Hendriks et al., <xref ref-type="bibr" rid="B57">2007</xref>) have been argued to influence linguistic performance. Thus, it seems apparent that language processing is not an isolated capacity but is embedded in other aspects of cognition. This claim conflicts with the traditional view that language is a specialized faculty (cf. Chomsky, <xref ref-type="bibr" rid="B31">1980</xref>; Fodor, <xref ref-type="bibr" rid="B44">1983</xref>). It is therefore important to note that computational cognitive models can be used to investigate both viewpoints, i.e., to investigate to what extent general cognitive resources can be used in language processing but also to investigate to what extent language is a specialized process. It has also been argued that language processing is a specialized process that is nevertheless influenced by a range of general cognitive resources (cf. Newell, <xref ref-type="bibr" rid="B81">1990</xref>; Lewis, <xref ref-type="bibr" rid="B75">1996</xref>). Therefore, we argue that the potential influence and limitations of general cognitive resources should be taken into account when studying theories of language processing.</p>
<p>To be able to account for the processing limitations imposed by a scarcity of cognitive resources, theories of language need to be specified as explicitly as possible with regards to, for example, processing steps, the incrementality of processing, memory retrievals, and representations. This allows for a specification of what belongs to linguistic competence and what belongs to linguistic performance (Chomsky, <xref ref-type="bibr" rid="B29">1965</xref>): competence is the knowledge a language user has, whereas performance is the output that a language user produces, which results from his competence in combination with other (cognitive) factors (see Figure <xref ref-type="fig" rid="F1">1</xref> for examples). Many linguistic theories have been argued to be theories of linguistic competence that abstract away from details of linguistic performance (Fromkin, <xref ref-type="bibr" rid="B46">2000</xref>). These theories rarely make explicit how the step from competence to performance is made. In order to create a distinction between competence and performance, an increasing emphasis is placed on grounding linguistic theories empirically by creating the step from an abstract theory to concrete, testable predictions (cf. e.g., Kempen and Hoenkamp, <xref ref-type="bibr" rid="B66">1987</xref>; Roelofs, <xref ref-type="bibr" rid="B95">1992</xref>; Baayen et al., <xref ref-type="bibr" rid="B9">1997</xref>; Reitter et al., <xref ref-type="bibr" rid="B90">2011</xref>). Formalizing language processing theories explicitly thus means that the distinction between linguistic competence and linguistic performance can be explained and makes it possible to examine which cognitive resources, according to a language processing theory, are needed to process language (see also Hale, <xref ref-type="bibr" rid="B51">2011</xref>).</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The above graphs show four possible relationships between competence, cognition and performance. Performance is influenced by competence and cognition. If someone&#x02019;s performance (black solid line) increases over age, this could be due to the competence (red dashed line) increasing (as displayed in the upper left graph), or due to cognition (shaded area) increasing, while competence stays constant (as displayed in the upper right graph). Cognitive limitations can prevent performance from reaching full competence (lower left graph). Competence and cognition can also both change over age and influence performance. The lower right graph shows the classical performance curve of U-shaped learning, in which performance initially decreases even though competence is increasing. The graphs are a simplification, as factors other than competence and cognition could also influence performance, for example motor skills.</p></caption>
<graphic xlink:href="fcomm-02-00011-g001.tif"/>
</fig>
<p>The importance of explicitly specified linguistic theories that distinguish between competence and performance can be seen in the acquisition of verbs. Children show a U-shaped learning curve (see Pauls et al., <xref ref-type="bibr" rid="B83">2013</xref> for an overview, U-shaped learning curve is depicted in Figure <xref ref-type="fig" rid="F1">1</xref>) when learning past tenses of verbs, using the correct irregular form first (e.g., the past tense <italic>ate</italic> for <italic>eat</italic>), then using the incorrect regular form of irregular verbs (e.g., <italic>eated</italic>), before using the correct irregular form again. It is conceivable that whereas children&#x02019;s performance initially decreases, children are in the process of learning how to correctly form irregular past tenses and therefore have increasing competence (cf. Taatgen and Anderson, <xref ref-type="bibr" rid="B107">2002</xref>). In this example, explicitly specifying the processing that is needed to form verb tenses and how this processing uses general cognitive resources could explain why children&#x02019;s performance does not match their competence. Another example of performance deviating from competence can be seen in the comprehension and production of pronouns: whereas 6-year-old children generally produce pronouns correctly (they have the competence, see Spenader et al., <xref ref-type="bibr" rid="B102">2009</xref>), they often make mistakes in pronoun interpretation (they show reduced performance, Chien and Wexler, <xref ref-type="bibr" rid="B28">1990</xref>).</p>
<p>Especially when different linguistic theories have been put forward to explain similar phenomena, it is important to be able to compare and test the theories on the basis of concrete predictions. Linguistic theories are often postulated without considering cognitive resources. Therefore, it is important to investigate how well these theories perform under realistic cognitive constraints; this will provide information about their cognitive plausibility. Cognitively constrained computational models (from now on: cognitive models) are a useful tool to compare linguistic theories while taking into account the limitations imposed by a scarcity of cognitive resources and can be used to investigate the relation between underlying linguistic competence and explicit predictions about performance. Thus, by implementing a linguistic theory into a cognitive model, language processing is embedded in other aspects of cognition, and the extent can be investigated to which assumptions about general cognitive resources influence models of linguistic competence.</p>
<p>As cognitive models, we will consider computational models simulating human processing that are constrained by realistic and validated assumptions about human processing. Such cognitive models can generate new predictions that can be tested in further experiments, generating new data that can give rise to new implementations. This theory-model-experiment cycle is a promising method for investigating aspects of language processing that are difficult to investigate with standard experimental techniques, which usually provide insight into performance (e.g., behavior, responses, response times), but not competence. Cognitive models require linguistic theories, that usually describe competence, to be explicitly specified. This way, the performance of competing linguistic theories, which often have different approaches to the structure and interpretation of language, can be investigated using cognitive models. Contrary to other computational modeling methods, cognitive models simulate the processing of a single individual. Because of this, it can be investigated how individual variations in cognitive resources (which can be manipulated in a model) influence a linguistic theory&#x02019;s performance.</p>
<p>The comparison of cognitive models that use different linguistic approaches is most straightforward when they make use of the same assumptions about cognitive resources, and thus are implemented in the same cognitive architecture. This review will therefore focus on cognitive models developed in the same domain-general cognitive architecture, ACT-R (Anderson et al., <xref ref-type="bibr" rid="B5">2004</xref>). There are several other cognitive architectures available (e.g., EPIC: Kieras and Meyer, <xref ref-type="bibr" rid="B67">1997</xref>; NENGO: Stewart et al., <xref ref-type="bibr" rid="B105">2009</xref>), but in order to keep the assumptions about general cognitive resources roughly constant, this review will only consider models implemented in ACT-R. Over the past years, several linguistic phenomena have been implemented in ACT-R, such as metaphors (Budiu and Anderson, <xref ref-type="bibr" rid="B23">2002</xref>), agrammatism (Stocco and Crescentini, <xref ref-type="bibr" rid="B106">2005</xref>), pronominal binding (Hendriks et al., <xref ref-type="bibr" rid="B57">2007</xref>), and presupposition resolution (Brasoveanu and Dotla&#x0010D;il, <xref ref-type="bibr" rid="B21">2015</xref>). In order to obtain a broad view of cognitively constrained models of linguistic theories, we will examine three models of different linguistic modalities (comprehension, production, perspective taking), that all take a different linguistic approach, in depth: the syntactic processing model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>), the syntactic priming model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>), and the pronoun processing model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>). By examining models of different linguistic modalities that take different linguistic approaches, we aim to provide a more unified understanding of how language processing is embedded within general cognition, and investigate how proficient language use is achieved. The selected models are all bounded by the same assumptions about cognitive capacities and seriality of processing as provided by the cognitive architecture ACT-R, which makes them optimally comparable. Their comparison will provide insight into the extent to which assumptions about general cognitive resources influence models of linguistic competence.</p>
<p>This paper is organized as follows. First, we will discuss the components of ACT-R that are most relevant in our discussion of language processing models, in order to explain how cognitive resources play a role in this architecture. Then, we will outline the different linguistic approaches that are used in the models. Finally, we will discuss the selected ACT-R models of language processing in more detail. Importantly, it will be examined how general cognitive resources are used in the models and how these cognitive resources and linguistic principles interact.</p>
</sec>
<sec id="S2">
<title>Basic ACT-R Components</title>
<p>Adaptive Control of Thought&#x02014;Rational (Anderson, <xref ref-type="bibr" rid="B3">1993</xref>, <xref ref-type="bibr" rid="B4">2007</xref>; Anderson et al., <xref ref-type="bibr" rid="B5">2004</xref>) is a cognitive architecture in which models can be implemented to simulate a certain process or collection of processes. Of specific interest for this review is the simulation of language-related processes, such as interpreting or producing a sentence. Cognitive models in ACT-R are restricted by general cognitive resources and constraints embedded in the ACT-R architecture. Examples of such cognitive resources, that are of importance when modeling language, are memory, processing speed, and attention. By implementing a model of a linguistic theory in ACT-R, one can thus examine how this linguistic theory behaves in interaction with other aspects of cognition.</p>
<p>Adaptive Control of Thought&#x02014;Rational aims to explain human cognition as the interaction between a set of functional modules. Each module has a specific function, such as perception, action, memory, and executive function [see Anderson et al. (<xref ref-type="bibr" rid="B5">2004</xref>) for an overview]. Modules can be accessed by the model through buffers. The information in these buffers represents information that is in the focus of attention. Only the information that is in a buffer can be readily used by the model. An overview of the standard ACT-R modules and buffers is shown in Figure <xref ref-type="fig" rid="F2">2</xref>. The modules most relevant for language processing, the declarative memory module and the procedural memory module, will be discussed in more detail below.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>An overview of the standard modules and buffers in Adaptive Control of Thought&#x02014;Rational [based on Anderson et al. (<xref ref-type="bibr" rid="B5">2004</xref>)].</p></caption>
<graphic xlink:href="fcomm-02-00011-g002.tif"/>
</fig>
<p>The declarative memory stores factual information as chunks. Chunks are pieces of knowledge that can store multiple properties, such as that there is a cat with the <italic>name</italic> &#x0201C;Coco,&#x0201D; whose <italic>color</italic> is &#x0201C;gray.&#x0201D; The information in a chunk can only be used after the chunk has been retrieved from the declarative memory and has been placed in the corresponding retrieval buffer. In order to retrieve information from memory, a retrieval request must be made. Only chunks with an activation that exceeds a predetermined activation threshold can be retrieved. The higher the activation of a chunk, the more likely it is to be retrieved. The base-level activation of a chunk increases when a chunk is retrieved from memory, but decays over time. This way, the recency and frequency of a chunk influence a chunk&#x02019;s activation, and thereby its chance of recall and its retrieval time (in line with experimental findings, e.g., Deese and Kaufman, <xref ref-type="bibr" rid="B39">1957</xref>; Allen and Hulme, <xref ref-type="bibr" rid="B2">2006</xref>). Additionally, information that is currently in the focus of attention (i.e., in a buffer) can increase the probability that associated chunks are recalled by adding spreading activation to a chunk&#x02019;s base-level activation. The activation of chunks can additionally be influenced by noise, occasionally causing a chunk with less activation to be retrieved over a chunk with more activation.</p>
<p>Whereas the declarative memory represents factual knowledge, the procedural memory represents knowledge about how to perform actions. The procedural memory consists of production rules, which have an if-then structure. An example of the basic structure of a production rule is as follows:
<list list-type="simple">
<list-item><p>IF</p></list-item>
<list-item><p>a new word is attended</p></list-item>
<list-item><p>THEN</p></list-item>
<list-item><p>retrieve lexical information about this word from memory</p></list-item>
</list></p>
<p>The THEN-part of a production rule is executed when the IF-part matches the current buffer contents. Production rules are executed one by one. If the conditions of several production rules are met, the one with the highest utility is selected. This utility reflects the usefulness the rule has had in the past and can be used to learn from feedback, both positively and negatively (for more detail on utilities, see Anderson et al., <xref ref-type="bibr" rid="B5">2004</xref>). New production rules can be learned on the basis of existing rules and declarative knowledge (production compilation, Taatgen and Anderson, <xref ref-type="bibr" rid="B107">2002</xref>).</p>
<p>Several general cognitive resources and further resources that are important for language processing are incorporated in the ACT-R architecture, such as memory, speed of processing, and attention. Long-term memory corresponds to the declarative module in ACT-R. Short-term or working memory is not incorporated as a separate component in ACT-R (Borst et al., <xref ref-type="bibr" rid="B18">2010</xref>) but emanates from the interaction between the buffers and the declarative memory. Daily et al. (<xref ref-type="bibr" rid="B37">2001</xref>) proposed that the function of working memory can be simulated in ACT-R by associating relevant information with information that is currently in focus (through spreading activation). Thus, working memory capacity can change as a result of a change in the amount of spreading activation in a model.</p>
<p>Crucially, all above mentioned operations take time. Processing in ACT-R is serial, meaning that only one retrieval from declarative memory and only one production rule execution can be done at any point in time (this is known as the serial processing bottleneck, see Anderson, <xref ref-type="bibr" rid="B4">2007</xref>). The retrieval of information from declarative memory is faster and more likely to succeed if a chunk has a high activation (for details see Anderson et al., <xref ref-type="bibr" rid="B5">2004</xref>). Because a chunk&#x02019;s activation increases when it is retrieved, chunks that have been retrieved often will have a high activation and will therefore be retrieved more quickly. Production rules in ACT-R take a standard amount of time to fire (50&#x02009;ms). Rules that are often used in succession can merge into a new production rule. These new rules are a combination of the old rules that were previously fired in sequence, making the model more efficient. Thus, increasing activation and production compilation allow a model&#x02019;s processing speed to increase through practice and experience.</p>
<p>As described, memory and processing speed are examples of general cognitive principles in ACT-R, that will be important when implementing models that perform language processing. In the next section, three linguistic approaches will be discussed. These approaches are relevant for the three cognitive models reviewed in the remainder of the paper.</p>
</sec>
<sec id="S3">
<title>Linguistic Approaches</title>
<p>Cognitive models can be used to implement any linguistic approach, and as such are not bound to one method or theory. In principle any of the theories that have been proposed in linguistics to account for a speaker&#x02019;s linguistic competence, such as Combinatorial Categorial Grammar (Steedman, <xref ref-type="bibr" rid="B103">1988</xref>), construction grammar (Fillmore et al., <xref ref-type="bibr" rid="B43">1988</xref>), generative syntax (Chomsky, <xref ref-type="bibr" rid="B30">1970</xref>), Head-driven Phrase Structure Grammar (Pollard and Sag, <xref ref-type="bibr" rid="B87">1994</xref>), Lexical Functional Grammar (Bresnan, <xref ref-type="bibr" rid="B22">2001</xref>), Optimality Theory (OT) (Prince and Smolensky, <xref ref-type="bibr" rid="B88">1993</xref>), Tree-Adjoining Grammar (Joshi et al., <xref ref-type="bibr" rid="B63">1975</xref>), and usage-based grammar (Bybee and Beckner, <xref ref-type="bibr" rid="B24">2009</xref>) could be implemented in a cognitive model. Note that this does not imply that any linguistic theory or approach can be implemented in any cognitive model, as cognitive models place restrictions on what can and cannot be modeled. Different linguistic approaches tend to entertain different assumptions, for example about what linguistic knowledge looks like (universal principles, violable constraints, structured lexical categories, grammatical constructions), the relation between linguistic forms and their meanings, and the levels of representation needed. This then determines whether and how a particular linguistic approach can be implemented in a particular cognitive model.</p>
<p>In this review, we will discuss three specific linguistic approaches that have been implemented in cognitive models, which allows us to compare how general cognitive resources influence the implementation and output (e.g., responses, response times) of these modeled linguistic approaches. The three linguistic approaches that will be discussed have several features in common but also differ in a number of features: X-bar theory (Chomsky, <xref ref-type="bibr" rid="B30">1970</xref>), Combinatorial Categorial Grammar (Steedman, <xref ref-type="bibr" rid="B103">1988</xref>), and OT (Prince and Smolensky, <xref ref-type="bibr" rid="B88">1993</xref>). These linguistic approaches are implemented in the cognitive models discussed in the next section.</p>
<p>Generative syntax uses X-bar theory to build syntactic structures (Chomsky, <xref ref-type="bibr" rid="B30">1970</xref>). X-bar theory reflects the assumption that the syntactic representation of a clause is hierarchical and can be presented as a binary branching tree. Phrases are built up around a head, which is the principal category. For example, the head of a verb phrase is the verb, and the head of a prepositional phrase is a preposition. To the left or right of this head, other phrases can be attached in the hierarchical structure.</p>
<p>Combinatory Categorial Grammar (CCG) (Steedman, <xref ref-type="bibr" rid="B103">1988</xref>) builds the syntactic structure of a sentence in tandem with the representation of the meaning of the sentence. It is a strongly lexicalized grammar formalism, that proceeds from the assumption that the properties of the grammar follow from the properties of the words in the sentence. That is, each word has a particular lexical category that specifies how that word can combine with other words, and what the resulting meaning will be. In addition, CCG is surface-driven and reflects the assumption that language is processed and interpreted directly, without appealing to an underlying&#x02014;invisible&#x02014;level of representation. For one sentence, CCG can produce multiple representations (Steedman, <xref ref-type="bibr" rid="B103">1988</xref>; Reitter et al., <xref ref-type="bibr" rid="B90">2011</xref>). This allows CCG to build syntactic representations incrementally, from left to right.</p>
<p>The linguistic framework of OT (Prince and Smolensky, <xref ref-type="bibr" rid="B88">1993</xref>) reflects the assumption that language is processed based on constraints on possible outputs (words, sentences, meanings). Based on an input, a set of output candidates is generated. Subsequently, these potential outputs are evaluated based on hierarchically ranked constraints; stronger constraints have priority over weaker constraints. The optimal output is the candidate that satisfies the set of constraints best. The optimal output may be a form (in language production) or a meaning (in language comprehension).</p>
<sec id="S3-1">
<title>Commonalities and Differences</title>
<p>X-bar theory, CCG, and OT have different assumptions about how language is structured. X-bar theory builds a syntactic structure, whereas CCG builds both a syntactic and a semantic representation, and OT builds either a syntactic representation (in language production) or a semantic representation (in language comprehension). Nevertheless, these theories can all be used for the implementation of cognitive models of language processing. In the next section, three cognitive models of language processing will be discussed in detail, with a focus on how the linguistic approaches are implemented and how they interact with other aspects of cognition.</p>
</sec>
</sec>
<sec id="S4">
<title>Cognitive Models of Language Processing</title>
<p>In the following sections, three cognitive language models will be described: the sentence processing model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>), the syntactic priming model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>), and the pronoun processing model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>). The model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) uses a parsing strategy that is based on X-bar theory, the model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>) uses CCG, and the model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) uses OT. The models will be evaluated based on their predictions of novel empirical outcomes and how they achieve these predictions (for example how many parameters are fitted, cf. Roberts and Pashler, <xref ref-type="bibr" rid="B94">2000</xref>). After describing the models separately, the commonalities and differences between these models will be discussed. Based on this, we will review how the interaction between general cognitive resources in ACT-R and linguistic principles from specific linguistic theories can be fruitful in studying cognitive assumptions of linguistic theories.</p>
<sec id="S4-1">
<title>Modeling Sentence Processing as Skilled Memory Retrieval</title>
<p>The first model that we discuss is the sentence processing model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>). This model is a seminal model forming the basis for many later language processing models (a.o., Salvucci and Taatgen, <xref ref-type="bibr" rid="B96">2008</xref>; Engelmann et al., <xref ref-type="bibr" rid="B40">2013</xref>; J&#x000E4;ger et al., <xref ref-type="bibr" rid="B62">2015</xref>). Lewis and Vasishth&#x02019;s (Lewis and Vasishth, <xref ref-type="bibr" rid="B76">2005</xref>) sentenced processing model (henceforth the L&#x00026;V model) performs syntactic parsing based on memory principles: when processing a complete sentence, maintaining the part of the sentence that is already processed in order to integrate it with new incoming information requires (working) memory. The aim of the L&#x00026;V model is to investigate how working memory processes play a role in sentence processing.</p>
<sec id="S4-1-1">
<title>Theoretical Approach</title>
<p>The L&#x00026;V model uses left-corner parsing (Aho and Ullman, <xref ref-type="bibr" rid="B1">1972</xref>), based on X-bar theory (Chomsky, <xref ref-type="bibr" rid="B30">1970</xref>), to build a syntactic representation of the sentence. The left corner (LC) parser builds a syntactic structure of the input sentence incrementally, and predicts the upcoming syntactic structure as new words are encountered. Thus, LC parsing uses information from the words in the sentence to predict what the syntactic structure of that sentence will be. In doing this, LC parsing combines top-down processing, based on syntactic rules, and bottom-up processing, based on the words in a sentence. An example sentence is (1).
<list list-type="simple">
<list-item><label>(1)</label> <p>The dog ran.</p></list-item>
</list></p>
<p>Left corner parsing is based on structural rules, such as those given below as (a)&#x02013;(d). These structural rules for example state that a sentence can be made up of a noun phrase (NP) and a verb phrase [rule (a)], and that a NP can be made up of a determiner and a noun [rule (b)]. An input (word) is nested under the lefthand-side (generally an overarching category) of a structural rule if that rule contains the input on its LC. For example, in sentence (1), <italic>the</italic> is a determiner (Det) according to structural rule (c), which itself is on the LC of rule (b) and thus it is nested under an NP. This NP is on the LC of rule (a). The result of applying these rules is the phrase-structure tree shown in Figure <xref ref-type="fig" rid="F3">3</xref>.</p>
<list list-type="simple">
<list-item><label>(a)</label> <p>S &#x02192; NP VP</p></list-item>
<list-item><label>(b)</label> <p>NP &#x02192; Det N</p></list-item>
<list-item><label>(c)</label> <p>Det &#x02192; the</p></list-item>
<list-item><label>(d)</label> <p>N &#x02192; dog</p></list-item>
</list>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>A tree structure generated by left corner parsing of the word <italic>the</italic> from Example (1) by applying rules (c), (b), and (a) consecutively [based on Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>)].</p></caption>
<graphic xlink:href="fcomm-02-00011-g003.tif"/>
</fig>
<p>Importantly, the generated tree also contains syntactic categories that have not been encountered yet (like N and VP in Figure <xref ref-type="fig" rid="F3">3</xref>), so it contains a prediction of the upcoming sentence structure. When the next word, <italic>dog</italic>, is now encountered, it can be integrated with the existing tree immediately after applying rule (d).</p>
</sec>
<sec id="S4-1-2">
<title>Implementation</title>
<p>The L&#x00026;V model parses a sentence on the basis of guided memory retrievals. Declarative memory is used as the short- and long-term memory needed for sentence processing. The declarative memory holds lexical information as well as any syntactic structures that are built during sentence processing. The activation of these chunks is influenced by the standard ACT-R declarative memory functions, and so their activation (and with this their retrieval probability and latency) is influenced by the recency and frequency with which they were used. Similarity-based interference occurs because the effectiveness of a retrieval request is reduced as the number of items associated with the specific request increases.</p>
<p>Grammatical knowledge however is not stored in the declarative memory but is implemented as procedural knowledge in production rules. That is, the knowledge about how sentences are parsed is stored in a large number of production rules, which interact with the declarative memory when retrieving lexical information or constituents (syntactic structures).</p>
<p>The L&#x00026;V model processes a sentence word for word using the LC parsing algorithm described in Section &#x0201C;<xref ref-type="sec" rid="S4-1-1">Theoretical Approach</xref>.&#x0201D; An overview of the model&#x02019;s processing steps is shown in Figure <xref ref-type="fig" rid="F4">4</xref>. After a word is attended [for example, <italic>the</italic> from Example (1), Box 1], lexical information about this word is retrieved from memory and stored in the lexical buffer (Box 2). Based on the syntactic category of the word and the current state of the model, the model looks for a prior constituent that the new syntactic category could be attached to (Box 3). In our example, <italic>the</italic> is a determiner and it is the first word, so a syntactic structure with a determiner will be retrieved. The model then creates a new syntactic structure by attaching the new word to the retrieved constituent (Box 4). A new word is then attended [<italic>dog</italic> in Example (1), Box 1]. This cycle continues until no new words are left to attend.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Overview of the processing steps of L&#x00026;V&#x02019;s sentence processing model [based on Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>)]. The model processes one word at a time when processing a sentence such as Example (1), first retrieving its lexical information and then retrieving a prior constituent for the new word to be attached to.</p></caption>
<graphic xlink:href="fcomm-02-00011-g004.tif"/>
</fig>
</sec>
<sec id="S4-1-3">
<title>Evaluation</title>
<p>Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) presented several simulation studies, showing that their model can account for reading times from experiments. The model also accounts for the effects of the length of a sentence (short sentences are read faster than long sentences) and structural interference (high interference creates a bigger delay in reading times than low interference) on unambiguous and garden-path sentences. With a number of additions (that are outside the scope of this review), the model can be made to cope with gapped structures and embedded structures, as well as local ambiguity (see Lewis and Vasishth, <xref ref-type="bibr" rid="B76">2005</xref>, for more detail).</p>
<sec id="S4-1-3-1">
<title>Predictions</title>
<p>Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) compared their output to existing experiments, rather than making explicit predictions about new experiments. The model does however provide ideas about <italic>why</italic> any discrepancies between the model and the fitted data occur, which could be seen as predictions, although these predictions have not been tested in new experiments. For example, in a simulation comparing the L&#x00026;V models&#x02019; simulated reading times of subject relative clauses vs. object relative clauses to data from Grodner and Gibson (<xref ref-type="bibr" rid="B49">2005</xref>), the model overestimates the cost of object-gap filling for object relative clauses. The prediction following from the model is that adjusting the latency, a standard ACT-R parameter that influences the time it takes to perform a chunk retrieval, would reduce the difference between model and data. Thus, the prediction is that the retrieval latency of chunks may be lower in this type of language processing than in other cognitive processes.</p>
</sec>
<sec id="S4-1-3-2">
<title>Linguistic Principles</title>
<p>X-bar theory is a widely known approach to syntactic structure. Although already previously implemented as an LC parser (Aho and Ullman, <xref ref-type="bibr" rid="B1">1972</xref>), it is interesting to examine this linguistic approach in interaction with memory functions. Importantly, the use of LC parsing allowed the L&#x00026;V model to use a top-down (prediction-based, cf. Chesi, <xref ref-type="bibr" rid="B27">2015</xref>) as well as bottom-up (input-based, cf. Chomsky, <xref ref-type="bibr" rid="B33">1993</xref>) processing, which increases its efficiency.</p>
</sec>
<sec id="S4-1-3-3">
<title>Cognitive Principles</title>
<p>Many of the cognitive principles used in the L&#x00026;V model are taken directly from ACT-R: memory retrievals are done from declarative memory, the grammatical knowledge needed for parsing is incorporated in production rules, and sentences are processed serially (word by word). Memory plays a very important role in the model, as processing sentences requires memory of the recent past. For all memory functions, the same principles of declarative memory are used as would be used for non-linguistic processes. For the L&#x00026;V model, the standard ACT-R architecture was expanded with a lexical buffer, which holds a lexical chunk after it is retrieved from the declarative memory. Thus, the model assumes the use of general memory functions for language processing, but added a specific attention component to store linguistic (lexical) information that is in the focus of attention.</p>
<p>The speed of processing required for language processing is achieved in the L&#x00026;V model by keeping the model&#x02019;s processing down to the most efficient way to do things: the processing of a word takes a maximum of three production rules and two memory retrievals, serially. This however includes only the syntactic processing, and not, for example, any semantic processing. It remains to be investigated therefore how the model would function if more language processing elements, that take additional time to be executed due to the serial processing bottleneck, are added.</p>
</sec>
<sec id="S4-1-3-4">
<title>Limitations and Future Directions</title>
<p>Although the simulations show a decent fit when compared to data from several empirical experiments, there are a number of phenomena for which a discrepancy is found between the simulation data and some of the experimental data. Specifically, the L&#x00026;V model overestimates effects of the length of a sentence and underestimates interference effects. Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) indicated that part of this discrepancy may be resolved by giving more weight to decay and less weight to interference in the model, but leave the mechanisms responsible for length effects and interference effects open for future research.</p>
<p>Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) acknowledged that the model is a first step to modeling complete sentence comprehension and indicated that future extensions might lie in the fields of semantic and discourse processing, the interaction between lexical and syntactic processing, and investigating individual performance based on working memory capacity differences. Indeed, this sentence processing model is an influential model that has served as a building block for further research. For example, Engelmann et al. (<xref ref-type="bibr" rid="B40">2013</xref>) used the sentence processing model to study the relation between syntactic processing and eye movements, Salvucci and Taatgen (<xref ref-type="bibr" rid="B96">2008</xref>) used the model in their research of multitasking, and Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) and Vogelzang (<xref ref-type="bibr" rid="B114">2017</xref>) build their OT model of pronoun resolution on top of L&#x00026;V&#x02019;s syntactic processing model.</p>
</sec>
</sec>
</sec>
<sec id="S4-2">
<title>Modeling Syntactic Priming in Language Production</title>
<p>A second model discussed in this paper is the ACT-R model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>). Their model (henceforth the RK&#x00026;M model) investigates syntactic priming in language production. Speakers have a choice between different words and grammatical structures to express their ideas. They tend to repeat previously encountered grammatical structures, a pattern of linguistic behavior that is referred to as syntactic or structural priming (for a review, see Pickering and Ferreira, <xref ref-type="bibr" rid="B86">2008</xref>). For example, Bock (<xref ref-type="bibr" rid="B14">1986</xref>) found that when speakers were presented with a passive construction such as <italic>The boy was kissed by the girl</italic> as a description of a picture, they were more likely to describe a new picture using a similar syntactic structure. Effects of priming have been detected with a range of syntactic constructions, including NP variants (Cleland and Pickering, <xref ref-type="bibr" rid="B35">2003</xref>), the order of main and auxiliary verbs (Hartsuiker and Westenberg, <xref ref-type="bibr" rid="B54">2000</xref>), and other structures, in a variety of languages (Pickering and Ferreira, <xref ref-type="bibr" rid="B86">2008</xref>), and in children (Huttenlocher et al., <xref ref-type="bibr" rid="B59">2004</xref>; Van Beijsterveldt and Van Hell, <xref ref-type="bibr" rid="B109">2009</xref>), but also syntactic phrase-structure rules in general (Reitter et al., <xref ref-type="bibr" rid="B93">2006</xref>; Reitter and Moore, <xref ref-type="bibr" rid="B92">2014</xref>).</p>
<p>In the literature, a number of factors that interact with priming have been identified:
<list list-type="bullet">
<list-item><p>Cumulativity: priming strengthens with each copy of the primed construction (Jaeger and Snider, <xref ref-type="bibr" rid="B60">2008</xref>).</p></list-item>
<list-item><p>Decay: the probability of occurrence of a syntactic construction decays over time (Branigan et al., <xref ref-type="bibr" rid="B19">1999</xref>).</p></list-item>
<list-item><p>Lexical boost: lexically similar materials increase the chance that priming will occur (Pickering and Branigan, <xref ref-type="bibr" rid="B85">1998</xref>).</p></list-item>
<list-item><p>Inverse frequency interaction: priming by less frequent constructions is stronger (Scheepers, <xref ref-type="bibr" rid="B97">2003</xref>).</p></list-item>
</list></p>
<p>Besides these factors, differences have been found between fast, short-term priming and slow, long-term adaptation, which is a learning effect that can persist over several days (Bock et al., <xref ref-type="bibr" rid="B15">2007</xref>; Kaschak et al., <xref ref-type="bibr" rid="B65">2011b</xref>). These two different priming effects have been suggested to use separate underlying mechanisms (Hartsuiker et al., <xref ref-type="bibr" rid="B52">2008</xref>), and as such may rely on different cognitive resources.</p>
<p>Syntactic priming is seen as an important effect by which to validate models of syntactic representations and associated learning. Several other models of syntactic priming were proposed (Chang et al., <xref ref-type="bibr" rid="B26">2006</xref>; Snider, <xref ref-type="bibr" rid="B101">2008</xref>; Malhotra, <xref ref-type="bibr" rid="B78">2009</xref>), but none of these are able to account for all mentioned factors as well as short and long term priming. The goal of the RK&#x00026;M model is thus to account for all types of syntactic priming within a cognitive architecture.</p>
<sec id="S4-2-1">
<title>Theoretical Approach</title>
<p>The RK&#x00026;M model is based on a theoretical approach that explains priming as facilitation of lexical-syntactic access. The model bases its syntactic composition process on a broad-coverage grammar framework, CCG (see <xref ref-type="sec" rid="S3">Linguistic Approaches</xref>, Steedman, <xref ref-type="bibr" rid="B103">1988</xref>, <xref ref-type="bibr" rid="B104">2000</xref>). Categorial Grammars use a small set of combinatory rules and a set of parameters to define the basic operations that yield sentences in a specific language. Most specific information is stored in the lexicon. With the use of CCG, the RK&#x00026;M model implements the idea of combinatorial categories as in Pickering and Branigan&#x02019;s (Pickering and Branigan, <xref ref-type="bibr" rid="B85">1998</xref>) model.</p>
<p>In CCG, the syntactic process is the result of combinations of adjacent words and phrases (in constituents). Unlike classical phrase-structure trees, however, the categories that classify each constituent reflect its syntactic and semantic status by stating what other components are needed before a sentence results. For example, the phrase <italic>loves toys</italic> needs to be combined with a NP to its left, as in Example 2. This phrase is assigned the category S&#x0005C;NP. Similarly, the phrase <italic>Dogs love</italic> requires a NP to its right to be complete, thus, its category is S//NP. Many analyses (derivations) of a given sentence are possible in CCG.</p>
<list list-type="simple">
<list-item><label>(2)</label> <p>Dogs love toys.</p></list-item>
</list>
<p>Combinatory Categorial Grammar allows the RK&#x00026;M model to generate a syntactic construction incrementally, so that a speaker can start speaking before the entire sentence is planned. However, it also allows the planning of a full sentence before a speaker starts speaking. CCG is generally underspecified and generates more sentences than would be judged acceptable. The RK&#x00026;M model at least partially addresses this over-generation by employing memory-based ACT-R mechanisms, which also help in providing a cognitively plausible version of a language model.</p>
</sec>
<sec id="S4-2-2">
<title>Implementation</title>
<p>In the RK&#x00026;M model, lexical forms and syntactic categories are stored in chunks in declarative memory. The activation of any chunk in ACT-R is determined by previous occurrences, which causes previously used, highly active chunks to have a higher retrieval probability, creating a priming effect.</p>
<p>The RK&#x00026;M model additionally uses spreading activation to activate all syntax chunks that are associated with a lexical form, creating the possibility to express a meaning in multiple ways. Some ways of expressing a meaning are more frequent in language than others, and therefore the amount of spreading activation from a lexical form to a syntax chunk is mediated by the frequency of the syntactic construction. This causes more frequent forms to have a higher activation and therefore to be more likely to be selected. However, a speaker&#x02019;s choice of syntactic construction can vary on the basis of priming and noise.</p>
<p>To make its theoretical commitments to cue-based, frequency- and recency-governed declarative retrieval, as well as its non-commitments to specific production rules and their timing more clear, the RK&#x00026;M model was implemented first in ACT-R 6, and then in the ACT-UP implementation of the ACT-R theory (Reitter and Lebiere, <xref ref-type="bibr" rid="B91">2010</xref>).</p>
<sec id="S4-2-2-1">
<title>Syntactic Realization</title>
<p>The RK&#x00026;M model takes a semantic description of a sentence as input and creates a syntactic structure for this input. The serially executed processing steps of the model are shown in Figure <xref ref-type="fig" rid="F5">5</xref> and will be explained on the basis of Example (3).</p>
<list list-type="simple">
<list-item><label>(3)</label> <p>Sharks bite.</p></list-item>
</list>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Overview of the processing steps of RK&#x00026;M&#x02019;s syntactic priming model, which produces the syntactic structure of a sentence such as Example (3) [based on Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>)]. First, retrievals of the lexical form of the head and a thematic role are done. Then, the model selects an argument for the thematic role and retrieves a syntax chunk before combining the information according to combinatorial rules of Combinatory Categorial Grammar in the adjoin phase.</p></caption>
<graphic xlink:href="fcomm-02-00011-g005.tif"/>
</fig>
<p>First, the model retrieves a lexical form for the head of the sentence (Box 1). In Example (3), this head will be the verb <italic>bite</italic>. Then the most active thematic role is retrieved from memory (Box 2), which would be the &#x0201C;agent-role&#x0201D; in our example. If no next thematic role can be retrieved, the entire sentence has been generated and an output can be given. The model then identifies the argument associated with the retrieved thematic role and retrieves a lexical form for this argument (Box 3). In the case of the agent-role in Example (3), this will be <italic>sharks</italic>. Following, the model retrieves a syntax chunk that is associated with the retrieved lexical form (Box 4). The lexical form was <italic>sharks</italic>, and the corresponding syntax chunk will thus indicate that this is an NP, and that it needs a verb to its right (S/VP). Finally, the model adjoins the new piece of syntactic information with the syntactic structure of the phrase thus far (Box 5), according to the combinatorial rules of CCG. The model then goes back to retrieving the next thematic role (Box 2) and repeats this process until the entire sentence has been generated.</p>
</sec>
<sec id="S4-2-2-2">
<title>Priming</title>
<p>Within the language production process, syntactic choice points (Figure <xref ref-type="fig" rid="F5">5</xref>, Box 4) will occur, during which a speaker decides between several possible syntactic variants. The model needs to explicate the probability distribution over possible decisions at that point. This can be influenced by priming.</p>
<p>The time course of priming is of concern in the RK&#x00026;M model. Immediately after a prime, repetition probability is strongly elevated. The model uses two default ACT-R mechanisms, base-level learning and spreading activation, to account for long-term adaptation and short-term priming. Short-term priming emerges from a combination of two general memory effects: (1) rapid temporal decay of syntactic information and (2) cue-based memory retrieval subject to interfering and facilitating semantic information (Reitter et al., <xref ref-type="bibr" rid="B90">2011</xref>). Long-term priming effects in the model emerge from the increase in base-level activation that occurs when a chunk is retrieved.</p>
</sec>
</sec>
<sec id="S4-2-3">
<title>Evaluation</title>
<p>In the RK&#x00026;M model, base-level learning and spreading activation account for long-term adaptation and short-term priming, respectively. By simulating a restricted form of incremental language production, it accounts for (a) the inverse frequency interaction (Scheepers, <xref ref-type="bibr" rid="B97">2003</xref>; Reitter, <xref ref-type="bibr" rid="B89">2008</xref>; Jaeger and Snider, <xref ref-type="bibr" rid="B61">2013</xref>); (b) the absence of a decay in long-term priming (Hartsuiker and Kolk, <xref ref-type="bibr" rid="B53">1998</xref>; Bock and Griffin, <xref ref-type="bibr" rid="B16">2000</xref>; Branigan et al., <xref ref-type="bibr" rid="B20">2000</xref>; Bock et al., <xref ref-type="bibr" rid="B15">2007</xref>); and (c) the cumulativity of long-term adaptation (Jaeger and Snider, <xref ref-type="bibr" rid="B60">2008</xref>). The RK&#x00026;M model also explains the lexical boost effect and the fact that it only applies to short-term priming, because semantic information is held in short-term memory and serves as a source of activation for associated syntactic material.</p>
<p>The model uses lexical-syntactic associations as in the residual-activation account (Pickering and Branigan, <xref ref-type="bibr" rid="B85">1998</xref>). However, learning remains an implicit process, and routinization (acquisition of highly trained sequences of actions) may still occur, as it would in implicit learning accounts.</p>
<p>The RK&#x00026;M model accounts for a range of priming effects, but despite providing an account of grammatical encoding, it has not been implemented to explain how speakers construct complex sentences using the broad range of syntactic constructions found in a corpus.</p>
<sec id="S4-2-3-1">
<title>Predictions</title>
<p>Because semantic information is held in short-term memory and serves as a source of activation for associated syntactic material, the RK&#x00026;M model predicts that lexical boost occurs with the repetition of any lexical material with semantic content, rather than just with repeated head words. This prediction was confirmed with corpus data (Reitter et al., <xref ref-type="bibr" rid="B90">2011</xref>) and also experimentally (Scheepers et al., <xref ref-type="bibr" rid="B98">2017</xref>). The RK&#x00026;M model also predicts that only content words cause a lexical boost effect. This prediction was not tested on the corpus, although it is compatible with prior experimental results using content words (Corley and Scheepers, <xref ref-type="bibr" rid="B36">2002</xref>; Schoonbaert et al., <xref ref-type="bibr" rid="B99">2007</xref>; Kootstra et al., <xref ref-type="bibr" rid="B71">2012</xref>) and semantically related words (Cleland and Pickering, <xref ref-type="bibr" rid="B35">2003</xref>), and the insensitivity of priming to closed-class words (Bock and Kroch, <xref ref-type="bibr" rid="B17">1989</xref>; Pickering and Branigan, <xref ref-type="bibr" rid="B85">1998</xref>; Ferreira, <xref ref-type="bibr" rid="B42">2003</xref>).</p>
<p>The model predicted cumulativity of prepositional-object construction priming, and it suggested that double-object constructions are ineffective as primes to the point where cumulativity cannot be detected. In an experimental study published later by another lab (Kaschak et al., <xref ref-type="bibr" rid="B64">2011a</xref>), this turned out to be the case.</p>
</sec>
<sec id="S4-2-3-2">
<title>Linguistic Principles</title>
<p>An important aspect of the RK&#x00026;M model is that it uses CCG. This allows the model to realize syntactic constructions both incrementally and non-incrementally, without storing large amounts of information. CCG can produce multiple representations of the input at the same time, which reflect the choices that a speaker can make. CCG has enjoyed substantial use on large-scale problems in computational linguistics in recent years. Still, how much does this theoretical commitment (of CCG) limit the model&#x02019;s applicability? The RK&#x00026;M model relies, for its account of grammatical encoding, on the principles of incremental planning made possible by categorial grammars. However, for its account of syntactic priming, the deciding principle is that the grammar is lexicalized, and that syntactic decisions involve lower-frequency constructions that are retrieved from declarative (lexical) memory. Of course, ACT-R as a cognitive framework imposes demands on what the grammatical encoder can and cannot do, chiefly in terms of working memory: large, complex symbolic representations such as those necessary to process subtrees in Tree-Adjoining Grammar (Joshi et al., <xref ref-type="bibr" rid="B63">1975</xref>), or large feature structures of unification-based formalisms such as Head-driven Phrase Structure Grammar (Pollard and Sag, <xref ref-type="bibr" rid="B87">1994</xref>) would be implausible under the assumptions of ACT-R.</p>
</sec>
<sec id="S4-2-3-3">
<title>Cognitive Principles</title>
<p>The RK&#x00026;M model&#x02019;s linguistic principles are intertwined with cognitive principles in order to explain priming effects. Declarative memory retrievals and the accompanying activation boost cause frequently used constructions to be preferred. Additionally, the model uses the default ACT-R component of spreading activation to give additional activation to certain syntax chunks, increasing the likelihood that a specific syntactic structure will be used. Working memory capacity is not specified in the RK&#x00026;M model.</p>
<p>The RK&#x00026;M model is silent with respect to the implementation of its grammatical encoding algorithms. Standard ACT-R provides for production rules that represent routinized skills. These rules are executed at a rate of one every 50&#x02009;ms. Whether that is fast enough for grammatical encoding when assuming a serial processing bottleneck, and how production compilation can account for fast processing, is unclear at this time. Production compilation, in ACT-R, can combine a sequence of rule invocations and declarative retrievals into a single, large and efficient production rule. An alternative explanation may be that the production rule system associated with the syntactic process is not implemented by the <italic>basal ganglia</italic>, the brain structure normally associated with ACT-R&#x02019;s production rules, but by a language-specific region such as <italic>Broca&#x02019;s area</italic>. This language-specific region may allow for faster processing.</p>
</sec>
<sec id="S4-2-3-4">
<title>Limitations and Future Directions</title>
<p>Some effects related to syntactic priming remain unexplained by the RK&#x00026;M model. For example, the repetition of thematic and semantic assignments between sentences (Chang et al., <xref ref-type="bibr" rid="B25">2003</xref>) is not a consequence of retrieval of lexical-syntactic material. A future ACT-R model can make use of working memory accounts (cf. Van Rij et al., <xref ref-type="bibr" rid="B113">2013</xref>) to explain repetition preferences leading to such effects.</p>
</sec>
</sec>
</sec>
<sec id="S4-3">
<title>Modeling the Acquisition of Object Pronouns</title>
<p>The third and final model that is discussed, is Van Rij et al.&#x02019;s (<xref ref-type="bibr" rid="B112">2010</xref>) model for the acquisition of the interpretation of object pronouns (henceforth the RR&#x00026;H model). In languages such as English and Dutch, an object pronoun (<italic>him</italic> in Example 4) cannot refer to the local subject (<italic>the penguin</italic> in Example 4, cf. e.g., Chomsky, <xref ref-type="bibr" rid="B32">1981</xref>). Instead, it must refer to another referent in the context, in our example <italic>the sheep</italic>. In contrast, reflexives such as &#x0201C;zichzelf&#x0201D; (<italic>himself, herself</italic>) can only refer to the local subject.</p>
<list list-type="simple">
<list-item><label>(4)</label> <p>Look, a penguin and a sheep. The penguin is hitting him/himself.</p></list-item>
</list>
<p>Children up to age seven allow the unacceptable interpretation of the object pronoun &#x0201C;him&#x0201D; (<italic>the penguin</italic>), although children perform adult-like on the interpretation of reflexives from the age of four (e.g., Chien and Wexler, <xref ref-type="bibr" rid="B28">1990</xref>; Philip and Coopmans, <xref ref-type="bibr" rid="B84">1996</xref>). Interestingly, children as early as 4&#x02009;years old show adult-like production of object pronouns and reflexives (e.g., De Villiers et al., <xref ref-type="bibr" rid="B38">2006</xref>; Spenader et al., <xref ref-type="bibr" rid="B102">2009</xref>). The ACT-R model is used to investigate why children show difficulties interpreting object pronouns, but not interpreting reflexives or producing object pronouns or reflexives.</p>
<sec id="S4-3-1">
<title>Theoretical Account</title>
<p>To explain the described findings on the interpretation of object pronouns and reflexives, Hendriks and Spenader (<xref ref-type="bibr" rid="B56">2006</xref>) proposed that children do not lack the linguistic knowledge needed for object pronoun interpretation but fail to take into account the speaker&#x02019;s perspective. According to this account, formulated within OT (Prince and Smolensky, <xref ref-type="bibr" rid="B88">1993</xref>, see <xref ref-type="sec" rid="S3">Linguistic Approaches</xref>), object pronouns compete with reflexives in their use and interpretation.</p>
<p>In the account of Hendriks and Spenader (<xref ref-type="bibr" rid="B56">2006</xref>), two grammatical constraints guide the production and interpretation of pronouns and reflexives. &#x0201C;Principle A&#x0201D; is the strongest constraint, which states that reflexives have the same reference as the subject of the clause. In production, Hendriks and Spenader assume a general preference for producing reflexives over pronouns, which is formulated in the constraint &#x0201C;Avoid Pronouns.&#x0201D;</p>
<p>Hendriks and Spenader (<xref ref-type="bibr" rid="B56">2006</xref>) argue that the interpretation of object pronouns is not ambiguous for adults, because they take into account the speakers&#x02019; perspective: if the speaker wanted to refer to the subject (e.g., <italic>the penguin</italic> in Example 4), then the speaker would have used a reflexive in accordance with the constraint Principle A. When the speaker did not use a reflexive, therefore, an adult listener should be able to conclude that the speaker must have wanted to refer to another referent. Although this account can explain the asymmetry in children&#x02019;s production and interpretation of object pronouns, it does not provide a theory on how children acquire the interpretation of object pronouns. To investigate this question, the theoretical account of Hendriks and Spenader was implemented in ACT-R (Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>; see also Hendriks et al., <xref ref-type="bibr" rid="B57">2007</xref>).</p>
</sec>
<sec id="S4-3-2">
<title>Implementation</title>
<p>An overview of the RR&#x00026;H model is presented in Figure <xref ref-type="fig" rid="F6">6</xref>. The process of finding the optimal meaning for a form (in comprehension) or finding the optimal form for a meaning (in production) was implemented in ACT-R as a serial process. To illustrate the process, consider the interpretation of the pronoun <italic>him</italic>.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>The optimization process as implemented in the RR&#x00026;H model to determine the optimal meaning for a form such as <italic>him</italic> in Example (4) (in comprehension), and the optimal form for a meaning such as co-reference [(in production), based on Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>)].</p></caption>
<graphic xlink:href="fcomm-02-00011-g006.tif"/>
</fig>
<sec id="S4-3-2-1">
<title>Using Grammatical Constraints</title>
<p>When interpreting a pronoun, two consecutive production rules request the retrieval of two candidate interpretations from the model&#x02019;s declarative memory (Box 1 and Box 2 in Figure <xref ref-type="fig" rid="F6">6</xref>). The two candidate interpretations are the co-referential interpretation (i.e., reference to the referent expressed by the local subject, e.g., <italic>the penguin</italic> in Example 4) and the disjoint interpretation (i.e., reference to another referent in the discourse, such as <italic>the sheep</italic> in Example 4). Consequently, a production rule requests the retrieval of a grammatical constraint from declarative memory. The chunk that represents the constraint Principle A has the highest activation because it is the strongest constraint and is retrieved from memory first (see Box 3).</p>
<p>On the basis of the retrieved constraint, the two candidate interpretations are evaluated (Box 4 and 5). If one of the candidates violates the constraint, the RR&#x00026;H model tries to replace that candidate by a new candidate (Box 4 and Box 2). If it cannot find a new candidate in memory, the remaining candidate is selected as the optimal interpretation.</p>
<p>If the input was a pronoun, however, none of the candidate interpretations violates Principle A. Therefore, both candidate interpretations are still possible (Box 5). In this situation, the RR&#x00026;H model retrieves a new constraint (Box 3), Avoid Pronouns. This constraint cannot distinguish between the two candidate meanings either, because it only applies to forms. As both the co-referential and the disjoint interpretation are still possible, the model randomly selects one of the two candidates as the optimal interpretation. The random choice between two optimal candidates reflects children&#x02019;s behavior in the interpretation of object pronouns.</p>
</sec>
<sec id="S4-3-2-2">
<title>Perspective Taking</title>
<p>After selecting the optimal interpretation, the RR&#x00026;H model takes the speaker&#x02019;s perspective to verify whether the speaker indeed intended to express the selected interpretation (see Figure <xref ref-type="fig" rid="F7">7</xref>). Taking the speaker&#x02019;s perspective, the model uses the same optimization mechanism, but now the input is the <italic>meaning</italic> (optimal interpretation) selected in the previous step when taking the listener&#x02019;s perspective (m<sub>1</sub>), and the output is the optimal form to express that meaning (f<sub>2</sub>).</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>An optimization process from the perspective of the listener as well as an optimization process from the perspective of the speaker is performed [based on Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>)].</p></caption>
<graphic xlink:href="fcomm-02-00011-g007.tif"/>
</fig>
<p>Continuing with the example of processing an object pronoun, the model could have selected the co-referential interpretation as the interpretation of the object pronoun when taking the listener&#x02019;s perspective. In that situation, the input (m<sub>1</sub>) for the second optimization step, using the speaker&#x02019;s perspective, would be the co-referential interpretation. The output of the second optimization step (f<sub>2</sub>) is the reflexive form, because the constraint Avoid Pronouns favors the use of a reflexive over a pronoun.</p>
<p>After the two optimization steps, a new production rule fires that compares the initial input (the object pronoun) with the output (a reflexive, Figure <xref ref-type="fig" rid="F7">7</xref> Box 3). As these forms are not identical in our example, the model concludes that a co-referential interpretation is not intended by the speaker: the speaker would have used a reflexive rather than a pronoun to express a co-referential interpretation. As a consequence, the model will take an alternative candidate interpretation, the disjoint interpretation, and will check if the speaker could have intended a disjoint interpretation.</p>
<p>Alternatively, if the model had selected a disjoint interpretation for a pronoun during the first optimization step, the input for the speaker&#x02019;s perspective (m<sub>1</sub>) would be a disjoint interpretation. The constraint Principle A would cause the model to select a pronoun rather than a reflexive for expressing the disjoint interpretation (f<sub>2</sub>). As the original input (f<sub>1</sub>, a pronoun) and the output (f<sub>2</sub>, also a pronoun) are identical, the model concludes that the speaker indeed intended a disjoint interpretation.</p>
<p>Although children are expected to use the same perspective taking mechanism as adults, it is assumed that children&#x02019;s processing is initially too slow to complete this process. The time for pronoun resolution is limited: When the next word comes, the model stops processing the pronoun and redirects its attention to the new word. Gradually however, children&#x02019;s processing becomes more efficient due to ACT-R&#x02019;s default mechanism of production compilation (Taatgen and Anderson, <xref ref-type="bibr" rid="B107">2002</xref>). This way, the process becomes more efficient, and over time it is possible to take the perspective of the speaker into account in interpretation.</p>
</sec>
</sec>
<sec id="S4-3-3">
<title>Evaluation</title>
<p>The RR&#x00026;H model explains the delay in object pronoun acquisition as arising from the interaction between general cognitive principles and specific linguistic constraints. The model simulations show that children&#x02019;s non-adult-like performance does not necessarily arise from differences in linguistic knowledge or differences in processing mechanism but may arise because children lack processing efficiency.</p>
<sec id="S4-3-3-1">
<title>Predictions</title>
<p>From the RR&#x00026;H model simulations, a new prediction was formulated: when children receive sufficient time for pronoun interpretation, they will show more adult-like performance on object pronoun interpretation. Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) tested this prediction by slowing down the speech rate. They found that children indeed performed significantly more adult-like on object pronoun interpretation when they were presented with slowed-down speech compared to normal speech. A second prediction of the RR&#x00026;H model is that the use of perspective taking in pronoun interpretation is dependent on the input frequency of pronouns. With higher input frequency, the process becomes more efficient in a shorter time (Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>; Hendriks, <xref ref-type="bibr" rid="B55">2014</xref>).</p>
</sec>
<sec id="S4-3-3-2">
<title>Linguistic Principles</title>
<p>The linguistic principles incorporated in the RR&#x00026;H model is rooted in OT. The underlying idea in OT is that an in principle infinite set of potential candidates is evaluated on the basis of all constraints of the grammar. The serial optimization mechanism implemented in the model is a more constrained version of optimization: the two most likely candidates are compared using the constraints that are most relevant in the context. In this respect, the optimization mechanism could be applied to other linguistic (and non-linguistic) phenomena and is thus potentially generalizable.</p>
</sec>
<sec id="S4-3-3-3">
<title>Cognitive Principles</title>
<p>Several general cognitive principles are used in the RR&#x00026;H model. Production compilation learning allowed the model to gradually derive an efficient variant of the general cognitive skill of perspective taking that is specialized for object pronoun interpretation. This specialization mechanism has been applied to model other linguistic and non-linguistic phenomena (e.g., Taatgen and Anderson, <xref ref-type="bibr" rid="B107">2002</xref>). Through the increased efficiency of production rules, as well as through increasing activation of candidates and constraints that were used for pronoun interpretation, the model&#x02019;s processing speed increases over time.</p>
<p>The RR&#x00026;H model uses ACT-R&#x02019;s declarative memory for the storage and retrieval of candidates and constraints. However, no discourse processing was included in the model, and no working memory component was used. Therefore, a remaining question is whether, contrary to what is assumed in other research (Christiansen and Chater, <xref ref-type="bibr" rid="B34">2016</xref>), processing speed limitations on pronoun processing are not imposed by working memory limitations, but by processing efficiently limitations (cf. Kuijper, <xref ref-type="bibr" rid="B72">2016</xref>).</p>
<p>RR&#x00026;H&#x02019;s account of the difference between children&#x02019;s and adults&#x02019; processing of pronouns crucially follows from the serial processing bottleneck assumption, as it assumes that children have the knowledge necessary to use bidirectional optimization, including all relevant linguistic knowledge, but cannot make use of it due to time limitations. Proceduralization is used as the explanation for how children arrive at adult performance given the serial processing bottleneck.</p>
</sec>
<sec id="S4-3-3-4">
<title>Limitations and Future Directions</title>
<p>A potential limitation of RR&#x00026;H&#x02019;s object pronoun processing model is that it is not yet clear how to determine the two most likely candidates or how the model can decide what the most relevant constraint is. Another simplification is that both candidate referents were introduced in the previous sentence. An interesting extension of the model would be one in which the discourse status of the referents would also be taken into account (cf. Van Rij et al., <xref ref-type="bibr" rid="B113">2013</xref>). The extended model would need to integrate factors such as first-mention, frequency, recency, grammatical role and role parallelism (Lappin and Leass, <xref ref-type="bibr" rid="B74">1994</xref>), and semantic role (Kong et al., <xref ref-type="bibr" rid="B70">2009</xref>) to account for topicality and the discourse prominence of referents (Grosz et al., <xref ref-type="bibr" rid="B50">1995</xref>), which plays an important role in pronoun resolution (Spenader et al., <xref ref-type="bibr" rid="B102">2009</xref>).</p>
<p>Another future direction for this research would be to investigate why children as early as 4&#x02009;years old in languages such as Italian and Spanish do not allow unacceptable reference to the local subject for object pronouns (Italian: McKee, <xref ref-type="bibr" rid="B80">1992</xref>; for an overview on Italian see Belletti and Guasti, <xref ref-type="bibr" rid="B13">2015</xref>; Spanish: Baauw, <xref ref-type="bibr" rid="B8">2002</xref>), in contrast to children in languages such as English and Dutch. Thus, this cognitive model could be applied to investigate cross-linguistic variation.</p>
</sec>
</sec>
</sec>
<sec id="S4-4">
<title>Commonalities and Differences</title>
<p>In the previous sections, we discussed three language processing models in ACT-R that were based on different linguistic approaches. The models were all implemented in the same cognitive architecture, so they are all constrained by the same limitations on cognitive resources. This allows for their comparison, which can provide information about how different aspects of language processing interact with non-linguistic aspects of cognition, and how models addressing different linguistic phenomena can be integrated. In this section, we will discuss the commonalities and differences between these models in more detail, so it can be examined to which extent assumptions about general cognitive resources influence implementations of these specific linguistic approaches. Additionally, their comparison will provide an overview of some choices that can be made when implementing a language processing model, such as how to represent (grammatical) knowledge, and how these choices can directly impact how cognitive resources influence the model. The models&#x02019; main differences lie in (1) the language modality, (2) the linguistic approach they take, and (3) how grammatical knowledge is represented.</p>
<p>As for the different language modalities investigated in the three models, the model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) focuses on sentence interpretation and builds the syntactic representations needed for interpretation. In contrast, the model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>) focuses on sentence production. The model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) again focuses on sentence interpretation but includes a sentence production component in its implementation of perspective taking. So, the selected models show that cognitive models can perform both sentence processing as needed for interpretation and sentence processing as needed for production. As the selected models are merely example implementations of linguistic approaches, this shows how versatile cognitive modeling can be.</p>
<p>A second difference between the three models is that the models all take a different linguistic approach, as Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) used LC parsing based on X-bar theory, Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>) used CCG, and Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) used OT. Although a working cognitive model does not prove the necessity of a particular linguistic approach, it shows its sufficiency: the model of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>), for example, shows that LC parsing is sufficient to account for experimental data on sentence processing. It should be noted that the three linguistic approaches need not be mutually exclusive. For example, it is conceivable that a model processes sentences based on LC parsing and uses OT to interpret ambiguous pronouns (cf. Van Rij, <xref ref-type="bibr" rid="B111">2012</xref>; Vogelzang, <xref ref-type="bibr" rid="B114">2017</xref>). Additionally, it should be noted that all three theories have been treated as approaches that have remained unquestioned, whereas variations of these approaches may be worth while to consider (cf., e.g., Osborne et al., <xref ref-type="bibr" rid="B82">2011</xref>).</p>
<p>A final important difference between the models is how grammatical knowledge is represented. In Lewis and Vasishth&#x02019;s (Lewis and Vasishth, <xref ref-type="bibr" rid="B76">2005</xref>) model, lexical information and syntactic structures are stored in declarative memory, but grammatical rules are incorporated as procedural knowledge in production rules. Therefore, their grammatical rules are not subject to the activation functions associated with the declarative memory but are subject to the time constraints of production rule execution. This is different from the model of Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>), which stores lexical forms as well as syntactic categories as chunks in the declarative memory, and therefore also incorporates the grammatical rules in the declarative memory. The model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) incorporates grammatical rules as chunks in the declarative memory. So, the models incorporate grammatical knowledge in different ways, which has consequences for the influence of general cognitive resources on grammatical knowledge. Specifically, knowledge stored in declarative memory is subject to ACT-R&#x02019;s principles concerning memory activation and retrieval time, whereas knowledge stored in procedural memory is subject to ACT-R&#x02019;s principles concerning production rule execution time.</p>
<p>Although the three models differ in several respects, they also have a number of important features in common. The most important ones that we will discuss are (1) the restrictions placed on the model performance by general cognitive resources, (2) the assumption of a serial processing bottleneck, and (3) the generation of quantitative predictions.</p>
<p>As all models were implemented in ACT-R, the performance of all models is constrained by the same restrictions on cognitive resources. So, although the models focus on different linguistic phenomena and use different representations, they all use, for example, the same functions of declarative memory for the activation of chunks. Furthermore, they all use the same distinction between procedural and declarative memory and incorporate the constraint that information can only be actively used by the model once it is retrieved from declarative memory. Using the same cognitive architecture therefore makes these different models comparable with regard to how the representations are influenced by cognitive resources.</p>
<p>Another constraint within all the models, also imposed by the cognitive architecture, is the serial processing bottleneck (Anderson, <xref ref-type="bibr" rid="B4">2007</xref>). In ACT-R, only one production rule execution or memory retrieval can be performed at a time. Using serial processing increases the time it takes to perform multiple processing steps. Therefore, the serial processing bottleneck creates timing constraints for the models, influencing predictions about performance. We will discuss the implications of this serial processing bottleneck in more detail in the Section &#x0201C;<xref ref-type="sec" rid="S5">Discussion</xref>.&#x0201D;</p>
<p>Finally, the last commonality is that all models can generate quantitative predictions. In general, linguistic theories only discuss competence and do not address performance and do not explain why the observed performance may not match the competence. Thus, linguistic theories do not explain, for example, why speakers may use a certain form in 80% of the cases, but a different form in the other cases. By implementing theoretical approaches in cognitive models, quantitative predictions about why performance does not match competence can be generated.</p>
</sec>
</sec>
<sec id="S5" sec-type="discussion">
<title>Discussion</title>
<p>In this review, we investigated to what extent general cognitive resources influence concretely implemented models of linguistic competence. To this end, we examined the language processing models of Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>), Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>), and Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>). In this section, we will discuss the benefits and limitations of using a cognitive architecture to implement and investigate theories of linguistic competence, and to what extent general cognitive resources influence performance on the basis of these theories.</p>
<p>Cognitive architectures provide a framework for implementing theories of linguistic competence in a validated account of general cognitive resources related to learning and memory. The three specific models that we discussed showed that the cognitive architecture ACT-R on the one hand provides sufficient freedom to implement different linguistic theories in a plausible manner, and on the other hand sufficiently constrains these theories to account for several differences between linguistic competence and performance. Implementing a linguistic theory in a cognitive architecture forces one to specify, among other things, assumptions about how lexical, syntactic, and semantic knowledge is represented and processed in our mind. These specifications are necessarily constrained by general cognitive resources. Therefore, general cognitive resources such as memory and processing speed also constrain performance on the basis of linguistic theories and are crucial for investigating this performance in a cognitively plausible framework.</p>
<p>By implementing a theory of linguistic competence in a cognitive model, it can be evaluated whether a linguistic theory can account for experimental performance data. The distinction between competence and performance is an advantage of cognitive models over abstract linguistic theories (reflecting competence) and standard experimental measures (measuring performance). A cognitive model thus can not only be used to model performance but can also be used to investigate the reason why full competence may not be reached (e.g., because of memory retrieval limitations: Van Maanen and Van Rijn, <xref ref-type="bibr" rid="B110">2010</xref>, processing speed limitations: Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>, or the use of an incorrect strategy: Arslan et al., <xref ref-type="bibr" rid="B6">2017</xref>). As such, cognitive models can account for patterns of linguistic performance that were traditionally accounted for by positing a separate parsing module in the mind specifically for language processing (e.g., Kimball, <xref ref-type="bibr" rid="B68">1973</xref>; Frazier and Fodor, <xref ref-type="bibr" rid="B45">1978</xref>). This line of argumentation has also been explored by Hale (<xref ref-type="bibr" rid="B51">2011</xref>), who argues that linguistic theories need to be specified not just on Marr&#x02019;s computational level, but that it is necessary to specify theories at a level of detail so that they can be implemented, step-by-step, in an algorithmic-level framework and yield precise predictions about behavior. The comparison of models described in this review makes explicit which assumptions have to be made in the cognitive model to incorporate particular linguistic theories. All three cognitive models discussed in this review have been applied to fit human data. In many of these cases, the model could account for the general trends in the data, if not the complete data set. As such, all three models provided an explicit relation between data, theory, and explanation. Although not all models made novel predictions that could be tested in new experiments, this is a strength of cognitive modeling and therefore something every paper on cognitive modeling should include. Adding novel predictions shows that (1) the model was not just fitted to existing data and (2) the model is falsifiable. The latter is important, because falsifiable models allow a theory to be disproven. Providing novel predictions allows other researchers to test these, and gather either support for or evidence against a specific theory.</p>
<p>An additional benefit of cognitive modeling is that individual differences can be investigated. By manipulating, for example, the amount of experience (Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>), the amount of working memory capacity (Van Rij et al., <xref ref-type="bibr" rid="B113">2013</xref>), or the rate of forgetting in memory (Sense et al., <xref ref-type="bibr" rid="B100">2016</xref>), different performance levels can be achieved. This way, different individuals can be modeled and it can be investigated why certain mistakes may be made (explanations could be, for example limited experience, limited memory capacity, limited attention span). By combining different simulated individuals, group effects may be explained (Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>).</p>
<p>There are, however, also some limitations to modeling language processing in a cognitive architecture. First, all three models that were discussed can account for specific linguistic phenomena, but these only form a small part of language. Scalability is an issue for many models, as expanding their coverage and making them more complex (for example, by combining a model that performs full semantic processing with a model that performs full syntactic processing) will make models slower in any architecture that assumes serial processing. Specifically, although the model of Van Rij et al. (<xref ref-type="bibr" rid="B112">2010</xref>) uses the serial processing bottleneck explicitly to account for children&#x02019;s performance errors, both Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) and Reitter et al. (<xref ref-type="bibr" rid="B90">2011</xref>) suggest that their models may struggle with this assumption when expanded. It is thus important to keep in mind that the discussed, relatively small, serially implemented models of language processing were sufficient to fit to experimental data, but the serial processing bottleneck may prove to be too strict for sentence processing when a complete language processing model is developed. Moreover, the discussed models are abstractions and simplifications of reality and take into account neither additional internal factors influencing language processing, such as attentional state or focus (Lappin and Leass, <xref ref-type="bibr" rid="B74">1994</xref>; Lewis et al., <xref ref-type="bibr" rid="B77">2006</xref>), emotion (Belavkin et al., <xref ref-type="bibr" rid="B12">1999</xref>), and motivation (Belavkin, <xref ref-type="bibr" rid="B11">2001</xref>), nor external factors such as visual context (Tanenhaus et al., <xref ref-type="bibr" rid="B108">1995</xref>). Once a model has found support for underlying mechanisms of sentence processing, it can be used as a basis for investigating the effects of these additional factors. Therefore, the models discussed can be seen as a first step toward investigating such factors in the future.</p>
<p>A second limitation is related to a concern that Lewis and Vasishth (<xref ref-type="bibr" rid="B76">2005</xref>) raised: the degrees of freedom in cognitive models. For any set of cognitive models to be optimally comparable, they should be restricted by the same cognitive resources. However, cognitive architectures provide much freedom regarding different parameters (for example, the memory decay parameter in ACT-R can be changed manually). Therefore, models should generally strive to keep the quantitative parameters constant. If this is done, any variation between models will originate from the production rules and the content of the declarative memory, which is also where (linguistic) theory is implemented.</p>
<p>As a final limitation, any cognitive architecture that does not specify different types of memory (short-term memory, episodic long-term memory, semantic long-term memory) will make it difficult to model language processing in all its complexity. For example, long-term memory is difficult to implement in ACT-R, because all chunks are subject to the same decay in activation over time. Thus, it is a puzzle why people do not forget certain pieces of knowledge that are not retrieved frequently (like, for example, what a hedgehog is). Recent research has found that different types of facts may actually have different decay rates (Sense et al., <xref ref-type="bibr" rid="B100">2016</xref>). This can be important for language processing, because even infrequent words are not forgotten and can still be recognized and used after a long time. A related issue is that cognitive architectures with only one type of memory make it challenging to implement and manipulate working memory capacity. So, although the possibility of manipulating cognitive resources in cognitive models can be seen as a benefit, not restricting how these cognitive resources should be modeled limits its application. As language processing is known to be constrained by working memory capacity, manipulations of working memory capacity would be useful in order to study its effects on linguistic performance. Moreover, when modeling language acquisition or language attrition, working memory may be of great influence, as it can differ between ages (Grivol and Hage, <xref ref-type="bibr" rid="B48">2011</xref>) and in clinical populations (e.g., ADHD: Martinussen et al., <xref ref-type="bibr" rid="B79">2005</xref>; autism spectrum disorder: Barendse et al., <xref ref-type="bibr" rid="B10">2013</xref>; cochlear implant users: AuBuchon et al., <xref ref-type="bibr" rid="B7">2015</xref>). Although the function of working memory can be simulated indirectly through other processes like spreading activation (Daily et al., <xref ref-type="bibr" rid="B37">2001</xref>), restrictions on their implementation in the cognitive architecture would make models more comparable and potentially more cognitively plausible.</p>
<p>Thus, using a cognitive architecture to investigate theories of linguistic competence has clear benefits as well as a number of current limitations. The main question in this review was to what extent general cognitive resources influence concretely implemented models of linguistic competence. An examination of the different cognitive models of linguistic performance provides evidence that well-studied general cognitive resources such as working memory influence language processing. In addition, less well-studied cognitive factors may also play a role, such as number of processing steps (Lewis and Vasishth, <xref ref-type="bibr" rid="B76">2005</xref>) and processing efficiency (Van Rij et al., <xref ref-type="bibr" rid="B112">2010</xref>). The influence of these factors can differ due to differences in, for example, experience, processing strategy, or possibly developmental disorder. Thus, our investigation of different cognitive models emphasizes that not only memory-related resources but also other timing-related resources and factors influence language processing.</p>
<p>As stated, implementations of linguistic theories into a cognitive model can, on the one hand, provide information about whether the theory can sufficiently account for observed performance. On the other hand, they can also be used to investigate cognitive processes. For example, the speed of language processing is so high that it may not be met by the time-consuming processing steps provided by a cognitive model (cf. Vogelzang, <xref ref-type="bibr" rid="B114">2017</xref>), or by the same memory processes that underlie other cognitive processes. So, from the viewpoint of linguistics, but also from the viewpoint of cognitive modeling, the puzzle of highly fast and efficient language processing compared to other cognitive processes is an interesting direction for future research.</p>
<p>Overall, cognitively constrained models can be used to investigate whether a linguistic theory can account for specific linguistic data. The interactions between a particular linguistic approach and general cognitive resources can be investigated through such models, which formalize of relation between competence and performance. Additionally, cognitive models can generate quantitative predictions of the basis of theories of linguistic competence. Because of this, cognitive models of linguistic theories are very suitable for investigating the relation between data, theory and experiments. Moreover, the possibility to model differences in cognitive resources allows for the investigation of individual differences in performance, as well as deviating performance due to aging or developmental disorders. In some cases, the high efficiency of language processing is currently not met by some of the constraining assumptions about cognitive resources. In this sense, cognitive models of language processing can also be used to investigate human cognition, for example in which ways currently adopted cognitive assumptions fail to meet the requirements for language processing. In conclusion, investigating specific linguistic phenomena through cognitive modeling can provide new insights that can complement findings from standard experimental techniques.</p>
</sec>
<sec id="S6" sec-type="author-contributor">
<title>Author Contributions</title>
<p>All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.</p>
</sec>
<sec id="S7">
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<ack>
<p>The authors would like to thank Michael Putnam for his comments on an earlier draft of the paper.</p>
</ack>
<fn-group>
<fn fn-type="financial-disclosure">
<p><bold>Funding.</bold> This research was supported by a grant from the Netherlands Organization for Scientific Research (NWO) awarded to Jacolien Van Rij (grant no. 275-70-044). David Reitter acknowledges support from the National Science Foundation grant BCS-1734304.</p></fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Aho</surname> <given-names>A. V.</given-names></name> <name><surname>Ullman</surname> <given-names>J. D.</given-names></name></person-group> (<year>1972</year>). <source>The Theory of Parsing, Translation, and Compiling</source>. <publisher-loc>Upper Saddle River, NJ</publisher-loc>: <publisher-name>Prentice-Hall, Inc</publisher-name>.</citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Allen</surname> <given-names>R.</given-names></name> <name><surname>Hulme</surname> <given-names>C.</given-names></name></person-group> (<year>2006</year>). <article-title>Speech and language processing mechanisms in verbal serial recall</article-title>. <source>J. Mem. Lang.</source> <volume>55</volume>, <fpage>64</fpage>&#x02013;<lpage>88</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2006.02.002</pub-id></citation></ref>
<ref id="B3"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name></person-group> (<year>1993</year>). <source>Rules of the Mind</source>. <publisher-loc>Hillsdale, NJ</publisher-loc>: <publisher-name>Erlbaum</publisher-name>.</citation></ref>
<ref id="B4"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name></person-group> (<year>2007</year>). <source>How Can the Human Mind Occur in the Physical Universe?</source> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name> <name><surname>Bothell</surname> <given-names>D.</given-names></name> <name><surname>Byrne</surname> <given-names>M. D.</given-names></name> <name><surname>Douglass</surname> <given-names>S.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name> <name><surname>Qin</surname> <given-names>Y.</given-names></name></person-group> (<year>2004</year>). <article-title>An integrated theory of mind</article-title>. <source>Psychol. Rev.</source> <volume>111</volume>, <fpage>1036</fpage>&#x02013;<lpage>1060</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.111.4.1036</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Arslan</surname> <given-names>B.</given-names></name> <name><surname>Taatgen</surname> <given-names>N. A.</given-names></name> <name><surname>Verbrugge</surname> <given-names>R.</given-names></name></person-group> (<year>2017</year>). <article-title>Five-year-olds&#x02019; systematic errors in second-order false belief tasks are due to first-order theory of mind strategy selection: a computational modeling study</article-title>. <source>Front. Psychol.</source> <volume>8</volume>:<fpage>275</fpage>.<pub-id pub-id-type="doi">10.3389/fpsyg.2017.00275</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>AuBuchon</surname> <given-names>A. M.</given-names></name> <name><surname>Pisoni</surname> <given-names>D. B.</given-names></name> <name><surname>Kronenberger</surname> <given-names>W. G.</given-names></name></person-group> (<year>2015</year>). <article-title>Short-term and working memory impairments in early-implanted, long-term cochlear implant users are independent of audibility and speech production</article-title>. <source>Ear Hear.</source> <volume>36</volume>, <fpage>733</fpage>&#x02013;<lpage>737</lpage>.<pub-id pub-id-type="doi">10.1097/AUD.0000000000000189</pub-id></citation></ref>
<ref id="B8"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Baauw</surname> <given-names>S.</given-names></name></person-group> (<year>2002</year>). <source>Grammatical Features and the Acquisition of Reference. A Comparative Study of Dutch and Spanish</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Routledge</publisher-name>.</citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Baayen</surname> <given-names>R. H.</given-names></name> <name><surname>Dijkstra</surname> <given-names>T.</given-names></name> <name><surname>Schreuder</surname> <given-names>R.</given-names></name></person-group> (<year>1997</year>). <article-title>Singulars and plurals in Dutch: evidence for a parallel dual-route model</article-title>. <source>J. Mem. Lang.</source> <volume>37</volume>, <fpage>94</fpage>&#x02013;<lpage>117</lpage>.<pub-id pub-id-type="doi">10.1006/jmla.1997.2509</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barendse</surname> <given-names>E. M.</given-names></name> <name><surname>Hendriks</surname> <given-names>M. P.</given-names></name> <name><surname>Jansen</surname> <given-names>J. F.</given-names></name> <name><surname>Backes</surname> <given-names>W. H.</given-names></name> <name><surname>Hofman</surname> <given-names>P. A.</given-names></name> <name><surname>Thoonen</surname> <given-names>G.</given-names></name> <etal/></person-group> (<year>2013</year>). <article-title>Working memory deficits in high-functioning adolescents with autism spectrum disorders: neuropsychological and neuroimaging correlates</article-title>. <source>J. Neurodev. Disord.</source> <volume>5</volume>, <fpage>14</fpage>.<pub-id pub-id-type="doi">10.1186/1866-1955-5-14</pub-id><pub-id pub-id-type="pmid">23731472</pub-id></citation></ref>
<ref id="B11"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Belavkin</surname> <given-names>R. V.</given-names></name></person-group> (<year>2001</year>). <article-title>&#x0201C;Modelling the inverted-U effect in ACT-R,&#x0201D;</article-title> in <conf-name>Proceedings of the 2001 Fourth International Conference on Cognitive Modelling</conf-name>, eds <person-group person-group-type="editor"><name><surname>Altmann</surname> <given-names>E. M.</given-names></name> <name><surname>Cleeremans</surname> <given-names>A.</given-names></name> <name><surname>Schunn</surname> <given-names>C. D.</given-names></name> <name><surname>Gray</surname> <given-names>W. D.</given-names></name></person-group> (<conf-loc>Mahwah, NJ, London</conf-loc>: <conf-sponsor>Lawrence Erlbaum</conf-sponsor>), <fpage>275</fpage>&#x02013;<lpage>276</lpage>.</citation></ref>
<ref id="B12"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Belavkin</surname> <given-names>R. V.</given-names></name> <name><surname>Ritter</surname> <given-names>F. E.</given-names></name> <name><surname>Elliman</surname> <given-names>D. G.</given-names></name></person-group> (<year>1999</year>). <article-title>&#x0201C;Towards including simple emotions in a cognitive architecture in order to fit children&#x02019;s behaviour better,&#x0201D;</article-title> in <conf-name>Proceedings of the 1999 Conference of the Cognitive Science Society (CogSci)</conf-name> (<conf-loc>Mahwah, NJ</conf-loc>: <conf-sponsor>Erlbaum</conf-sponsor>).</citation></ref>
<ref id="B13"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Belletti</surname> <given-names>A.</given-names></name> <name><surname>Guasti</surname> <given-names>M. T.</given-names></name></person-group> (<year>2015</year>). <source>The Acquisition of Italian: Morphosyntax and Its Interfaces in Different Modes of Acquisition</source>. <publisher-loc>Amsterdam, PA</publisher-loc>: <publisher-name>John Benjamins Publishing Company</publisher-name>.</citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bock</surname> <given-names>J. K.</given-names></name></person-group> (<year>1986</year>). <article-title>Syntactic persistence in language production</article-title>. <source>Cogn. Psychol.</source> <volume>18</volume>, <fpage>355</fpage>&#x02013;<lpage>387</lpage>.<pub-id pub-id-type="doi">10.1016/0010-0285(86)90004-6</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bock</surname> <given-names>J. K.</given-names></name> <name><surname>Dell</surname> <given-names>G. S.</given-names></name> <name><surname>Chang</surname> <given-names>F.</given-names></name> <name><surname>Onishi</surname> <given-names>K. H.</given-names></name></person-group> (<year>2007</year>). <article-title>Persistent structural priming from language comprehension to language production</article-title>. <source>Cognition</source> <volume>104</volume>, <fpage>437</fpage>&#x02013;<lpage>458</lpage>.<pub-id pub-id-type="doi">10.1016/j.cognition.2006.07.003</pub-id><pub-id pub-id-type="pmid">16973143</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bock</surname> <given-names>J. K.</given-names></name> <name><surname>Griffin</surname> <given-names>Z.</given-names></name></person-group> (<year>2000</year>). <article-title>The persistence of structural priming: transient activation or implicit learning?</article-title> <source>J. Exp. Psychol. Gen.</source> <volume>129</volume>, <fpage>177</fpage>&#x02013;<lpage>192</lpage>.<pub-id pub-id-type="doi">10.1037/0096-3445.129.2.177</pub-id><pub-id pub-id-type="pmid">10868333</pub-id></citation></ref>
<ref id="B17"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Bock</surname> <given-names>J. K.</given-names></name> <name><surname>Kroch</surname> <given-names>A. S.</given-names></name></person-group> (<year>1989</year>). <article-title>&#x0201C;The isolability of syntactic processing,&#x0201D;</article-title> in <source>Linguistic Structure in Language Processing</source>, eds <person-group person-group-type="editor"><name><surname>Carlson</surname> <given-names>G. N.</given-names></name> <name><surname>Tanenhaus</surname> <given-names>M. K.</given-names></name></person-group> (<publisher-loc>Dordrecht, The Netherlands</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>157</fpage>&#x02013;<lpage>196</lpage>.</citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Borst</surname> <given-names>J. P.</given-names></name> <name><surname>Taatgen</surname> <given-names>N. A.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name></person-group> (<year>2010</year>). <article-title>The problem state: a cognitive bottleneck in multitasking</article-title>. <source>J. Exp. Psychol. Learn. Mem. Cogn.</source> <volume>36</volume>, <fpage>363</fpage>&#x02013;<lpage>382</lpage>.<pub-id pub-id-type="doi">10.1037/a0018106</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Branigan</surname> <given-names>H. P.</given-names></name> <name><surname>Pickering</surname> <given-names>M. J.</given-names></name> <name><surname>Cleland</surname> <given-names>A. A.</given-names></name></person-group> (<year>1999</year>). <article-title>Syntactic priming in language production: evidence for rapid decay</article-title>. <source>Psychon. Bull. Rev.</source> <volume>6</volume>, <fpage>635</fpage>&#x02013;<lpage>640</lpage>.<pub-id pub-id-type="doi">10.3758/BF03212972</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Branigan</surname> <given-names>H. P.</given-names></name> <name><surname>Pickering</surname> <given-names>M. J.</given-names></name> <name><surname>Stewart</surname> <given-names>A. J.</given-names></name> <name><surname>McLean</surname> <given-names>J. F.</given-names></name></person-group> (<year>2000</year>). <article-title>Syntactic priming in spoken production: linguistic and temporal interference</article-title>. <source>Mem. Cognit.</source> <volume>28</volume>, <fpage>1297</fpage>&#x02013;<lpage>1302</lpage>.<pub-id pub-id-type="doi">10.3758/BF03211830</pub-id><pub-id pub-id-type="pmid">11219957</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brasoveanu</surname> <given-names>A.</given-names></name> <name><surname>Dotla&#x0010D;il</surname> <given-names>J.</given-names></name></person-group> (<year>2015</year>). <article-title>Incremental and predictive interpretation: experimental evidence and possible accounts</article-title>. <source>Proc. SALT</source> <volume>25</volume>, <fpage>57</fpage>&#x02013;<lpage>81</lpage>.<pub-id pub-id-type="doi">10.3765/salt.v25i0.3047</pub-id></citation></ref>
<ref id="B22"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Bresnan</surname> <given-names>J.</given-names></name></person-group> (<year>2001</year>). <article-title>&#x0201C;The emergence of the unmarked pronoun,&#x0201D;</article-title> in <source>Optimality-Theoretic Syntax</source>, eds <person-group person-group-type="editor"><name><surname>Legendre</surname> <given-names>G.</given-names></name> <name><surname>Grimshaw</surname> <given-names>J.</given-names></name> <name><surname>Vikner</surname> <given-names>S.</given-names></name></person-group> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>), <fpage>113</fpage>&#x02013;<lpage>142</lpage>.</citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Budiu</surname> <given-names>R.</given-names></name> <name><surname>Anderson</surname> <given-names>J. R.</given-names></name></person-group> (<year>2002</year>). <article-title>Comprehending anaphoric metaphors</article-title>. <source>Mem. Cognit.</source> <volume>30</volume>, <fpage>158</fpage>&#x02013;<lpage>165</lpage>.<pub-id pub-id-type="doi">10.3758/BF03195275</pub-id><pub-id pub-id-type="pmid">11958349</pub-id></citation></ref>
<ref id="B24"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Bybee</surname> <given-names>J.</given-names></name> <name><surname>Beckner</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;Usage-based theory,&#x0201D;</article-title> in <source>The Oxford Handbook of Linguistic Analysis</source>, eds <person-group person-group-type="editor"><name><surname>Heine</surname> <given-names>B.</given-names></name> <name><surname>Narrog</surname> <given-names>H.</given-names></name></person-group> (<publisher-loc>Oxford</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>), <fpage>827</fpage>&#x02013;<lpage>855</lpage>.</citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname> <given-names>F.</given-names></name> <name><surname>Bock</surname> <given-names>J. K.</given-names></name> <name><surname>Goldberg</surname> <given-names>A. E.</given-names></name></person-group> (<year>2003</year>). <article-title>Can thematic roles leave traces of their places?</article-title> <source>Cognition</source> <volume>99</volume>, <fpage>29</fpage>&#x02013;<lpage>49</lpage>.<pub-id pub-id-type="doi">10.1016/S0010-0277(03)00123-9</pub-id><pub-id pub-id-type="pmid">14597269</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname> <given-names>F.</given-names></name> <name><surname>Dell</surname> <given-names>G. S.</given-names></name> <name><surname>Bock</surname> <given-names>K.</given-names></name></person-group> (<year>2006</year>). <article-title>Becoming syntactic</article-title>. <source>Psychol. Rev.</source> <volume>113</volume>, <fpage>234</fpage>&#x02013;<lpage>272</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.113.2.234</pub-id><pub-id pub-id-type="pmid">16637761</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chesi</surname> <given-names>C.</given-names></name></person-group> (<year>2015</year>). <article-title>On directionality of phrase structure building</article-title>. <source>J. Psycholinguist. Res.</source> <volume>44</volume>, <fpage>65</fpage>&#x02013;<lpage>89</lpage>.<pub-id pub-id-type="doi">10.1007/s10936-014-9330-6</pub-id><pub-id pub-id-type="pmid">25408515</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chien</surname> <given-names>Y.-C.</given-names></name> <name><surname>Wexler</surname> <given-names>K.</given-names></name></person-group> (<year>1990</year>). <article-title>Children&#x02019;s knowledge of locality conditions in binding as evidence for the modularity of syntax and pragmatics</article-title>. <source>Lang. Acquis.</source> <volume>1</volume>, <fpage>225</fpage>&#x02013;<lpage>295</lpage>.<pub-id pub-id-type="doi">10.1207/s15327817la0103_2</pub-id></citation></ref>
<ref id="B29"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name></person-group> (<year>1965</year>). <source>Aspects of the Theory of Syntax</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B30"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name></person-group> (<year>1970</year>). <article-title>&#x0201C;Remarks on nominalization,&#x0201D;</article-title> in <source>Reading in English Transformational Grammar</source>, eds <person-group person-group-type="editor"><name><surname>Jacobs</surname> <given-names>R.</given-names></name> <name><surname>Rosenbaum</surname> <given-names>P.</given-names></name></person-group> (<publisher-loc>Waltham, MA</publisher-loc>: <publisher-name>Ginn</publisher-name>), <fpage>184</fpage>&#x02013;<lpage>221</lpage>.</citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name></person-group> (<year>1980</year>). <article-title>Rules and representations</article-title>. <source>Behav. Brain Sci.</source> <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>15</lpage>.<pub-id pub-id-type="doi">10.1017/S0140525X00001515</pub-id></citation></ref>
<ref id="B32"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name></person-group> (<year>1981</year>). <source>Lectures on Government and Binding: The Pisa Lectures</source>. <publisher-loc>Dordrecht, The Netherlands</publisher-loc>: <publisher-name>Foris Publications</publisher-name>.</citation></ref>
<ref id="B33"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name></person-group> (<year>1993</year>). <article-title>&#x0201C;A minimalist program for linguistic theory,&#x0201D;</article-title> in <source>Tile View from Building 20: Essays in Linguistics in Honor of Sylvain Bromberger</source>, eds <person-group person-group-type="editor"><name><surname>Hale</surname> <given-names>K.</given-names></name> <name><surname>Keyser</surname> <given-names>S. J.</given-names></name></person-group> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>52</lpage>.</citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Christiansen</surname> <given-names>M.</given-names></name> <name><surname>Chater</surname> <given-names>N.</given-names></name></person-group> (<year>2016</year>). <article-title>The now-or-never bottleneck: a fundamental constraint on language</article-title>. <source>Brain Behav. Sci.</source> <volume>39</volume>, <fpage>e62</fpage>.<pub-id pub-id-type="doi">10.1017/S0140525X1500031X</pub-id><pub-id pub-id-type="pmid">25869618</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cleland</surname> <given-names>A. A.</given-names></name> <name><surname>Pickering</surname> <given-names>M. J.</given-names></name></person-group> (<year>2003</year>). <article-title>The use of lexical and syntactic information in language production: evidence from the priming of noun-phrase structure</article-title>. <source>J. Mem. Lang.</source> <volume>49</volume>, <fpage>214</fpage>&#x02013;<lpage>230</lpage>.<pub-id pub-id-type="doi">10.1016/S0749-596X(03)00060-3</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Corley</surname> <given-names>M.</given-names></name> <name><surname>Scheepers</surname> <given-names>C.</given-names></name></person-group> (<year>2002</year>). <article-title>Syntactic priming in English sentence production: categorical and latency evidence from an internet-based study</article-title>. <source>Psychon. Bull. Rev.</source> <volume>9</volume>, <fpage>126</fpage>&#x02013;<lpage>131</lpage>.<pub-id pub-id-type="doi">10.3758/BF03196267</pub-id><pub-id pub-id-type="pmid">12026944</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Daily</surname> <given-names>L. Z.</given-names></name> <name><surname>Lovett</surname> <given-names>M. C.</given-names></name> <name><surname>Reder</surname> <given-names>L. M.</given-names></name></person-group> (<year>2001</year>). <article-title>Modeling individual differences in working memory performance: a source activation account</article-title>. <source>Cogn. Sci.</source> <volume>25</volume>, <fpage>315</fpage>&#x02013;<lpage>353</lpage>.<pub-id pub-id-type="doi">10.1016/S0364-0213(01)00039-8</pub-id><pub-id pub-id-type="pmid">19079561</pub-id></citation></ref>
<ref id="B38"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>De Villiers</surname> <given-names>J.</given-names></name> <name><surname>Cahillane</surname> <given-names>J.</given-names></name> <name><surname>Altreuter</surname> <given-names>E.</given-names></name></person-group> (<year>2006</year>). <article-title>&#x0201C;What can production reveal about principle B?</article-title>&#x0201D; in <conf-name>Proceedings of the Inaugural Conference on Generative Approaches to Language Acquisition &#x02013; North America</conf-name>, Vol. <volume>1</volume>, eds <person-group person-group-type="editor"><name><surname>Deen</surname> <given-names>K.</given-names></name> <name><surname>Nomura</surname> <given-names>J.</given-names></name> <name><surname>Schulz</surname> <given-names>B.</given-names></name> <name><surname>Schwartz</surname> <given-names>B.</given-names></name></person-group> (<conf-loc>Honolulu, HI</conf-loc>: <conf-sponsor>University of Connecticut</conf-sponsor>), <fpage>89</fpage>&#x02013;<lpage>100</lpage>.</citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deese</surname> <given-names>J.</given-names></name> <name><surname>Kaufman</surname> <given-names>R. A.</given-names></name></person-group> (<year>1957</year>). <article-title>Serial effects in recall of unorganized and sequentially organized verbal material</article-title>. <source>J. Exp. Psychol.</source> <volume>54</volume>, <fpage>180</fpage>&#x02013;<lpage>187</lpage>.<pub-id pub-id-type="doi">10.1037/h0040536</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Engelmann</surname> <given-names>F.</given-names></name> <name><surname>Vasishth</surname> <given-names>S.</given-names></name> <name><surname>Engbert</surname> <given-names>R.</given-names></name> <name><surname>Kliegl</surname> <given-names>R.</given-names></name></person-group> (<year>2013</year>). <article-title>A framework for modeling the interaction of syntactic processing and eye movement control</article-title>. <source>Top. Cogn. Sci.</source> <volume>5</volume>, <fpage>452</fpage>&#x02013;<lpage>474</lpage>.<pub-id pub-id-type="doi">10.1111/tops.12026</pub-id><pub-id pub-id-type="pmid">23681560</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fedorenko</surname> <given-names>E.</given-names></name> <name><surname>Woodbury</surname> <given-names>R.</given-names></name> <name><surname>Gibson</surname> <given-names>E.</given-names></name></person-group> (<year>2013</year>). <article-title>Direct evidence of memory retrieval as a source of difficulty in non-local dependencies in language</article-title>. <source>Cogn. Sci.</source> <volume>37</volume>, <fpage>378</fpage>&#x02013;<lpage>394</lpage>.<pub-id pub-id-type="doi">10.1111/cogs.12021</pub-id><pub-id pub-id-type="pmid">23362990</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ferreira</surname> <given-names>V. S.</given-names></name></person-group> (<year>2003</year>). <article-title>The persistence of optional complementizer production: why saying &#x0201C;that&#x0201D; is not saying &#x0201C;that&#x0201D; at all</article-title>. <source>J. Mem. Lang.</source> <volume>48</volume>, <fpage>379</fpage>&#x02013;<lpage>398</lpage>.<pub-id pub-id-type="doi">10.1016/s0749-596x(02)00523-5</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fillmore</surname> <given-names>C. J.</given-names></name> <name><surname>Kay</surname> <given-names>P.</given-names></name> <name><surname>O&#x02019;Connor</surname> <given-names>M. C.</given-names></name></person-group> (<year>1988</year>). <article-title>Regularity and idiomaticity in grammatical constructions: the case of let alone</article-title>. <source>Language</source> <volume>64</volume>, <fpage>501</fpage>&#x02013;<lpage>538</lpage>.<pub-id pub-id-type="doi">10.2307/414531</pub-id></citation></ref>
<ref id="B44"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Fodor</surname> <given-names>J. A.</given-names></name></person-group> (<year>1983</year>). <source>Modularity of Mind: An Essay on Faculty Psychology</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Frazier</surname> <given-names>L.</given-names></name> <name><surname>Fodor</surname> <given-names>J. D.</given-names></name></person-group> (<year>1978</year>). <article-title>The sausage machine: a new two-stage parsing model</article-title>. <source>Cognition</source> <volume>6</volume>, <fpage>291</fpage>&#x02013;<lpage>325</lpage>.<pub-id pub-id-type="doi">10.1016/0010-0277(78)90002-1</pub-id></citation></ref>
<ref id="B46"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Fromkin</surname> <given-names>V. A.</given-names></name></person-group> (<year>2000</year>). <source>Linguistics: An Introduction to Linguistic Theory</source>. <publisher-loc>Hoboken, NJ</publisher-loc>: <publisher-name>Wiley-Blackwell Publishing</publisher-name>.</citation></ref>
<ref id="B47"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Gibson</surname> <given-names>E.</given-names></name></person-group> (<year>2000</year>). <article-title>&#x0201C;The dependency locality theory: a distance-based theory of linguistic complexity,&#x0201D;</article-title> in <source>Image, Language, Brain</source>, eds <person-group person-group-type="editor"><name><surname>Marantz</surname> <given-names>A.</given-names></name> <name><surname>Miyashita</surname> <given-names>Y.</given-names></name> <name><surname>O&#x02019;Neil</surname> <given-names>W.</given-names></name></person-group> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>), <fpage>95</fpage>&#x02013;<lpage>126</lpage>.</citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grivol</surname> <given-names>M. A.</given-names></name> <name><surname>Hage</surname> <given-names>S. R. d. V.</given-names></name></person-group> (<year>2011</year>). <article-title>Phonological working memory: a comparative study between different age groups</article-title>. <source>J. Soc. Bras. Fonoaudiol.</source> <volume>23</volume>, <fpage>245</fpage>&#x02013;<lpage>251</lpage>.<pub-id pub-id-type="doi">10.1590/S2179-64912011000300010</pub-id><pub-id pub-id-type="pmid">22012159</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grodner</surname> <given-names>D.</given-names></name> <name><surname>Gibson</surname> <given-names>E.</given-names></name></person-group> (<year>2005</year>). <article-title>Consequences of the serial nature of linguistic input for sentenial complexity</article-title>. <source>Cogn. Sci.</source> <volume>29</volume>, <fpage>261</fpage>&#x02013;<lpage>290</lpage>.<pub-id pub-id-type="doi">10.1207/s15516709cog0000_7</pub-id><pub-id pub-id-type="pmid">21702774</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grosz</surname> <given-names>B. J.</given-names></name> <name><surname>Weinstein</surname> <given-names>S.</given-names></name> <name><surname>Joshi</surname> <given-names>A. K.</given-names></name></person-group> (<year>1995</year>). <article-title>Centering: a framework for modeling the local coherence of discourse</article-title>. <source>Comput. Linguist.</source> <volume>21</volume>, <fpage>203</fpage>&#x02013;<lpage>225</lpage>.</citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hale</surname> <given-names>J. T.</given-names></name></person-group> (<year>2011</year>). <article-title>What a rational parser would do</article-title>. <source>Cogn. Sci.</source> <volume>35</volume>, <fpage>399</fpage>&#x02013;<lpage>443</lpage>.<pub-id pub-id-type="doi">10.1111/j.1551-6709.2010.01145.x</pub-id></citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hartsuiker</surname> <given-names>R. J.</given-names></name> <name><surname>Bernolet</surname> <given-names>S.</given-names></name> <name><surname>Schoonbaert</surname> <given-names>S.</given-names></name> <name><surname>Speybroeck</surname> <given-names>S.</given-names></name> <name><surname>Vanderelst</surname> <given-names>D.</given-names></name></person-group> (<year>2008</year>). <article-title>Syntactic priming persists while the lexical boost decays: evidence from written and spoken dialogue</article-title>. <source>J. Mem. Lang.</source> <volume>58</volume>, <fpage>214</fpage>&#x02013;<lpage>238</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2007.07.003</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hartsuiker</surname> <given-names>R. J.</given-names></name> <name><surname>Kolk</surname> <given-names>H. H. J.</given-names></name></person-group> (<year>1998</year>). <article-title>Syntactic persistence in Dutch</article-title>. <source>Lang. Speech</source> <volume>41</volume>, <fpage>143</fpage>&#x02013;<lpage>184</lpage>.<pub-id pub-id-type="doi">10.1177/002383099804100202</pub-id><pub-id pub-id-type="pmid">10194875</pub-id></citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hartsuiker</surname> <given-names>R. J.</given-names></name> <name><surname>Westenberg</surname> <given-names>C.</given-names></name></person-group> (<year>2000</year>). <article-title>Word order priming in written and spoken sentence production</article-title>. <source>Cognition</source> <volume>75B</volume>, <fpage>27</fpage>&#x02013;<lpage>39</lpage>.<pub-id pub-id-type="doi">10.1016/S0010-0277(99)00080-3</pub-id><pub-id pub-id-type="pmid">10771278</pub-id></citation></ref>
<ref id="B55"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Hendriks</surname> <given-names>P.</given-names></name></person-group> (<year>2014</year>). <source>Asymmetries between Language Production and Comprehension</source>. <publisher-loc>Dordrecht, The Netherlands</publisher-loc>: <publisher-name>Springer</publisher-name>.</citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hendriks</surname> <given-names>P.</given-names></name> <name><surname>Spenader</surname> <given-names>J.</given-names></name></person-group> (<year>2006</year>). <article-title>When production precedes comprehension: an optimization approach to the acquisition of pronouns</article-title>. <source>Lang. Acquis.</source> <volume>13</volume>, <fpage>319</fpage>&#x02013;<lpage>348</lpage>.<pub-id pub-id-type="doi">10.1207/s15327817la1304_3</pub-id></citation></ref>
<ref id="B57"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hendriks</surname> <given-names>P.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name> <name><surname>Valkenier</surname> <given-names>B.</given-names></name></person-group> (<year>2007</year>). <article-title>Learning to reason about speakers&#x02019; alternatives in sentence comprehension: a computational account</article-title>. <source>Lingua</source> <volume>117</volume>, <fpage>1879</fpage>&#x02013;<lpage>1896</lpage>.<pub-id pub-id-type="doi">10.1016/j.lingua.2006.11.008</pub-id></citation></ref>
<ref id="B58"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huettig</surname> <given-names>F.</given-names></name> <name><surname>Janse</surname> <given-names>E.</given-names></name></person-group> (<year>2016</year>). <article-title>Individual differences in working memory and processing speed predict anticipatory spoken language processing in the visual world</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>31</volume>, <fpage>80</fpage>&#x02013;<lpage>93</lpage>.<pub-id pub-id-type="doi">10.1080/23273798.2015.1047459</pub-id></citation></ref>
<ref id="B59"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huttenlocher</surname> <given-names>J.</given-names></name> <name><surname>Vasilyeva</surname> <given-names>M.</given-names></name> <name><surname>Shimpi</surname> <given-names>P.</given-names></name></person-group> (<year>2004</year>). <article-title>Syntactic priming in young children</article-title>. <source>J. Mem. Lang.</source> <volume>50</volume>, <fpage>182</fpage>&#x02013;<lpage>195</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2003.09.003</pub-id></citation></ref>
<ref id="B60"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>T. F.</given-names></name> <name><surname>Snider</surname> <given-names>N. E.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Implicit learning and syntactic persistence: surprisal and cumulativity,&#x0201D;</article-title> in <conf-name>Proceedings of the 30th Annual Conference of the Cognitive Science Society (CogSci)</conf-name> (<conf-loc>Washington, DC</conf-loc>: <conf-sponsor>Cognitive Science Society</conf-sponsor>), <fpage>1061</fpage>&#x02013;<lpage>1066</lpage>.</citation></ref>
<ref id="B61"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>T. F.</given-names></name> <name><surname>Snider</surname> <given-names>N. E.</given-names></name></person-group> (<year>2013</year>). <article-title>Alignment as a consequence of expectation adaptation: syntactic priming is affected by the prime&#x02019;s prediction error given both prior and recent experience</article-title>. <source>Cognition</source> <volume>127</volume>, <fpage>57</fpage>&#x02013;<lpage>83</lpage>.<pub-id pub-id-type="doi">10.1016/j.cognition.2012.10.013</pub-id></citation></ref>
<ref id="B62"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>J&#x000E4;ger</surname> <given-names>L. A.</given-names></name> <name><surname>Engelmann</surname> <given-names>F.</given-names></name> <name><surname>Vasishth</surname> <given-names>S.</given-names></name></person-group> (<year>2015</year>). <article-title>Retrieval interference in reflexive processing: experimental evidence from Mandarin, and computational modeling</article-title>. <source>Front. Psychol.</source> <volume>6</volume>:<fpage>617</fpage>.<pub-id pub-id-type="doi">10.3389/fpsyg.2015.00617</pub-id></citation></ref>
<ref id="B63"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Joshi</surname> <given-names>A. K.</given-names></name> <name><surname>Levy</surname> <given-names>L. S.</given-names></name> <name><surname>Takahashi</surname> <given-names>M.</given-names></name></person-group> (<year>1975</year>). <article-title>Tree adjunct grammars</article-title>. <source>J. Comput. Syst. Sci.</source> <volume>10</volume>, <fpage>136</fpage>&#x02013;<lpage>163</lpage>.<pub-id pub-id-type="doi">10.1016/S0022-0000(75)80019-5</pub-id></citation></ref>
<ref id="B64"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kaschak</surname> <given-names>M. P.</given-names></name> <name><surname>Kutta</surname> <given-names>T. J.</given-names></name> <name><surname>Jones</surname> <given-names>J. L.</given-names></name></person-group> (<year>2011a</year>). <article-title>Structural priming as implicit learning: cumulative priming effects and individual differences</article-title>. <source>Psychon. Bull. Rev.</source> <volume>18</volume>, <fpage>1133</fpage>&#x02013;<lpage>1139</lpage>.<pub-id pub-id-type="doi">10.3758/s13423-011-0157-y</pub-id><pub-id pub-id-type="pmid">21913001</pub-id></citation></ref>
<ref id="B65"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kaschak</surname> <given-names>M. P.</given-names></name> <name><surname>Kutta</surname> <given-names>T. J.</given-names></name> <name><surname>Schatschneider</surname> <given-names>C.</given-names></name></person-group> (<year>2011b</year>). <article-title>Long-term cumulative structural priming persists for (at least) one week</article-title>. <source>Mem. Cognit.</source> <volume>39</volume>, <fpage>381</fpage>&#x02013;<lpage>388</lpage>.<pub-id pub-id-type="doi">10.3758/s13421-010-0042-3</pub-id><pub-id pub-id-type="pmid">21264596</pub-id></citation></ref>
<ref id="B66"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kempen</surname> <given-names>G.</given-names></name> <name><surname>Hoenkamp</surname> <given-names>E.</given-names></name></person-group> (<year>1987</year>). <article-title>An incremental procedural grammar for sentence formulation</article-title>. <source>Cogn. Sci.</source> <volume>11</volume>, <fpage>201</fpage>&#x02013;<lpage>258</lpage>.<pub-id pub-id-type="doi">10.1207/s15516709cog1102_5</pub-id></citation></ref>
<ref id="B67"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kieras</surname> <given-names>D. E.</given-names></name> <name><surname>Meyer</surname> <given-names>D. E.</given-names></name></person-group> (<year>1997</year>). <article-title>An overview of the EPIC architecture for cognition and performance with application to human-computer interaction</article-title>. <source>Hum. Comput. Interact.</source> <volume>12</volume>, <fpage>391</fpage>&#x02013;<lpage>438</lpage>.<pub-id pub-id-type="doi">10.1207/s15327051hci1204_4</pub-id></citation></ref>
<ref id="B68"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kimball</surname> <given-names>J.</given-names></name></person-group> (<year>1973</year>). <article-title>Seven principles of surface structure parsing in natural language</article-title>. <source>Cognition</source> <volume>2</volume>, <fpage>15</fpage>&#x02013;<lpage>47</lpage>.<pub-id pub-id-type="doi">10.1016/0010-0277(72)90028-5</pub-id></citation></ref>
<ref id="B69"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>King</surname> <given-names>J.</given-names></name> <name><surname>Just</surname> <given-names>M. A.</given-names></name></person-group> (<year>1991</year>). <article-title>Individual differences in syntactic processing: the role of working memory</article-title>. <source>J. Mem. Lang.</source> <volume>30</volume>, <fpage>580</fpage>&#x02013;<lpage>602</lpage>.<pub-id pub-id-type="doi">10.1016/0749-596X(91)90027-H</pub-id></citation></ref>
<ref id="B70"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Kong</surname> <given-names>F.</given-names></name> <name><surname>Zhou</surname> <given-names>G.</given-names></name> <name><surname>Zhu</surname> <given-names>Q.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;Employing the centering theory in pronoun resolution from the semantic perspective,&#x0201D;</article-title> in <conf-name>Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing</conf-name> (<conf-loc>Singapore</conf-loc>: <conf-sponsor>Association for Computational Linguistics</conf-sponsor>), <fpage>987</fpage>&#x02013;<lpage>996</lpage>.</citation></ref>
<ref id="B71"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kootstra</surname> <given-names>G. J.</given-names></name> <name><surname>Van Hell</surname> <given-names>J. G.</given-names></name> <name><surname>Dijkstra</surname> <given-names>T.</given-names></name></person-group> (<year>2012</year>). <article-title>Priming of code-switches in sentences: the role of lexical repetition, cognates, and language proficiency</article-title>. <source>Bilingualism Lang. Cogn.</source> <volume>15</volume>, <fpage>797</fpage>&#x02013;<lpage>819</lpage>.<pub-id pub-id-type="doi">10.1017/S136672891100068X</pub-id></citation></ref>
<ref id="B72"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Kuijper</surname> <given-names>S. J. M.</given-names></name></person-group> (<year>2016</year>). <source>Communication Abilities of Children with ASD and ADHD</source>. <publisher-name>Doctoral dissertation, University of Groningen</publisher-name>, <publisher-loc>Groningen, The Netherlands</publisher-loc>.</citation></ref>
<ref id="B73"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kuijper</surname> <given-names>S. J. M.</given-names></name> <name><surname>Hartman</surname> <given-names>C. A.</given-names></name> <name><surname>Hendriks</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>Who is he? Children with ASD and ADHD take the listener into account in their production of ambiguous pronouns</article-title>. <source>PLoS ONE</source> <volume>10</volume>:<fpage>e0132408</fpage>.<pub-id pub-id-type="doi">10.1371/journal.pone.0132408</pub-id></citation></ref>
<ref id="B74"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lappin</surname> <given-names>S.</given-names></name> <name><surname>Leass</surname> <given-names>H. J.</given-names></name></person-group> (<year>1994</year>). <article-title>An algorithm for pronominal anaphora resolution</article-title>. <source>Comput. Linguist.</source> <volume>20</volume>, <fpage>535</fpage>&#x02013;<lpage>561</lpage>.</citation></ref>
<ref id="B75"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewis</surname> <given-names>R. L.</given-names></name></person-group> (<year>1996</year>). <article-title>Interference in short-term memory: the magical number two (or three) in sentence processing</article-title>. <source>J. Psycholinguist. Res.</source> <volume>25</volume>, <fpage>93</fpage>&#x02013;<lpage>115</lpage>.<pub-id pub-id-type="doi">10.1007/BF01708421</pub-id><pub-id pub-id-type="pmid">8789368</pub-id></citation></ref>
<ref id="B76"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewis</surname> <given-names>R. L.</given-names></name> <name><surname>Vasishth</surname> <given-names>S.</given-names></name></person-group> (<year>2005</year>). <article-title>An activation-based model of sentence processing as skilled memory retrieval</article-title>. <source>Cogn. Sci.</source> <volume>29</volume>, <fpage>375</fpage>&#x02013;<lpage>419</lpage>.<pub-id pub-id-type="doi">10.1207/s15516709cog0000_25</pub-id><pub-id pub-id-type="pmid">21702779</pub-id></citation></ref>
<ref id="B77"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewis</surname> <given-names>R. L.</given-names></name> <name><surname>Vasishth</surname> <given-names>S.</given-names></name> <name><surname>Van Dyke</surname> <given-names>J. A.</given-names></name></person-group> (<year>2006</year>). <article-title>Computational principles of working memory in sentence comprehension</article-title>. <source>Trends Cogn. Sci.</source> <volume>10</volume>, <fpage>447</fpage>&#x02013;<lpage>454</lpage>.<pub-id pub-id-type="doi">10.1016/j.tics.2006.08.007</pub-id><pub-id pub-id-type="pmid">16949330</pub-id></citation></ref>
<ref id="B78"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Malhotra</surname> <given-names>G.</given-names></name></person-group> (<year>2009</year>). <source>Dynamics of Structural Priming</source>. <publisher-name>Doctoral dissertation, University of Edinburgh</publisher-name>, <publisher-loc>Edinburgh</publisher-loc>.</citation></ref>
<ref id="B79"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Martinussen</surname> <given-names>R.</given-names></name> <name><surname>Hayden</surname> <given-names>J.</given-names></name> <name><surname>Hogg-Johnson</surname> <given-names>S.</given-names></name> <name><surname>Tannock</surname> <given-names>R.</given-names></name></person-group> (<year>2005</year>). <article-title>A meta-analysis of working memory impairments in children with attention-deficit/hyperactivity disorder</article-title>. <source>J. Am. Acad. Child Adolesc. Psychiatry</source> <volume>44</volume>, <fpage>377</fpage>&#x02013;<lpage>384</lpage>.<pub-id pub-id-type="doi">10.1097/01.chi.0000153228.72591.73</pub-id><pub-id pub-id-type="pmid">15782085</pub-id></citation></ref>
<ref id="B80"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>McKee</surname> <given-names>C.</given-names></name></person-group> (<year>1992</year>). <article-title>A comparison of pronouns and anaphors in Italian and English acquisition</article-title>. <source>Lang. Acquis.</source> <volume>2</volume>, <fpage>21</fpage>&#x02013;<lpage>54</lpage>.<pub-id pub-id-type="doi">10.1207/s15327817la0201_2</pub-id></citation></ref>
<ref id="B81"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Newell</surname> <given-names>A.</given-names></name></person-group> (<year>1990</year>). <source>Unified Theories of Cognition</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>Harvard University Press</publisher-name>.</citation></ref>
<ref id="B82"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Osborne</surname> <given-names>T.</given-names></name> <name><surname>Putnam</surname> <given-names>M.</given-names></name> <name><surname>Gross</surname> <given-names>T. M.</given-names></name></person-group> (<year>2011</year>). <article-title>Bare phrase structure, label-less trees, and specifier-less syntax. Is minimalism becoming a dependency grammar?</article-title> <source>Linguist. Rev.</source> <volume>28</volume>, <fpage>315</fpage>&#x02013;<lpage>364</lpage>.<pub-id pub-id-type="doi">10.1515/tlir.2011.009</pub-id></citation></ref>
<ref id="B83"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pauls</surname> <given-names>F.</given-names></name> <name><surname>Macha</surname> <given-names>T.</given-names></name> <name><surname>Petermann</surname> <given-names>F.</given-names></name></person-group> (<year>2013</year>). <article-title>U-shaped development: an old but unsolved problem</article-title>. <source>Front. Psychol.</source> <volume>4</volume>:<fpage>301</fpage>.<pub-id pub-id-type="doi">10.3389/fpsyg.2013.00301</pub-id><pub-id pub-id-type="pmid">23750146</pub-id></citation></ref>
<ref id="B84"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Philip</surname> <given-names>W.</given-names></name> <name><surname>Coopmans</surname> <given-names>P.</given-names></name></person-group> (<year>1996</year>). <article-title>&#x0201C;The role of lexical feature acquisition in the development of pronominal anaphora,&#x0201D;</article-title> in <source>Amsterdam Series on Child Language Development</source>, Vol. <volume>5</volume>, eds <person-group person-group-type="editor"><name><surname>Philip</surname> <given-names>W.</given-names></name> <name><surname>Wijnen</surname> <given-names>F.</given-names></name></person-group> (<publisher-loc>Amsterdam, The Netherlands</publisher-loc>: <publisher-name>University of Amsterdam</publisher-name>), <fpage>73</fpage>&#x02013;<lpage>106</lpage>.</citation></ref>
<ref id="B85"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pickering</surname> <given-names>M. J.</given-names></name> <name><surname>Branigan</surname> <given-names>H. P.</given-names></name></person-group> (<year>1998</year>). <article-title>The representation of verbs: evidence from syntactic priming in language production</article-title>. <source>J. Mem. Lang.</source> <volume>39</volume>, <fpage>633</fpage>&#x02013;<lpage>651</lpage>.<pub-id pub-id-type="doi">10.1006/jmla.1998.2592</pub-id></citation></ref>
<ref id="B86"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pickering</surname> <given-names>M. J.</given-names></name> <name><surname>Ferreira</surname> <given-names>V. S.</given-names></name></person-group> (<year>2008</year>). <article-title>Structural priming: a critical review</article-title>. <source>Psychol. Bull.</source> <volume>134</volume>, <fpage>427</fpage>&#x02013;<lpage>459</lpage>.<pub-id pub-id-type="doi">10.1037/0033-2909.134.3.427</pub-id><pub-id pub-id-type="pmid">18444704</pub-id></citation></ref>
<ref id="B87"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Pollard</surname> <given-names>C.</given-names></name> <name><surname>Sag</surname> <given-names>I. A.</given-names></name></person-group> (<year>1994</year>). <source>Head-Driven Phrase Structure Grammar</source>. <publisher-loc>Chicago, IL</publisher-loc>: <publisher-name>University of Chicago Press</publisher-name>.</citation></ref>
<ref id="B88"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Prince</surname> <given-names>A.</given-names></name> <name><surname>Smolensky</surname> <given-names>P.</given-names></name></person-group> (<year>1993</year>). <source>Optimality Theory: Constraint Interaction in Generative Grammar</source>. Technical Report CU-CS-696-93, <publisher-name>Department of Computer Science, University of Colorado at Boulder, and Technical Report TR-2, Rutgers Center for Cognitive Science, Rutgers University</publisher-name>, <publisher-loc>New Brunswick, NJ</publisher-loc>. Published by Blackwell in 2004.</citation></ref>
<ref id="B89"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Reitter</surname> <given-names>D.</given-names></name></person-group> (<year>2008</year>). <source>Context Effects in Language Production: Models of Syntactic Priming in Dialogue Corpora</source>. <publisher-name>Doctoral dissertation, University of Edinburgh</publisher-name>, <publisher-loc>Edinburgh</publisher-loc>.</citation></ref>
<ref id="B90"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Reitter</surname> <given-names>D.</given-names></name> <name><surname>Keller</surname> <given-names>F.</given-names></name> <name><surname>Moore</surname> <given-names>J. D.</given-names></name></person-group> (<year>2011</year>). <article-title>A computational cognitive model of syntactic priming</article-title>. <source>Cogn. Sci.</source> <volume>35</volume>, <fpage>587</fpage>&#x02013;<lpage>637</lpage>.<pub-id pub-id-type="doi">10.1111/j.1551-6709.2010.01165.x</pub-id><pub-id pub-id-type="pmid">21564266</pub-id></citation></ref>
<ref id="B91"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Reitter</surname> <given-names>D.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>&#x0201C;Accountable modeling in ACT-UP, a scalable, rapid-prototyping ACT-R implementation,&#x0201D;</article-title> in <conf-name>Proceedings of the 10th International Conference on Cognitive Modeling (ICCM)</conf-name> (<conf-loc>Philadelphia, PA</conf-loc>), <fpage>199</fpage>&#x02013;<lpage>204</lpage>.</citation></ref>
<ref id="B92"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Reitter</surname> <given-names>D.</given-names></name> <name><surname>Moore</surname> <given-names>J. D.</given-names></name></person-group> (<year>2014</year>). <article-title>Alignment and task success in spoken dialogue</article-title>. <source>J. Mem. Lang.</source> <volume>76</volume>, <fpage>29</fpage>&#x02013;<lpage>46</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2014.05.008</pub-id></citation></ref>
<ref id="B93"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Reitter</surname> <given-names>D.</given-names></name> <name><surname>Moore</surname> <given-names>J. D.</given-names></name> <name><surname>Keller</surname> <given-names>F.</given-names></name></person-group> (<year>2006</year>). <article-title>&#x0201C;Priming of syntactic rules in task-oriented dialogue and spontaneous conversation,&#x0201D;</article-title> in <conf-name>Proceedings of the 28th Annual Conference of the Cognitive Science Society (CogSci)</conf-name> (<conf-loc>Vancouver, Canada</conf-loc>: <conf-sponsor>Cognitive Science Society</conf-sponsor>), <fpage>685</fpage>&#x02013;<lpage>690</lpage>.</citation></ref>
<ref id="B94"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roberts</surname> <given-names>S.</given-names></name> <name><surname>Pashler</surname> <given-names>H.</given-names></name></person-group> (<year>2000</year>). <article-title>How persuasive is a good fit? A comment on theory testing</article-title>. <source>Psychol. Rev.</source> <volume>107</volume>, <fpage>358</fpage>&#x02013;<lpage>367</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.107.2.358</pub-id><pub-id pub-id-type="pmid">10789200</pub-id></citation></ref>
<ref id="B95"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roelofs</surname> <given-names>A.</given-names></name></person-group> (<year>1992</year>). <article-title>A spreading-activation theory of lemma retrieval in speaking</article-title>. <source>Cognition</source> <volume>42</volume>, <fpage>107</fpage>&#x02013;<lpage>142</lpage>.<pub-id pub-id-type="doi">10.1016/0010-0277(92)90041-F</pub-id></citation></ref>
<ref id="B96"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Salvucci</surname> <given-names>D. D.</given-names></name> <name><surname>Taatgen</surname> <given-names>N. A.</given-names></name></person-group> (<year>2008</year>). <article-title>Threaded cognition: an integrated theory of concurrent multitasking</article-title>. <source>Psychol. Rev.</source> <volume>115</volume>, <fpage>101</fpage>&#x02013;<lpage>130</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.115.1.101</pub-id><pub-id pub-id-type="pmid">18211187</pub-id></citation></ref>
<ref id="B97"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scheepers</surname> <given-names>C.</given-names></name></person-group> (<year>2003</year>). <article-title>Syntactic priming of relative clause attachments: persistence of structural configuration in sentence production</article-title>. <source>Cognition</source> <volume>89</volume>, <fpage>179</fpage>&#x02013;<lpage>205</lpage>.<pub-id pub-id-type="doi">10.1016/S0010-0277(03)00119-7</pub-id><pub-id pub-id-type="pmid">12963261</pub-id></citation></ref>
<ref id="B98"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scheepers</surname> <given-names>C.</given-names></name> <name><surname>Raffray</surname> <given-names>C. N.</given-names></name> <name><surname>Myachykov</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>The lexical boost effect is not diagnostic of lexically-specific syntactic representations</article-title>. <source>J. Mem. Lang.</source> <volume>95</volume>, <fpage>102</fpage>&#x02013;<lpage>115</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2017.03.001</pub-id></citation></ref>
<ref id="B99"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schoonbaert</surname> <given-names>S.</given-names></name> <name><surname>Hartsuiker</surname> <given-names>R. J.</given-names></name> <name><surname>Pickering</surname> <given-names>M. J.</given-names></name></person-group> (<year>2007</year>). <article-title>The representation of lexical and syntactic information in bilinguals: evidence from syntactic priming</article-title>. <source>J. Mem. Lang.</source> <volume>56</volume>, <fpage>153</fpage>&#x02013;<lpage>171</lpage>.<pub-id pub-id-type="doi">10.1016/j.jml.2006.10.002</pub-id></citation></ref>
<ref id="B100"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sense</surname> <given-names>F.</given-names></name> <name><surname>Behrens</surname> <given-names>F.</given-names></name> <name><surname>Meijer</surname> <given-names>R. R.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name></person-group> (<year>2016</year>). <article-title>An individual&#x02019;s rate of forgetting is stable over time but differs across materials</article-title>. <source>Top. Cogn. Sci.</source> <volume>8</volume>, <fpage>305</fpage>&#x02013;<lpage>321</lpage>.<pub-id pub-id-type="doi">10.1111/tops.12183</pub-id></citation></ref>
<ref id="B101"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Snider</surname> <given-names>N. E.</given-names></name></person-group> (<year>2008</year>). <source>An Exemplar Model of Syntactic Priming</source>. <publisher-name>Doctoral dissertation, Stanford University</publisher-name>, <publisher-loc>Stanford</publisher-loc>.</citation></ref>
<ref id="B102"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spenader</surname> <given-names>J.</given-names></name> <name><surname>Smits</surname> <given-names>E.-J.</given-names></name> <name><surname>Hendriks</surname> <given-names>P.</given-names></name></person-group> (<year>2009</year>). <article-title>Coherent discourse solves the pronoun interpretation problem</article-title>. <source>J. Child Lang.</source> <volume>36</volume>, <fpage>23</fpage>&#x02013;<lpage>52</lpage>.<pub-id pub-id-type="doi">10.1017/S0305000908008854</pub-id><pub-id pub-id-type="pmid">18761775</pub-id></citation></ref>
<ref id="B103"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Steedman</surname> <given-names>M.</given-names></name></person-group> (<year>1988</year>). <article-title>&#x0201C;Combinators and grammars,&#x0201D;</article-title> in <source>Categorial Grammars and Natural Language Structures</source>, eds <person-group person-group-type="editor"><name><surname>Oehrle</surname> <given-names>R. T.</given-names></name> <name><surname>Bach</surname> <given-names>E.</given-names></name> <name><surname>Wheeler</surname> <given-names>D.</given-names></name></person-group> (<publisher-loc>Dordrecht, The Netherlands</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>417</fpage>&#x02013;<lpage>442</lpage>.</citation></ref>
<ref id="B104"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Steedman</surname> <given-names>M.</given-names></name></person-group> (<year>2000</year>). <source>The Syntactic Process</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B105"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stewart</surname> <given-names>T.</given-names></name> <name><surname>Tripp</surname> <given-names>B.</given-names></name> <name><surname>Eliasmith</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>Python scripting in the Nengo simulator</article-title>. <source>Front. Neuroinformatics</source> <volume>3</volume>:<fpage>7</fpage>.<pub-id pub-id-type="doi">10.3389/neuro.11.007.2009</pub-id><pub-id pub-id-type="pmid">19352442</pub-id></citation></ref>
<ref id="B106"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stocco</surname> <given-names>A.</given-names></name> <name><surname>Crescentini</surname> <given-names>C.</given-names></name></person-group> (<year>2005</year>). <article-title>Syntactic comprehension in agrammatism: a computational model</article-title>. <source>Brain Lang.</source> <volume>95</volume>, <fpage>127</fpage>&#x02013;<lpage>128</lpage>.<pub-id pub-id-type="doi">10.1016/j.bandl.2005.07.069</pub-id></citation></ref>
<ref id="B107"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taatgen</surname> <given-names>N.</given-names></name> <name><surname>Anderson</surname> <given-names>J. R.</given-names></name></person-group> (<year>2002</year>). <article-title>Why do children learn to say &#x0201C;broke&#x0201D;? A model of learning the past tense without feedback</article-title>. <source>Cognition</source> <volume>86</volume>, <fpage>123</fpage>&#x02013;<lpage>155</lpage>.<pub-id pub-id-type="doi">10.1016/S0010-0277(02)00176-2</pub-id></citation></ref>
<ref id="B108"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tanenhaus</surname> <given-names>M. K.</given-names></name> <name><surname>Spivey-Knowlton</surname> <given-names>M. J.</given-names></name> <name><surname>Eberhard</surname> <given-names>K. M.</given-names></name> <name><surname>Sedivy</surname> <given-names>J. C.</given-names></name></person-group> (<year>1995</year>). <article-title>Integration of visual and linguistic information in spoken language comprehension</article-title>. <source>Science</source> <volume>268</volume>, <fpage>1632</fpage>&#x02013;<lpage>1634</lpage>.<pub-id pub-id-type="doi">10.1126/science.7777863</pub-id><pub-id pub-id-type="pmid">7777863</pub-id></citation></ref>
<ref id="B109"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Beijsterveldt</surname> <given-names>L. M.</given-names></name> <name><surname>Van Hell</surname> <given-names>J. G.</given-names></name></person-group> (<year>2009</year>). <article-title>Structural priming of adjective-noun structures in hearing and deaf children</article-title>. <source>J. Exp. Child Psychol.</source> <volume>104</volume>, <fpage>179</fpage>&#x02013;<lpage>196</lpage>.<pub-id pub-id-type="doi">10.1016/j.jecp.2009.05.002</pub-id><pub-id pub-id-type="pmid">19497585</pub-id></citation></ref>
<ref id="B110"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Maanen</surname> <given-names>L.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name></person-group> (<year>2010</year>). <article-title>The locus of the Gratton effect in picture-word interference</article-title>. <source>Top. Cogn. Sci.</source> <volume>2</volume>, <fpage>168</fpage>&#x02013;<lpage>180</lpage>.<pub-id pub-id-type="doi">10.1111/j.1756-8765.2009.01069.x</pub-id></citation></ref>
<ref id="B111"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Van Rij</surname> <given-names>J.</given-names></name></person-group> (<year>2012</year>). <source>Pronoun Processing: Computational, Behavioral, and Psychophysiological Studies in Children and Adults</source>. <publisher-name>Doctoral dissertation, University of Groningen</publisher-name>, <publisher-loc>Groningen, The Netherlands</publisher-loc>.</citation></ref>
<ref id="B112"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Rij</surname> <given-names>J.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name> <name><surname>Hendriks</surname> <given-names>P.</given-names></name></person-group> (<year>2010</year>). <article-title>Cognitive architectures and language acquisition: a case study in pronoun comprehension</article-title>. <source>J. Child Lang.</source> <volume>37</volume>, <fpage>731</fpage>&#x02013;<lpage>766</lpage>.<pub-id pub-id-type="doi">10.1017/S0305000909990560</pub-id><pub-id pub-id-type="pmid">20353614</pub-id></citation></ref>
<ref id="B113"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Rij</surname> <given-names>J.</given-names></name> <name><surname>Van Rijn</surname> <given-names>H.</given-names></name> <name><surname>Hendriks</surname> <given-names>P.</given-names></name></person-group> (<year>2013</year>). <article-title>How WM load influences linguistic processing in adults: a computational model of pronoun interpretation in discourse</article-title>. <source>Top. Cogn. Sci.</source> <volume>5</volume>, <fpage>564</fpage>&#x02013;<lpage>580</lpage>.<pub-id pub-id-type="doi">10.1111/tops.12029</pub-id><pub-id pub-id-type="pmid">23757182</pub-id></citation></ref>
<ref id="B114"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Vogelzang</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <source>Reference and Cognition: Experimental and Computational Cognitive Modeling Studies on Reference Processing in Dutch and Italian</source>. <publisher-name>Doctoral dissertation, University of Groningen</publisher-name>, <publisher-loc>Groningen, Netherlands</publisher-loc>.</citation></ref>
</ref-list>
</back>
</article>
