<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychology</journal-id>
<journal-title>Frontiers in Psychology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychology</abbrev-journal-title>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Research Foundation</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyg.2012.00177</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>The Role of Inertia in Modeling Decisions from Experience with Instance-Based Learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Dutt</surname> <given-names>Varun</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn001">&#x0002A;</xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Gonzalez</surname> <given-names>Cleotilde</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>School of Computing and Electrical Engineering and School of Humanities and Social Sciences, Indian Institute of Technology</institution> <country>Mandi, India</country></aff>
<aff id="aff2"><sup>2</sup><institution>Dynamic Decision Making Laboratory, Department of Social and Decision Sciences, Carnegie Mellon University</institution> <country>Pittsburgh, PA, USA</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Konstantinos Tsetsos, Oxford University, UK</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Christian C. Luhmann, Stony Brook University, USA; Adrian R. Camilleri, Duke University, USA</p></fn>
<fn fn-type="corresp" id="fn001"><p>&#x0002A;Correspondence: Varun Dutt, School of Computing and Electrical Engineering and School of Humanities and Social Sciences, Indian Institute of Technology, Mandi, PWD Rest House, Near Bus Stand, Mandi &#x02013; 175 001, Himachal Pradesh, India. e-mail: <email>varundutt&#x00040;yahoo.com</email></p></fn>
<fn fn-type="other" id="fn002"><p>This article was submitted to Frontiers in Cognitive Science, a specialty of Frontiers in Psychology.</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>06</day>
<month>06</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="collection">
<year>2012</year>
</pub-date>
<volume>3</volume>
<elocation-id>177</elocation-id>
<history>
<date date-type="received">
<day>15</day>
<month>02</month>
<year>2012</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>05</month>
<year>2012</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2012 Dutt and Gonzalez.</copyright-statement>
<copyright-year>2012</copyright-year>
<license license-type="open-access" xlink:href="http://www.frontiersin.org/licenseagreement"><p>This is an open-access article distributed under the terms of the <uri xlink:href="http://creativecommons.org/licenses/by-nc/3.0/">Creative Commons Attribution Non Commercial License</uri>, which permits non-commercial use, distribution, and reproduction in other forums, provided the original authors and source are credited.</p></license>
</permissions>
<abstract>
<p>One form of inertia is the tendency to repeat the last decision irrespective of the obtained outcomes while making decisions from experience (DFE). A number of computational models based upon the Instance-Based Learning Theory, a theory of DFE, have included different inertia implementations and have shown to simultaneously account for both risk-taking and alternations between alternatives. The role that inertia plays in these models, however, is unclear as the same model without inertia is also able to account for observed risk-taking quite well. This paper demonstrates the predictive benefits of incorporating one particular implementation of inertia in an existing IBL model. We use two large datasets, estimation and competition, from the Technion Prediction Tournament involving a repeated binary-choice task to show that incorporating an inertia mechanism in an IBL model enables it to account for the observed average risk-taking and alternations. Including inertia, however, does not help the model to account for the trends in risk-taking and alternations over trials compared to the IBL model without the inertia mechanism. We generalize the two IBL models, with and without inertia, to the competition set by using the parameters determined in the estimation set. The generalization process demonstrates both the advantages and disadvantages of including inertia in an IBL model.</p>
</abstract>
<kwd-group>
<kwd>decisions from experience</kwd>
<kwd>instance-based learning</kwd>
<kwd>binary-choice</kwd>
<kwd>inertia</kwd>
<kwd>risk-taking</kwd>
<kwd>alternations</kwd>
</kwd-group>
<counts>
<fig-count count="4"/>
<table-count count="2"/>
<equation-count count="5"/>
<ref-count count="40"/>
<page-count count="12"/>
<word-count count="10331"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="introduction">
<title>Introduction</title>
<p>People&#x02019;s reliance on inertia, the tendency to repeat the last decision irrespective of the obtained outcomes (successes or failures), has been documented in literature concerning managerial and organizational sciences as well as behavioral sciences (Samuelson, <xref ref-type="bibr" rid="B37">1994</xref>; Reger and Palmer, <xref ref-type="bibr" rid="B36">1996</xref>; Hodgkinson, <xref ref-type="bibr" rid="B25">1997</xref>; Tripsas and Gavetti, <xref ref-type="bibr" rid="B40">2000</xref>; Gladwell, <xref ref-type="bibr" rid="B18">2007</xref>; Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Nevo and Erev, <xref ref-type="bibr" rid="B33">2012</xref>). For example, inertia acts like a <italic>status quo</italic> bias and helps to account for the commonly observed phenomenon whereby managers fail to update and revise their understanding of a situation when it changes, a phenomenon that acts as a psychological barrier to organizational change (Reger and Palmer, <xref ref-type="bibr" rid="B36">1996</xref>; Tripsas and Gavetti, <xref ref-type="bibr" rid="B40">2000</xref>; Gladwell, <xref ref-type="bibr" rid="B18">2007</xref>). In these situations, inertia is generally believed to have a negative effect on decision making (Sandri et al., <xref ref-type="bibr" rid="B38">2010</xref>).</p>
<p>Inertia has also been incorporated to account for human behavior in existing computational models of decisions from experience (DFE). DFE are choices that are based on previous encounters with one&#x02019;s alternatives; as opposed to decisions from description, which are based on summary descriptions detailing all possible outcomes and their respective likelihoods of each option (Hertwig and Erev, <xref ref-type="bibr" rid="B24">2009</xref>). In DFE, researchers have studied both the risk-taking behavior and alternations between alternatives in repeated binary-choice tasks, where decision makers consequentially choose between risky and safe alternatives repeatedly (Samuelson, <xref ref-type="bibr" rid="B37">1994</xref>; B&#x000F6;rgers and Sarin, <xref ref-type="bibr" rid="B8">2000</xref>; Barron and Erev, <xref ref-type="bibr" rid="B5">2003</xref>; Erev and Barron, <xref ref-type="bibr" rid="B14">2005</xref>; Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>; Hertwig and Erev, <xref ref-type="bibr" rid="B24">2009</xref>; Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Nevo and Erev, <xref ref-type="bibr" rid="B33">2012</xref>). The alternations explain how individuals search information and how this search pattern changes over repeated trials. Thus, alternations tell us about the information-search patterns and learning in DFE (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>). Accounting for both risk-taking and alternations helps to develop a complete understanding about how decision makers reach certain long-term outcomes, which cannot be determined by solely studying one of these measures in the isolation of the other (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>).</p>
<p>Most recently, models based upon the Instance-Based Learning Theory (IBLT; and &#x0201C;IBL models&#x0201D; hereafter), a theory of dynamic DFE, have shown to account for both the observed risk-taking and alternations in a binary-choice task better than most of the best known computational models. A number of these IBL models have incorporated some form of the inertia mechanism (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>), while others have not incorporated inertia and still accounted for the risk-taking behavior (Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>). For example, Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>) have shown that a single IBL model, without inertia, is able to explain observed risk-taking and generalize across several variants of the repeated binary-choice task. Therefore, it appears that inertia may not be needed in computational models to account for the observed risk-taking. However, Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>) model does not demonstrate how alternations are accounted for or how alternations and risk-taking are accounted for simultaneously. As discussed above, people&#x02019;s experiential decisions may likely rely on inertia, and computational models might need some form of inertia to account for both observed risk-taking and alternations. Yet, the role that inertia mechanisms play in existing computational models is unclear and needs to be systematically investigated.</p>
<p>In this paper, we evaluate the role of an inertia mechanism in an IBL model. We evaluate a model with inertia and another without inertia for their ability to account for observed risk-taking and alternation behaviors. In order to evaluate the inertia mechanism, we use two large human datasets that were collected in the Technion Prediction Tournament (TPT) involving the repeated binary-choice task (Erev et al., <xref ref-type="bibr" rid="B16">2010b</xref>). In what follows, we first discuss the current understanding of the role of inertia in accounting for DFE. Next, we present the results of calibrating two existing IBL models, with and without inertia, in the TPT&#x02019;s estimation dataset and evaluate the added value and contribution of including inertia. Finally, we present the results that generalize these models into the TPT&#x02019;s competition dataset. We close this paper by discussing our results and highlighting some future directions in this ongoing research program.</p>
<sec>
<title>The role of inertia in decisions from experience</title>
<p>Inertia may be a psychological barrier to changes in an organization if decision makers fail to update their understanding of a situation when it changes (Reger and Palmer, <xref ref-type="bibr" rid="B36">1996</xref>; Hodgkinson, <xref ref-type="bibr" rid="B25">1997</xref>; Tripsas and Gavetti, <xref ref-type="bibr" rid="B40">2000</xref>; Gladwell, <xref ref-type="bibr" rid="B18">2007</xref>). For example, Tripsas and Gavetti (<xref ref-type="bibr" rid="B40">2000</xref>) provided a popular example of inertia in a managerial setting concerning the Polaroid Corporation. Polaroid believed that it could only make money by producing consumables and not the hardware. Thus, it decided to stick to producing only consumables. This decision led the company to neglect the growth in digital imaging technologies. Because of the prevailing inertial &#x0201C;mental model&#x0201D; of their business, the corporation failed to adapt effectively to market changes. Furthermore, Gladwell (<xref ref-type="bibr" rid="B18">2007</xref>) has suggested that inertia is one powerful explanation as to why established firms are not as innovative as young, less established firms. For example, as an established firm, Kodak&#x02019;s management is reported to have suffered from a <italic>status quo</italic> bias due to inertia: They believed that what has worked in the past will also work in the future (Gladwell, <xref ref-type="bibr" rid="B18">2007</xref>).</p>
<p>In judgment and decision making, inertia has been shown to play a role in determining the proportion of risk-taking due to the timing of a descriptive warning message (Barron et al., <xref ref-type="bibr" rid="B6">2008</xref>). Barron et al. (<xref ref-type="bibr" rid="B6">2008</xref>) compared the effect of a descriptive warning received <italic>before</italic> or <italic>after</italic> making risky decisions in a repeated binary-choice task. In this task, participants made a choice between a safe option with a sure gain and a risky option with the possibility of incurring a loss or a gain such that the probability of incurring the loss was very small (<italic>p</italic>&#x02009;&#x0003D;&#x02009;0.001). Thus, most of the time, the task offered gains for both safe and risky choices. These authors show that when an early warning coincides with the beginning of a decision making process, the warning is both weighted more heavily in future decisions and induces safer behavior (i.e., a decrease in the proportion of risky choices), which becomes the <italic>status quo</italic> for future choices. Thus, although the proportion of risk-taking is lower for an early warning message compared to a late warning message, the risky and safe choices in both cases show excessive reliance on inertia to repeat the last choice made. Here, inertia acts like a double-edged sword: It is likely to encourage or discourage ongoing risky behavior depending upon the timing of a warning.</p>
<p>Some researchers have depicted inertia as an irrational behavior in which individuals hold onto choices that clearly do not provide the maximizing outcome for too long (Sandri et al., <xref ref-type="bibr" rid="B38">2010</xref>). However, these authors have only shown that behavior may be inconsistent with one specific rational model of maximization, which may be an arbitrary standard that is difficult to generalize to other rational models of maximization. There are certain other situations where inertia is likely to produce positive effects as well. In psychology, inertia is also believed to be a key component of love, trust, and friendship (Cook et al., <xref ref-type="bibr" rid="B12">2005</xref>). If evidence shows that a friend is dishonest, then the decision to mistrust the friend in future interactions would demand much more instances of dishonesty from the friend than that required to form an opinion about a stranger. Thus, the inertia of continuing to trust the friend makes it difficult to break the friendship.</p>
<p>Inertia has been incorporated in a number of existing cognitive models of DFE. It is believed that inertia helps these models account for both observed risk-taking and alternations in the repeated binary-choice (Samuelson, <xref ref-type="bibr" rid="B37">1994</xref>; B&#x000F6;rgers and Sarin, <xref ref-type="bibr" rid="B8">2000</xref>; Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>; Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Nevo and Erev, <xref ref-type="bibr" rid="B33">2012</xref>). For example, Erev et al. (<xref ref-type="bibr" rid="B15">2010a</xref>) observed that in the repeated binary-choice task, participants selected the alternative that led to an observed high outcome in the last trial in 67.4% of the trials, while they repeated their last choice for an alternative, irrespective of it being high or low, in 75% of the trials. These observations suggest that participants tend to repeat their last choice even when it does not agree with the high outcome in their last experience, exhibiting robust reliance on inertia that seems to be independent of observed outcomes. Some researchers have suggested that in situations where estimating the choice that yields high outcomes from observation is costly, difficult, or time consuming, relying on inertia might be the most feasible course of action (Samuelson, <xref ref-type="bibr" rid="B37">1994</xref>). But other researchers have found this inertia effect even when the forgone outcome (i.e., what respondents would have gotten had they chosen the other alternative) is greater than the obtained outcome (Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>).</p>
<p>In order to account for these observations, recent computational models of DFE have explicitly incorporated three different forms of inertia as part of their specification (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). In the first form, inertia increases over time as a result of a decrease in surprise, where surprise is defined as the difference in expected values of the two alternatives (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>). This definition of inertia has been included in the Inertia Sampling and Weighting (I-SAW) model. The I-SAW model was designed for a repeated binary-choice market-entry task, and it distinguishes between three explicit response modes: exploration, exploitation, and inertia (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Chen et al., <xref ref-type="bibr" rid="B11">2011</xref>). The I-SAW model also provides reasonable predictions in the repeated binary-choice task (Nevo and Erev, <xref ref-type="bibr" rid="B33">2012</xref>). Inertia is represented in this model with the assumption that individuals tend to repeat their last choice, and the probability of inertia in a trial is a function of surprise. Surprise is calculated as the difference in the expected value of the two alternatives due to the observed outcomes in each alternative in previous trials. The probability of inertia is assumed to increase over trials, as surprise decreases over trials. This definition based upon surprise incorporates the idea of learning over repeated trials of game play where, due to repeated presentations of the same set of outcomes, participants tend to get increasingly less surprised and begin to stick to an option that they prefer (i.e., show inertia in their decisions).</p>
<p>In the second form (that is similar to the first form), inertia increases over time as a result of a decrease in surprise, which is based upon the difference in blended values (a measure of utility of alternatives based on past experience in Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref> model). This definition of inertia has been included in the IBL model that was runner-up in the market-entry competition (Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). This model includes an inertia mechanism that is driven by surprise like in the I-SAW model; however, surprise here is calculated as the difference between the blended values of two alternatives.</p>
<p>In the third and simpler form, inertia is a probabilistic process that is triggered randomly over trials, where the random occurrences of inertia are based upon a calibrated probability parameter, <italic>pInertia</italic> (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). This definition of inertia is the one we evaluate in this paper, as it was recently included in an IBL model that produced robust predictions superior to many existing models (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). According to Gonzalez and Dutt (<xref ref-type="bibr" rid="B20">2011</xref>), the IBL model with the <italic>pInertia</italic> parameter accounts for both observed risk-taking and alternations simultaneously in different paradigms of DFE and performs consistently better than most existing computational models of DFE that competed in the TPT.</p>
<p>Although computational models have included inertia in several forms, Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>) have recently shown that a single IBL model without any inertia assumption is also able to account for the observed risk-taking behavior in different tasks that included probability-learning, binary-choice with fixed probability, and binary-choice with changing probability. Although the use of some form of inertia seems necessary in many computational models of DFE (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Chen et al., <xref ref-type="bibr" rid="B11">2011</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>; Nevo and Erev, <xref ref-type="bibr" rid="B33">2012</xref>), its role in accounting for risk-taking and alternations in DFE is still unclear and a systematic investigation of its role in computational models is needed.</p>
<p>Given the wide use of inertia in computational models, it is likely that incorporating inertia assumptions might make them more ecologically valid. That seems likely because if a model accounts for risk-taking behavior already, then incorporating a form of inertia in its specification might directly influence its ability to account for alternations as well. However, we currently do not know how inertia in a model might impact its ability to account for both the risk-taking behavior and the alternations simultaneously. The incorporation of inertia in a model is likely to be beneficial only if it improves the model&#x02019;s ability to account for both risk-taking and alternations, and not solely one of these measures.</p>
</sec>
</sec>
<sec sec-type="materials|methods">
<title>Materials and Methods</title>
<sec>
<title>Risk-taking and alternations in the technion prediction tournament</title>
<p>The TPT (Erev et al., <xref ref-type="bibr" rid="B16">2010b</xref>) was a modeling competition organized in 2008 in which different models were submitted to predict choices made by human participants. Competing models were evaluated following the generalization criterion method (Busemeyer and Wang, <xref ref-type="bibr" rid="B10">2000</xref>), by which models were fitted to choices made by participants in 60 problems (the estimation set) and later tested in a new set of 60 problems (the competition set) with the parameters obtained in the estimation set. Although the TPT involved three different experimental paradigms, here we use data from the E-repeated paradigm that involved consequential choices in a repeated binary-choice task with immediate feedback on the chosen alternative. We use this dataset to evaluate the inertia mechanism in an IBL model.</p>
<p>The TPT dataset&#x02019;s 120 problems involved a choice between a safe alternative that offered a medium (M) outcome with certainty; and a risky alternative that offered a high (H) outcome with some probability (pH) and a low (L) outcome with the complementary probability. The M, H, pH, and L were generated randomly, and a selection algorithm assured that the 60 problems in each set were different in domain (positive, negative, and mixed outcomes) and probability (high, medium, and low pH). The positive domain was such that each of the M, H, and L outcomes in a problem were positive numbers (&#x0003E;0). The mixed domain was such that one or two of the outcomes among M, H, and L (but not all three) in a problem were negative (&#x0003C;0). The negative domain was such that each of the M, H, and L outcomes in a problem were negative numbers (&#x0003C;0). The low, medium, and high probability in a problem corresponded to the value of pH between 0.01&#x02013;0.09, 0.1&#x02013;0.9, and 0.91&#x02013;0.99, respectively. The selection algorithm ensured that there were 20 problems each for the three domains and about 20 problems each for the three probability values in the estimation and the competition sets. The resulting set of problems in the three domains and the three probability values was large and representative. For each of the 60 problems in the estimation and competition set, a sample of 100 participants was randomly assigned into 5 groups, and each group completed 12 of the 60 problems. Each participant was instructed to repeatedly and consequentially select between two unlabeled buttons on a computer screen in order to maximize long-term rewards for a block of 100 trials per problem (the end point on trials was not provided or known to participants). One button was associated with a risky alternative and the other button with a safe alternative. Clicking a button corresponding to either the safe or risky alternative generated an outcome associated with the selected button (i.e., there was only partial feedback and participants were not shown the foregone outcome on the unselected button). The alternative with the higher expected value, which could be either the safe or risky, could maximize a participant&#x02019;s long-term rewards. Other details about the E-repeated paradigm are reported in Erev et al. (<xref ref-type="bibr" rid="B16">2010b</xref>).</p>
<p>The models submitted to the TPT were not provided with the alternation data (i.e., the A-rate), and they were evaluated only according to their ability to account for risk-taking behavior (i.e., the R-rate; Erev et al., <xref ref-type="bibr" rid="B16">2010b</xref>). Gonzalez and Dutt (<xref ref-type="bibr" rid="B20">2011</xref>) had calculated the A-rate for analyses of alternations from the TPT datasets and we followed the exact same procedures in this paper. First, alternations were either coded as 1&#x02009;s (a respondent switched from making a risky or safe choice in the last trial to making a safe or risky choice in the current trial) or as 0&#x02009;s (the respondent repeated the same choice in the current trial as that in the last trial). Then, the A-rate is computed as the proportion of alternations in each trial starting in trial 2 (the A-rate in trial 1 is undefined as there is no preceding trial to calculate alternations). The proportion of alternations in each trial is computed by averaging the alternations over 20 participants per problem and 60 problems in each dataset. The R-rate is the proportion of risky choices (i.e., choices of the risky alternative) in each trial averaged over 20 participants per problem and 60 problems in each dataset.</p>
<p>Figure <xref ref-type="fig" rid="F1">1</xref> shows the overall R-rate and A-rate over 99 trials from trial 2 to 100 in the estimation and competition sets. As seen in both datasets, the R-rate decreases slightly across trials, although there is a sharp decrease in the A-rate. The sharp decrease in the A-rate shows a change in the exploration (information-search) pattern across repeated trials. Overall, the R-rate and A-rate curves suggest that participants&#x02019; risk-taking behavior remains relatively steady across trials, while they learn to alternate less and choose one of the two alternatives more often. Later in this paper, we evaluate the role of inertia mechanism to account for these R- and A-rate curves in Figure <xref ref-type="fig" rid="F1">1</xref> in a computational IBL model.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(A)</bold> The R-rate and A-rate across trials observed in human data in the estimation set of the TPT between trial 2 and trial 100. <bold>(B)</bold> The R-rate and A-rate across trials observed in human data in the competition set of the TPT between trial 2 and trial 100.</p></caption>
<graphic xlink:href="fpsyg-03-00177-g001.tif"/>
</fig>
</sec>
<sec>
<title>An instance-based learning model of repeated binary-choice</title>
<p>Instance-Based Learning Theory has been used for developing computational models that explain human behavior in a wide variety of dynamic decision making tasks. These tasks include dynamically complex tasks (Gonzalez and Lebiere, <xref ref-type="bibr" rid="B22">2005</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B23">2003</xref>; Martin et al., <xref ref-type="bibr" rid="B31">2004</xref>), training paradigms of simple and complex tasks (Gonzalez et al., <xref ref-type="bibr" rid="B19">2010</xref>), simple stimulus-response practice and skill acquisition tasks (Dutt et al., <xref ref-type="bibr" rid="B13">2009</xref>), and repeated binary-choice tasks (Lebiere et al., <xref ref-type="bibr" rid="B28">2007</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>; Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>) among others. Its applications to these diverse tasks illustrate its generality and its ability to explain DFE in multiple contexts.</p>
<p>Here, we briefly discuss an IBL model that has shown to successfully account for both risk-taking and alternation behaviors in DFE (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). This model assumes reliance on recency, frequency, and random inertia to make choice selections. Here, we evaluate how the same IBL model, with and without the random inertia mechanism, can simultaneously account for risk-taking and alternation in repeated binary-choice. This evaluation will enable us to better understand the role of this particular simpler formulation of inertia in computational IBL models.</p>
<sec>
<title>IBL model</title>
<p>All IBL models propose an &#x0201C;instance&#x0201D; as a key representation of cognitive information. An instance is a representation of each decision alternative and consists of three parts: a situation (a set of attributes that define the alternative), a decision for one of the many alternatives, and an outcome resulting from making that decision. An IBL model of repeated binary-choice has assumed a simple instantiation of inertia mechanism (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>): A free parameter, called <italic>pInertia</italic>, determines the repetition of the previous choice in the current decision according to this probabilistic mechanism (see the <xref ref-type="app" rid="A1">Appendix</xref> for the formal definitions of all the mechanisms of the IBL model for repeated binary-choice and the <italic>pInertia</italic> parameter). If a uniformly distributed random number is less than the probability <italic>pInertia</italic>, then the model repeats its last choice; otherwise, the model compares blended values for the risky and safe alternatives, and makes a choice for the alternative with the higher blended value.</p>
<p>In this paper, we call this IBL model implementation with the random inertia mechanism, the &#x0201C;IBL-Inertia model.&#x0201D; In addition, we consider another version of the same model, but without this inertia mechanism (Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>) as a baseline to compare against the IBL-Inertia model. We call this implementation without inertia, the &#x0201C;IBL model.&#x0201D; In the absence of inertia, this model relies solely on the comparison between the blended values for the risky and safe alternatives to make choice selections in each trial (the IBL-Inertia model also compares blended values to make choice selections; however, blended values are used only in the IBL-Inertia model when a random number is greater than or equal to the <italic>pInertia</italic> parameter in a trial). With the exception of the presence of <italic>pInertia</italic> in the IBL-Inertia model and its absence in the IBL model, both models are identical in all other respects.</p>
<p>Blending, as proposed in both model implementations, is a function of the probability of retrieving instances from memory multiplied by their respective outcomes stored in instances (Lebiere, <xref ref-type="bibr" rid="B27">1999</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>). Each instance consists of a label that identifies a decision alternative in the task and the outcome obtained. For example (A, &#x00024;32) is an instance where the decision was to choose the risky alternative (A) and the outcome obtained was &#x00024;32. The probability of retrieving an instance from memory is a function of its activation (Anderson and Lebiere, <xref ref-type="bibr" rid="B3">1998</xref>). A simplified version of the activation mechanism, which relies on recency and frequency of using instances and noise in retrieval, has been shown to be sufficient to capture human choices in several repeated binary-choice and probability-learning tasks (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>). The activation is influenced by the decay parameter <italic>d</italic> that captures the rate of forgetting or reliance on recency. The higher the value of the <italic>d</italic> parameter, the greater is the model&#x02019;s reliance on recent experiences. The activation is also influenced by a noise parameter <italic>s</italic> that is important for capturing the variability in human behavior from one participant to another.</p>
<p>For the first trial, both model implementations, IBL-Inertia and IBL, have no instances in memory from which to calculate blended values. Therefore, these implementations make a selection between instances that are pre-populated in their memory. We used a value of &#x0002B;30 in the outcome slot of the two alternatives&#x02019; instances (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). The &#x0002B;30 value is arbitrary, but most importantly, it should be greater than any possible outcomes in the TPT problems to trigger an initial exploration of the two alternatives. For the first trial, the choice between the two alternatives in both implementations is solely based on the blended values. From the second trial onward, the inertia mechanism is used along with blending in IBL-Inertia model and only blending is used in the IBL model.</p>
</sec>
</sec>
</sec>
<sec>
<title>Results</title>
<sec>
<title>Model calibration and evaluation of inertia</title>
<p>The IBL model is compared with the IBL-Inertia model for their ability to account for both the proportion of risk-taking (R-rate) and alternations (A-rate) across trials. We will first calibrate the models&#x02019; shared parameters, noise <italic>s</italic> and decay <italic>d</italic>, to the data in the TPT&#x02019;s estimation set. Then, we explore the role of adding the <italic>pInertia</italic> parameter to the IBL model (i.e., the IBL-Inertia model) by recalibrating all its parameters. Then, we generalize both the calibrated models, IBL and IBL-Inertia, to the TPT&#x02019;s competition set.</p>
<p>Calibrating a model to human data means finding the parameter values that minimize the mean-squared deviation (MSD) between the model&#x02019;s predictions and the observed human performance on a dependent measure. We used a genetic algorithm program to calibrate the model&#x02019;s parameters. The genetic algorithm tried different combinations of parameters to minimize the sum of MSDs between the model&#x02019;s average R-rate per problem and the average A-rate per problem measures and the corresponding values in human data (we call this sum as the combined R-rate and A-rate measure). Calibrating on the combined R-rate and A-rate measure is expected to produce the best account for both measures in human data compared to using only one of these measures (Dutt and Gonzalez, under review). Also, calibrating on the combined R-rate and A-rate measure allows us to test the IBL model&#x02019;s maximum potential to account for both these measures.</p>
<p>In order to compare results on the R-rate and A-rate during calibration, we use the AIC (Akaike Information Criterion) measure in addition to the MSD (mean-squared deviation) measure. The AIC definition takes into account both a model&#x02019;s complexity (estimated by the number of free parameters in the model), as well as its accuracy (estimated by <italic>G</italic><sup>2</sup>, defined the &#x0201C;lack of fit&#x0201D; between model and human data; Pitt and Myung, <xref ref-type="bibr" rid="B34">2002</xref>; Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>). The AIC definition and the computation procedures used here are the same as those used by Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; for more details on the AIC definition refer to the Appendix). The use of AIC during calibration is relevant because the IBL and IBL-Inertia models are hierarchical (or nested) models (Maruyama, <xref ref-type="bibr" rid="B32">1997</xref>; Loehlin, <xref ref-type="bibr" rid="B30">2003</xref>; Kline, <xref ref-type="bibr" rid="B26">2004</xref>) and they differ only in terms of the inertia mechanism. Thus, the IBL model can be simply derived from the IBL-Inertia model by restricting the <italic>pInertia</italic> parameter&#x02019;s value to 0 during model calibration. Furthermore, in order to capture the trend of R-rate and A-rate from a model over trials, we used the Pearson&#x02019;s correlation coefficient (<italic>r</italic>) between model and human data across trials (for the A-rate we used trials 2&#x02013;100 and for the R-rate we used trials 1&#x02013;100; the A-rate is undefined for trial 1). Also, we computed the MSE (mean-squared error) between model and human data across trials. For the MSE, we averaged the R-rate and A-rate in model and human data across all participants and problems in a dataset for each trial. Then, we calculated the mean of the squared differences between model and human data for each trial. Because the MSE is computed across trials, it measures the distance between the model and human data curves trial-by-trial (for more details on the MSE definition refer to the Appendix).</p>
<p>For the purpose of calibration, the average R-rate per problem and the average A-rate per problem were computed by averaging the risky choices and alternations in each problem over 20 participants per problem and 100 trials per problem (for the A-rate per problem, only 99 trials per problem were used for computing the average). Later, the MSDs were calculated across the 60 problems by using the average R-rate per problem and the average A-rate per problem measures from the model and human data in the estimation set. Some researchers suggest calibrating models to the data of each participant per problem rather than to aggregate measures (Pitt and Myung, <xref ref-type="bibr" rid="B34">2002</xref>; Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>); however, the calibration to aggregate behavior is quite common in the cognitive and behavioral sciences (e.g., Anderson et al., <xref ref-type="bibr" rid="B2">2004</xref>; Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>). In fact, calibrating to aggregate measures is especially meaningful when the participant-to-participant variability in the dependent measure is small compared to the value of the dependent measure itself (Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>). In the estimation and competition sets, the standard deviations for the A-rate and R-rate were similar and very small (&#x0223C;0.1) compared to the values of the R-rate (&#x0223C;0.5) and the A-rate (&#x0223C;0.3) measures themselves. Thus, we use the average dependent measures for the purposes of model calibration in this paper.</p>
<p>For calibrating the models, both the <italic>s</italic> parameter and the <italic>d</italic> parameters were varied between 0.0 and 10.0, and the <italic>pInertia</italic> parameter was varied between 0.0 and 1.0. Although the genetic algorithm can continue to indefinitely optimize parameters, it was stopped when there was no change in the parameter values obtained for a consecutive period of 200 generations. The assumed range of variation for the <italic>pInertia</italic>, <italic>s</italic>, and <italic>d</italic> parameters, and the decision process to stop the genetic algorithm are expected to provide good optimal parameter estimates (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). Also, the large range of parameters&#x02019; variation ensures that the optimization process does not miss the minimum sum of MSDs (for more details about genetic algorithm optimization, please see Gonzalez and Dutt (<xref ref-type="bibr" rid="B20">2011</xref>).</p>
<p>We calibrated both the IBL and IBL-Inertia models to the combined R-rate and A-rate in TPT&#x02019;s estimation set. The purpose of the calibration was to obtain optimized values of <italic>d</italic> and <italic>s</italic> parameters in the IBL model and <italic>pInertia</italic>, <italic>d</italic>, and <italic>s</italic> parameters in the IBL-Inertia model. Later, keeping <italic>d</italic> and <italic>s</italic> parameters at their optimized values in the IBL-Inertia model, we varied the <italic>pInertia</italic> parameter from 0.0 to 1.0 in increments of 0.05. By only varying the <italic>pInertia</italic> parameter and keeping the other parameter values fixed at their optimized values, we were able to determine the inertia mechanism&#x02019;s full contribution in the model.</p>
<p>Table <xref ref-type="table" rid="T1">1</xref> shows the values of calibrated parameters, MSD, <italic>r</italic>, AIC, and MSE compared to baseline for IBL and IBL-Inertia models in TPT&#x02019;s estimation set. First, both models&#x02019; <italic>d</italic> and <italic>s</italic> parameters have values in the same range as those reported by Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>). Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>) reported <italic>d</italic>&#x02009;&#x0003D;&#x02009;5 and <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.5 for a MSD&#x02009;&#x0003D;&#x02009;0.0056 calibrated on R-rate using the IBL model. As documented by Lejarraga et al. (<xref ref-type="bibr" rid="B29">2012</xref>), the values of both <italic>d</italic> and <italic>s</italic> reported in Table <xref ref-type="table" rid="T1">1</xref> are high compared to the ACT-R default values of <italic>d</italic>&#x02009;&#x0003D;&#x02009;0.5 and <italic>s</italic>&#x02009;&#x0003D;&#x02009;0.25 (the default values were reported by Anderson and Lebiere (<xref ref-type="bibr" rid="B3">1998</xref>, <xref ref-type="bibr" rid="B4">2003</xref>). A high <italic>d</italic> value points to a quick decay in memory and a strong dependence on recently experienced outcomes (i.e., reliance on recency). The high <italic>s</italic> value allows the model to exhibit participant-to-participant variability in capturing the R-rate and A-rate. The <italic>pInertia</italic> value in IBL-Inertia model (&#x0003D;0.62) is high and it shows that on a trial, this model is likely to repeat its previous choice with a 62% chance. In general, the results from both models are generally good (MSDs &#x0003C;0.05 and MSEs &#x0003C;0.05), where both models perform slightly better at capturing the human A-rate than the human R-rate.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p><bold>The values of calibrated parameters for IBL and IBL-Inertia models and the MSD, <italic>r</italic>, AIC, and MSE in TPT&#x02019;s estimation set</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left">Model</th>
<th align="left">Calibrated parameters</th>
<th align="left">MSD</th>
<th align="left"><italic>r</italic></th>
<th align="left">AIC</th>
<th align="left">MSE</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">IBL (calibrated upon R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td align="left"><italic>d</italic>&#x02009;&#x0003D;&#x02009;8.31; <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.26</td>
<td align="left">0.005 (A-rate)</td>
<td align="left">0.95 (A-rate)</td>
<td align="left">&#x02212;479.2 (A-rate)</td>
<td align="left">0.0076 (A-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="left">0.016 (R-rate)</td>
<td align="left">0.94 (R-rate)</td>
<td align="left">&#x02212;546.3 (R-rate)</td>
<td align="left">0.0041 (R-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="left">0.021 (R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td align="left"/>
<td align="left"/>
<td align="left"/>
</tr>
<tr>
<td align="left">IBL-Inertia (calibrated upon<break/> R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td align="left"><italic>d</italic>&#x02009;&#x0003D;&#x02009;6.71; <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.40;<break/> <italic>pInertia</italic>&#x02009;&#x0003D;&#x02009;0.62</td>
<td align="left">0.003 (A-rate)<break/>0.008 (R-rate)<break/>0.011 (R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td align="left">0.85 (A-rate)<break/>0.92 (R-rate)</td>
<td align="left">&#x02212;561.3 (A-rate)<break/>680.0 (R-rate)</td>
<td align="left">0.0032 (A-rate)<break/>0.0010 (R-rate)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Secondly, the individual MSDs, MSEs, and AICs on the R-rate and A-rate in the IBL model are larger than those in the IBL-Inertia model. For example, in the IBL-Inertia model, the MSDs for the R-rate, A-rate, and the sum of R-rate and A-rate are consistently smaller than those in the IBL model (0.008&#x02009;&#x0003C;&#x02009;0.016, an improvement of &#x0002B;0.008; 0.003&#x02009;&#x0003C;&#x02009;0.005, an improvement of &#x0002B;0.002; and, 0.011&#x02009;&#x0003C;&#x02009;0.021, an improvement of &#x0002B;0.010). Also, the relative AIC in the IBL-Inertia model is negative (i.e., better) for both the R-rate and the A-rate. Thus, even with an extraparametric complexity (the <italic>pInertia</italic> parameter), the IBL-Inertia model performs more accurately compared to the IBL model. Although the MSE in the IBL model is larger than that in the IBL-Inertia model for both R-rate and A-rate; however, as is also shown in Table <xref ref-type="table" rid="T1">1</xref>, the IBL-Inertia model does not account for the trends in the R-rate and the A-rate across trials compared with the IBL model (the <italic>r</italic> in the IBL model is greater than that in the IBL-Inertia model for both the R-rate and the A-rate).</p>
<p>Figure <xref ref-type="fig" rid="F2">2</xref> presents the R-rate and A-rate across trials predicted by the calibrated IBL and IBL-Inertia models and that observed in human data in the TPT&#x02019;s estimation set. In general, these results reveal that both models generate good accounts for both observed risk-taking and alternation behaviors. The IBL model is able to capture the gradual decreasing trend in the A-rate as well as the slightly decreasing trend in risk-taking across trials. However, the model&#x02019;s account for the R-rate exhibit as lightly greater decrease compared with that observed in human data across increasing number of trials. Also, the model&#x02019;s account for the A-rate shows more alternations during about the first half of the trials than that observed in human data. This latter observation is likely due to the &#x0002B;30 pre-populated instances initially put in model&#x02019;s memory, which make it explore both options for a longer time and causes a higher A-rate in the first few trials. However, with increasing trials, the activation of these pre-populated instances becomes weak (as these values are not observed in the problems) and their influence on the A-rate diminishes, causing the A-rate to decrease sharply and meet the human data.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p><bold>The R-rate and A-rate across trials predicted by the IBL and IBL-Inertia models and that observed in human data in the TPT&#x02019;s estimation set</bold>.</p></caption>
<graphic xlink:href="fpsyg-03-00177-g002.tif"/>
</fig>
<p>As shown in the bottom graphs of Figure <xref ref-type="fig" rid="F2">2</xref>, the IBL-Inertia model corrects for the under-estimation and over-estimation in the R-rate and A-rate. However, because of the <italic>pInertia</italic> parameter, the model is unable to account for the initial decrease in the A-rate in the first few trials as well as the IBL model, which does so naturally. A likely reason is the high calibrated value of <italic>pInertia</italic> parameter (&#x0003D;0.62) that overshadows the effect of pre-populated instances in the first few trials. Also, it seems that the random effect of <italic>pInertia</italic> across trials causes disruptions in IBL-Inertia model&#x02019;s R-rate trends over trials. Overall, these observations explain why the IBL-Inertia accounts for overall behavior better than the IBL model, but it does not account for the trends in the R-rate and the A-rate.</p>
</sec>
<sec>
<title>Evaluating the inertia mechanism</title>
<p>Although the analyses above provide some benefits of including <italic>pInertia</italic> in the IBL model, one would like to understand these benefits more thoroughly for different values of the <italic>pInertia</italic> parameter over its entire range. If including <italic>pInertia</italic> in the IBL model is beneficial, then we should observe smaller MSDs on the R-rate and A-rate across a large part of the parameter&#x02019;s range of variation compared with the IBL model without <italic>pInertia</italic>. Also, this analysis is important because the calibrated value of <italic>pInertia</italic> in the IBL model was found to be high (&#x0003D;0.62), minimizing the role of the blending mechanism.</p>
<p>For this investigation, we used the IBL-Inertia model with its optimized parameters calibrated on the combined R-rate and A-rate measure (i.e., <italic>d</italic>&#x02009;&#x0003D;&#x02009;6.71; <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.40) and varied the <italic>pInertia</italic> parameter from 0.0 to 1.0 in increments of 0.05 in TPT&#x02019;s estimation set. Varying <italic>pInertia</italic> like so allows us to determine the range of values for which the sum of the MSDs computed on the average R-rate per problem and the average A-rate per problem are minimized.</p>
<p>Figure <xref ref-type="fig" rid="F3">3</xref> shows the MSDs for the IBL-Inertia model calibrated on the combined R-rate and A-rate as a function of <italic>pInertia</italic> values in the estimation set. It also shows the three corresponding MSDs from the original IBL model (shown as dotted lines in Figure <xref ref-type="fig" rid="F3">3</xref>) for comparison purposes (these MSDs are also reported in Table <xref ref-type="table" rid="T1">1</xref>). The MSDs for the R-rate, the A-rate, and the sum of the MSDs for the R-rate and A-rate in the IBL-Inertia model are below the corresponding MSDs in the IBL model for all values of <italic>pInertia</italic> greater than 0.05 and less than 0.90. Thus, including inertia in the IBL model and calibrating all model parameters improves the model&#x02019;s ability to account for the average R-rate and A-rate compared with the IBL model without inertia. Also, the advantages of including <italic>pInertia</italic> parameter seem to be present over a large range of this parameter&#x02019;s variation.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p><bold>The MSD for the R-rate, the MSD for the A-rate, and the MSD for the combined R-rate and A-rate for different values of <italic>pInertia</italic> parameter in IBL-Inertia model (the corresponding MSDs for the IBL model are also plotted as dotted lines for comparison)</bold>. The IBL-Inertia model used the calibrated parameters for <italic>d</italic> and <italic>s</italic> parameters (i.e., <italic>d</italic>&#x02009;&#x0003D;&#x02009;6.41 and <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.40).</p></caption>
<graphic xlink:href="fpsyg-03-00177-g003.tif"/>
</fig>
</sec>
<sec>
<title>Generalizing the IBL models to the competition set</title>
<p>A popular method of comparing models of different complexity is through models&#x02019; generalization in novel conditions (Stone, <xref ref-type="bibr" rid="B39">1977</xref>; Busemeyer and Wang, <xref ref-type="bibr" rid="B10">2000</xref>; Ahn et al., <xref ref-type="bibr" rid="B1">2008</xref>). In generalization, the calibrated models with different complexities (number of free parameters) are run in novel conditions to compare their performance. The novel conditions would minimize any advantage the model with more parameters has over the model with fewer parameters. In fact, TPT also accounted for model complexity among submitted models by generalization, i.e., by running models in the new competition set with the parameters obtained in the estimation set (Erev et al., <xref ref-type="bibr" rid="B16">2010b</xref>). We used the same procedures as used in the TPT and generalized the calibrated IBL and IBL-Inertia models to TPT&#x02019;s competition set.</p>
<p>In related research, we have claimed that the TPT&#x02019;s estimation and competition data sets are too similar, raising questions regarding the value of using the competition set for generalization (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). These similarities arise because the problems used in the estimation and competition sets were generated by using the same algorithm. However, given that the TPT competition set was collected in a new experiment, involving new problems, and involving a different set of participants from that of the estimation set, testing the models in the competition set is still a relevant exercise to determine the robustness of the models. This generalization further helps us to take into account both models&#x02019; complexity (number of parameters) and their accuracy of predictions (MSDs; Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>).</p>
<p>The IBL model and IBL-Inertia model were run in the TPT&#x02019;s competition set problems using the parameters determined in the estimation set: <italic>d</italic>&#x02009;&#x0003D;&#x02009;6.71, <italic>s</italic>&#x02009;&#x0003D;&#x02009;1.40, and <italic>pInertia</italic>&#x02009;&#x0003D;&#x02009;0.62 (the <italic>pInertia</italic> parameter is only for the IBL-Inertia model). As previously mentioned, these parameter values had resulted in the lowest MSDs on the combined R-rate and A-rate measure for the two models in the estimation set. Table <xref ref-type="table" rid="T2">2</xref> shows the values of MSD, <italic>r</italic>, and MSE for the IBL and IBL-Inertia models upon their generalization in TPT&#x02019;s competition set. The IBL-Inertia model&#x02019;s predictions resulted in overall MSDs and MSEs for the R-rate and the A-rate that were smaller than those for the IBL model. Like in the estimation set, however, the IBL-Inertia model did not account for the over trial trend in the R-rate and the A-rate compared with the IBL model (demonstrated by the <italic>r</italic> calculations). These results demonstrate that the IBL-Inertia model can generalize to new problems more accurately (in terms of average overall performance in both the A-rate and R-rate measures across problems and across trials compared with the IBL model; but the IBL-Inertia model also cannot account for trends across trials in these measures compared with the IBL model without inertia).</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p><bold>The values of MSD, <italic>r</italic>, and MSE for IBL and IBL-Inertia models upon their generalization in TPT&#x02019;s competition set</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left">Model</th>
<th align="left">MSD</th>
<th align="left"><italic>r</italic></th>
<th align="left">MSE</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">IBL</td>
<td align="left">0.011 (A-rate)</td>
<td align="left">0.96 (A-rate)</td>
<td align="left">0.010 (A-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left">0.022 (R-rate)</td>
<td align="left">0.96 (R-rate)</td>
<td align="left">0.010 (R-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left">0.033 (R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td colspan="1" align="left"/>
</tr>
<tr>
<td align="left">IBL-Inertia</td>
<td align="left">0.003 (A-rate)</td>
<td align="left">0.87 (A-rate)</td>
<td align="left">0.003 (A-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left">0.007 (R-rate)</td>
<td align="left">0.94 (R-rate)</td>
<td align="left">0.001 (R-rate)</td>
</tr>
<tr>
<td align="left"/>
<td align="left">0.010 (R-rate&#x02009;&#x0002B;&#x02009;A-rate)</td>
<td colspan="1" align="left"/>
</tr>
</tbody>
</table>
</table-wrap>
<p>Figure <xref ref-type="fig" rid="F4">4</xref> shows the R-rate and the A-rate over trials for human data, and how the IBL and IBL-Inertia models generalized in the competition set. The IBL model, upon generalization, underestimates the observed R-rate and overestimates the observed A-rate in the competition set. These patterns of under- and over-estimations are similar to those observed in the model&#x02019;s predictions in the estimation set in Figure <xref ref-type="fig" rid="F2">2</xref>. The IBL-Inertia model&#x02019;s predictions about the human R-rate and A-rate in the competition set, however, were very good with very little under- and over-estimations of the observed R-rate and A-rate curves. Furthermore, because the <italic>pInertia</italic> parameter (&#x0003D;0.62) is fixed across trials at a high value in the IBL-Inertia model, the model does not alternate as much as humans in the first few trials. As seen in the lower right graph, the IBL-Inertia model&#x02019;s A-rate starts around 40%, rather than the 85% as observed in human data. Thus, the IBL-Inertia model is not able to account for the initially high A-rate and the rapid decrease in the A-rate in the first few trials compared with the IBL model in its predictions.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>The R-rate and A-rate over trials predicted by the IBL and IBL-Inertia models upon their generalization in the competition set</bold>. The R-rate and A-rate observed in human data in the competition set are also shown.</p></caption>
<graphic xlink:href="fpsyg-03-00177-g004.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion">
<title>Discussion</title>
<p>Some computational models of DFE do not include any inertia assumptions and are still able to account for the observed risk-taking behavior (Lejarraga et al., <xref ref-type="bibr" rid="B29">2012</xref>). However, a number of recent computational models have included some form of inertia to account for observed DFE (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). Three different inertial forms have been proposed: random inertia (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>); inertia as a function of surprise determined by the differences in expected values (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>); and inertia as a function of surprise determined by the differences in blended values (Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). This research uses the particular case of random inertia in an IBL model and determines the benefits of this mechanism by considering two IBL models with and without this mechanism. We selected the random inertia form for our evaluation because of its simplistic formulation, but also because an existing IBL model with this definition accounts for DFE better than other best known models of DFE (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>).</p>
<p>Our results reveal that a simple instantiation of the inertia mechanism can be used to improve the ability of the IBL model to account for the average risk-taking (R-rate) and alternations (A-rate; based upon MSDs, MSEs, and AICs) observed in human data. However, we also find that the inclusion of random inertia does not help the model to account for the trends across trials in the R-rate and A-rate compared with the same model without inertia (based upon correlation coefficients, <italic>r</italic>). We draw our conclusions based upon model calibration and model generalization that is known to account for increased model complexity (number of parameters) in novel test environments (Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>).</p>
<p>Most current models of DFE have been successful at capturing the risk-taking behavior, but not the underlying alternations observed in repeated binary-choice; such as the tendency to repeat choices irrespective of the obtained outcome in the last trial (Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>). This observation is perhaps not a coincidence, because predicting risk-taking behavior and alternation effects simultaneously is a very challenging task (Rapoport et al., <xref ref-type="bibr" rid="B35">1997</xref>; Erev and Barron, <xref ref-type="bibr" rid="B14">2005</xref>; Estes and Maddox, <xref ref-type="bibr" rid="B17">2005</xref>). In order to overcome some of the challenges, a number of computational models have considered the inclusion of some form of inertia with some initial success (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>; Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>; Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). As can be seen in our results, the random inertia&#x02019;s inclusion into the IBL model helps the model to account for both the average A-rate and R-rate in terms of MSDs, MSEs, and AICs, but not in terms of trends in these rates over trials. Because random inertia accounts for the average A-rate and R-rate in human data, it helps to reduce the observed under-estimation and over-estimation of the observed R-rate and A-rate, respectively, which is seen in the model without the inertia mechanism. This finding might suggest that the inclusion of some form of inertia into computational models might be ecologically plausible for capturing the average risk-taking and alternation behaviors more accurately, but not for the trend in these behaviors over time.</p>
<p>Although the introduction of inertia into the IBL model generally improves the fits to the average human data (based upon MSDs, MSEs, and AICs), it is likely that few modelers may be impressed by this particular result. It is well-known that a model with more parameters (i.e., greater model complexity) can fit a dataset better than a model with fewer parameters (Pitt and Myung, <xref ref-type="bibr" rid="B34">2002</xref>). We dealt with this issue through model generalization (Stone, <xref ref-type="bibr" rid="B39">1977</xref>). The generalization helped to test models with different parametric assumptions in a novel environment (Busemeyer and Diederich, <xref ref-type="bibr" rid="B9">2009</xref>). We used these procedures and generalized the IBL and the IBL-Inertia models to the TPT&#x02019;s competition dataset to compare their performance.</p>
<p>Although the error across trials between IBL-Inertia model and human data was smaller compared to that between IBL model and human data; however, unlike the IBL model, the IBL-Inertia model did not capture the trends in the R-rate and A-rate across trials. The most likely reason is that the inertia parameter in its current formulation is a noisy selection of choices across trials, which disregards the choices derived based upon blended values. Gonzalez et al. (<xref ref-type="bibr" rid="B21">2011</xref>) had assumed an inertia formulation that was based upon surprise, where surprise was a function of the difference in blended values of the two alternatives. Perhaps, if the inertia mechanism in the model is formulated as described by Gonzalez et al. (<xref ref-type="bibr" rid="B21">2011</xref>), then the trends across trials might be better accounted for compared to that based upon inertia&#x02019;s current formulation. Overall, these observations indicate that there are many aspects still left in the literature to explore. For example, it is unclear whether people exhibit inertia after receiving both rewards and punishments. Although inertia has been defined as the tendency to repeat the last choice irrespective of the obtained outcomes (Biele et al., <xref ref-type="bibr" rid="B7">2009</xref>), it is clear that inertia needs to be defined more precisely. Some researchers have argued inertia as an irrational behavior in which individuals hold onto choices that clearly do not provide the maximizing outcome for too long (Sandri et al., <xref ref-type="bibr" rid="B38">2010</xref>). Inertia has also been portrayed as desirable, however, as it believed to be a key component of love, trust, and friendship in the real world (Cook et al., <xref ref-type="bibr" rid="B12">2005</xref>). Even when we consider inertia as we defined it in this paper, it may be the result of strong preferences for the high outcomes or the result of an apparently irrational behavior of holding on too long to non-maximizing (low) outcomes. As part of our future research, we propose to define the reasons for inertia more precisely by investigating its relationship with the exploration of alternatives due to the nature of outcomes, high or low. One way we may do this analysis is by controlling for the nature of rewards or punishments received after a decision choice and by evaluating its effects on repeating the last choice as the current decision. Also, we would like to consider the alternation behaviors of individuals depending upon the nature of rewards or punishments received by them in the last trial.</p>
<p>Finally, as part of future research, we would also like to compare the different formulations of inertia in computational models of DFE. As detailed above, there have been at least three different inertia formulations proposed: A random variation across trials (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>), a function of surprise determined by the difference in expected values (Erev et al., <xref ref-type="bibr" rid="B15">2010a</xref>), and a function of surprise determined by the differences in blended values (Gonzalez et al., <xref ref-type="bibr" rid="B21">2011</xref>). Which one of these formulations performs best in different DFE tasks? How well do these different formulations account for the over trial trends in the R-rates and A-rates? Still, how are these formulations impacted by task complexity: by the nature and number of outcomes on each alternative, and the nature of the probability distribution of outcomes on each alternative? These are also some important questions that we would like to attend to as part of future research.</p>
</sec>
<sec>
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<app-group>
<app id="A1">
<title>Appendix</title>
<sec>
<title>IBL model equations</title>
<sec>
<title>Inertia mechanism</title>
<p>A choice is made in the model in trial <italic>t</italic>&#x02009;&#x0002B;&#x02009;1 as:</p>
<p><italic>If</italic></p>
<p>The draw of a random value in the uniform distribution <italic>U</italic> (0, 1)&#x02009;&#x0003C;&#x02009;<italic>pInertia</italic>,</p>
<p><italic>Then</italic></p>
<p>Repeat the choice as made in the previous trial</p>
<p><italic>Else</italic></p>
<disp-formula id="E1"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd class="eqnarray-2"><mml:mstyle class="text"><mml:mtext>Select</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>an</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>alternative</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>with</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>the</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>highest</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>blended</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>value</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>as</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>per</mml:mtext></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd columnalign="left"><mml:mspace width="1em" class="quad"/><mml:mstyle class="text"><mml:mtext>Eq.</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>&#x000A0;A2</mml:mtext></mml:mstyle><mml:mspace width="2.77695pt" class="tmspace"/><mml:mstyle class="text"><mml:mtext>(below)</mml:mtext></mml:mstyle></mml:mtd><mml:mtd columnalign="right"><mml:mtext class="eqnarray">(A1)</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The <italic>pInertia</italic> parameter could vary between 0 and 1, and it does not change across trials or participants.</p>
</sec>
<sec>
<title>Blending and activation mechanisms</title>
<p>The blended value of alternative <italic>j</italic> is defined as</p>
<disp-formula id="E2"><label>(A2)</label><mml:math id="M4"><mml:msub><mml:mrow><mml:mtext>V</mml:mtext></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-rel">=</mml:mo><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo mathsize="big">&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-rel">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></disp-formula>
<p>Where <italic>x<sub>i</sub></italic> is the value of the observed outcome in the outcome slot of an instance <italic>i</italic> corresponding to the alternative <italic>j</italic>, and <italic>p<sub>i</sub></italic> is the probability of that instance&#x02019;s retrieval from memory (for the case of our binary-choice task in the experience condition, the value of <italic>j</italic> Eq. A2 could be either risky or safe). The blended value of an alternative is the sum of all observed outcomes <italic>x<sub>i</sub></italic> in the outcome slot of corresponding instances, weighted by the instances&#x02019; probability of retrieval.</p>
</sec>
<sec>
<title>Probability of retrieving instances</title>
<p>In any trial <italic>t</italic>, the probability of retrieving instance <italic>i</italic> from memory is a function of that instance&#x02019;s activation relative to the activation of all other instances corresponding to that alternative, given by</p>
<disp-formula id="E3"><label>(A3)</label><mml:math id="M5"><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-rel">=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mstyle class="text"><mml:mtext>e</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-bin">&#x02215;</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:munder class="msub"><mml:mrow><mml:mo mathsize="big">&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:munder><mml:msup><mml:mrow><mml:mstyle class="text"><mml:mtext>e</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-bin">&#x02215;</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:math></disp-formula>
<p>Where &#x003C4; is random noise defined as <inline-formula><mml:math id="M1"><mml:mrow><mml:mi>s</mml:mi><mml:mo class="MathClass-bin">&#x000D7;</mml:mo><mml:msqrt><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msqrt><mml:mo class="MathClass-punc">,</mml:mo></mml:mrow></mml:math></inline-formula> and <italic>s</italic> is a free noise parameter. The noise parameter <italic>s</italic> captures the imprecision of retrieving instances from memory.</p>
</sec>
<sec>
<title>Activation of instances</title>
<p>The activation of each instance in memory depends upon the <italic>activation</italic> mechanism originally proposed in ACT-R (Anderson and Lebiere, <xref ref-type="bibr" rid="B4">2003</xref>). According to this mechanism, for each trial <italic>t</italic>, <italic>activation A<sub>i,t</sub></italic> of instance is:</p>
<disp-formula id="E4"><label>(A4)</label><mml:math id="M6"><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-rel">=</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:mfenced separators="" open="(" close=")"><mml:mrow><mml:munder class="msub"><mml:mrow><mml:mo mathsize="big">&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-rel">&#x02208;</mml:mo><mml:mrow><mml:mo class="MathClass-open">{</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo class="MathClass-punc">,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi><mml:mo class="MathClass-bin">-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo class="MathClass-close">}</mml:mo></mml:mrow></mml:mrow></mml:munder><mml:msup><mml:mrow><mml:mfenced separators="" open="(" close=")"><mml:mrow><mml:mi>t</mml:mi><mml:mo class="MathClass-bin">-</mml:mo><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mo class="MathClass-bin">-</mml:mo><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfenced><mml:mo class="MathClass-bin">&#x0002B;</mml:mo><mml:mi>s</mml:mi><mml:mo class="MathClass-bin">&#x000D7;</mml:mo><mml:mstyle class="text"><mml:mtext>1n</mml:mtext></mml:mstyle><mml:mfenced separators="" open="(" close=")"><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn><mml:mo class="MathClass-bin">-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mrow></mml:mfenced></mml:math></disp-formula>
<p>Where <italic>d</italic> is a free decay parameter, and <italic>t<sub>i</sub></italic> is a previous trial when the instance <italic>i</italic> was created or its activation was reinforced due to an outcome observed in the task (the instance <italic>i</italic> is the one that has the observed outcome as the value in its outcome slot). The summation will include a number of terms that coincides with the number of times an outcome has been observed in previous trials and the corresponding instance <italic>i</italic>&#x02019;s activation that has been reinforced in memory (by encoding a timestamp of the trial <italic>t<sub>i</sub></italic>). Therefore, the activation of an instance corresponding to an observed outcome increases with the frequency of observation and with the recency of those observations. The decay parameter <italic>d</italic> affects the activation of an instance directly, as it captures the rate of forgetting or reliance on recency.</p>
</sec>
<sec>
<title>Noise in activation</title>
<p>The &#x003B3;<italic><sub>i,t</sub></italic> term is a random draw from a uniform distribution <italic>U</italic> (0, 1), and the <inline-formula><mml:math id="M2"><mml:mrow><mml:mi>s</mml:mi><mml:mo class="MathClass-bin">&#x000D7;</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:mstyle class="text"><mml:mtext>(</mml:mtext></mml:mstyle><mml:mfrac><mml:mrow><mml:mn>1</mml:mn><mml:mo class="MathClass-bin">-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">.</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mstyle class="text"><mml:mtext>)</mml:mtext></mml:mstyle></mml:mrow></mml:math></inline-formula> term represents Gaussian noise important for capturing the variability of human behavior.</p>
</sec>
<sec>
<title>Definition of akaike information criterion</title>
<disp-formula id="E5"><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd class="eqnarray-1"><mml:mstyle class="text"><mml:mtext>AIC&#x000A0;=</mml:mtext></mml:mstyle></mml:mtd><mml:mtd columnalign="left"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo class="MathClass-bin">&#x0002B;</mml:mo><mml:mn>2</mml:mn><mml:mo class="MathClass-bin">&#x02217;</mml:mo><mml:mi>k</mml:mi></mml:mtd><mml:mtd columnalign="left"><mml:mtext class="eqnarray">(A5)</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd class="eqnarray-1"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo class="MathClass-rel">=</mml:mo></mml:mtd><mml:mtd columnalign="left"><mml:mi>t</mml:mi><mml:mo class="MathClass-bin">&#x02217;</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:mfrac><mml:mrow><mml:mstyle class="text"><mml:mtext>SSE</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mtd><mml:mtd class="eqnarray-4"><mml:mtext class="eqnarray">(A6)</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd class="eqnarray-1"><mml:mstyle class="text"><mml:mtext>SSE</mml:mtext></mml:mstyle><mml:mo class="MathClass-rel">=</mml:mo></mml:mtd><mml:mtd class="eqnarray-2"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo mathsize="big">&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo class="MathClass-rel">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:munderover><mml:msup><mml:mrow><mml:mfenced separators="" open="(" close=")"><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext>model</mml:mtext></mml:mstyle><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo class="MathClass-bin">-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext>human</mml:mtext></mml:mstyle><mml:mo class="MathClass-punc">,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd><mml:mtd class="eqnarray-4"><mml:mtext class="eqnarray">(A7)</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where, <italic>G</italic><sup>2</sup> is defined as the lack of fit between model and human data (Gonzalez and Dutt, <xref ref-type="bibr" rid="B20">2011</xref>). Furthermore, the <italic>x</italic><sub>model, <italic>i</italic></sub> and <italic>x</italic><sub>human,<italic>i</italic></sub> refer to the average dependent measure (e.g., average R-rate or A-rate) in the model and human data over <italic>t</italic> trials of a task (<italic>t</italic>&#x02009;&#x0003D;&#x02009;100 for the R-rate and <italic>t</italic>&#x02009;&#x0003D;&#x02009;99 for the A-rate). The average in the dependent measure (R-rate or A-rate) has been taken over all problems and participants. The SSE is the sum of squared errors between human and model datasets that is calculated for the average dependent measure (A-rate or R-rate). The mean-squared error (MSE) is defined as SSE/100 for the R-rate measure and SSE/99 for the A-rate measure. The <italic>t</italic> is the number of trials in the task, and <italic>k</italic> is the number of parameters in the model. The AIC in its formulation incorporates both the effect of an MSD (the <italic>G</italic><sup>2</sup> term) as well as the number of free parameters in a model (the 2&#x02009;&#x02217;&#x02009;<italic>k</italic> term). The smaller the value of AIC, the better the respective model is.</p>
</sec>
</sec>
</app>
</app-group>
<ack>
<p>This research is supported by the Defense Threat Reduction Agency (DTRA) grant number: HDTRA1-09-1-0053 to Dr. Cleotilde Gonzalez. We thank Ms. Hau-Yu Wong and Dr. Noam Ben-Asher of the Dynamic Decision Making Laboratory for their help in proofreading this manuscript and providing insightful comments. We would also like to thank Dr. Ido Erev of the Technion-Israel Institute of Technology for making the data from the Technion Prediction Tournament available.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahn</surname> <given-names>W. Y.</given-names></name> <name><surname>Busemeyer</surname> <given-names>J. R.</given-names></name> <name><surname>Wagenmakers</surname> <given-names>E. J.</given-names></name> <name><surname>Stout</surname> <given-names>J. C.</given-names></name></person-group> (<year>2008</year>). <article-title>Comparison of decision learning models using the generalization criterion method</article-title>. <source>Cogn. Sci.</source> <volume>32</volume>, <fpage>1376</fpage>&#x02013;<lpage>1402</lpage>.<pub-id pub-id-type="doi">10.1080/03640210802352992</pub-id><pub-id pub-id-type="pmid">21585458</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name> <name><surname>Bothell</surname> <given-names>D.</given-names></name> <name><surname>Byrne</surname> <given-names>M. D.</given-names></name> <name><surname>Douglass</surname> <given-names>S.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name> <name><surname>Qin</surname> <given-names>Y.</given-names></name></person-group> (<year>2004</year>). <article-title>An integrated theory of the mind</article-title>. <source>Psychol. Rev.</source> <volume>111</volume>, <fpage>1036</fpage>&#x02013;<lpage>1060</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.111.4.1036</pub-id><pub-id pub-id-type="pmid">15482072</pub-id></citation></ref>
<ref id="B3"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>1998</year>). <source>The Atomic Components of Thought</source>. <publisher-loc>Mahwah, NJ</publisher-loc>: <publisher-name>Erlbaum</publisher-name>.</citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. R.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2003</year>). <article-title>The Newell test for a theory of mind</article-title>. <source>Behav. Brain Sci.</source> <volume>26</volume>, <fpage>587</fpage>&#x02013;<lpage>639</lpage>.<pub-id pub-id-type="doi">10.1017/S0140525X0300013X</pub-id><pub-id pub-id-type="pmid">15179936</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barron</surname> <given-names>G.</given-names></name> <name><surname>Erev</surname> <given-names>I.</given-names></name></person-group> (<year>2003</year>). <article-title>Small feedback-based decisions and their limited correspondence to description-based decisions</article-title>. <source>J. Behav. Decis. Mak.</source> <volume>16</volume>, <fpage>215</fpage>&#x02013;<lpage>233</lpage>.<pub-id pub-id-type="doi">10.1002/bdm.443</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barron</surname> <given-names>G.</given-names></name> <name><surname>Leider</surname> <given-names>S.</given-names></name> <name><surname>Stack</surname> <given-names>J.</given-names></name></person-group> (<year>2008</year>). <article-title>The effect of safe experience on a warnings&#x02019; impact: sex, drugs, and rock-n-roll</article-title>. <source>Organ. Behav. Hum. Decis. Process.</source> <volume>106</volume>, <fpage>125</fpage>&#x02013;<lpage>142</lpage>.<pub-id pub-id-type="doi">10.1016/j.obhdp.2007.11.002</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Biele</surname> <given-names>G.</given-names></name> <name><surname>Erev</surname> <given-names>I.</given-names></name> <name><surname>Ert</surname> <given-names>E.</given-names></name></person-group> (<year>2009</year>). <article-title>Learning, risk attitude and hot stoves in restless bandit problems</article-title>. <source>J. Math. Psychol.</source> <volume>53</volume>, <fpage>155</fpage>&#x02013;<lpage>167</lpage>.<pub-id pub-id-type="doi">10.1016/j.jmp.2008.05.006</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>B&#x000F6;rgers</surname> <given-names>T.</given-names></name> <name><surname>Sarin</surname> <given-names>R.</given-names></name></person-group> (<year>2000</year>). <article-title>Naive reinforcement learning with endogenous aspirations</article-title>. <source>Int. Econ. Rev.</source> <volume>41</volume>, <fpage>921</fpage>&#x02013;<lpage>950</lpage>.<pub-id pub-id-type="doi">10.1111/1468-2354.00090</pub-id></citation></ref>
<ref id="B9"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Busemeyer</surname> <given-names>J. R.</given-names></name> <name><surname>Diederich</surname> <given-names>A.</given-names></name></person-group> (<year>2009</year>). <source>Cognitive Modeling</source>. <publisher-loc>Thousand Oaks</publisher-loc>: <publisher-name>Sage Publications, Inc</publisher-name>.</citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Busemeyer</surname> <given-names>J. R.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name></person-group> (<year>2000</year>). <article-title>Model comparisons and model selections based on the generalization criterion methodology</article-title>. <source>J. Math. Psychol.</source> <volume>44</volume>, <fpage>171</fpage>&#x02013;<lpage>189</lpage>.<pub-id pub-id-type="doi">10.1006/jmps.1999.1282</pub-id><pub-id pub-id-type="pmid">10733863</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Liu</surname> <given-names>S.-Y.</given-names></name> <name><surname>Chen</surname> <given-names>C.-H.</given-names></name> <name><surname>Lee</surname> <given-names>Y.-S.</given-names></name></person-group> (<year>2011</year>). <article-title>Bounded memory, inertia, sampling and weighting model for market entry games</article-title>. <source>Games</source> <volume>2</volume>, <fpage>187</fpage>&#x02013;<lpage>199</lpage>.<pub-id pub-id-type="doi">10.3390/g2010187</pub-id></citation></ref>
<ref id="B12"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Cook</surname> <given-names>K. S.</given-names></name> <name><surname>Hardin</surname> <given-names>R.</given-names></name> <name><surname>Levi</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <source>Cooperation without Trust?</source> <publisher-loc>New York</publisher-loc>: <publisher-name>Russell Sage Foundation Publications</publisher-name>.</citation></ref>
<ref id="B13"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Dutt</surname> <given-names>V.</given-names></name> <name><surname>Yamaguchi</surname> <given-names>M.</given-names></name> <name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Proctor</surname> <given-names>R. W.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;An instance-based learning model of stimulus-response compatibility effects in mixed location-relevant and location-irrelevant tasks,&#x0201D;</article-title> in <source>Proceedings of the 9th International Conference on Cognitive Modeling &#x02013; ICCM2009</source>, eds <person-group person-group-type="editor"><name><surname>Howes</surname> <given-names>A.</given-names></name> <name><surname>Peebles</surname> <given-names>D.</given-names></name> <name><surname>Cooper</surname> <given-names>R.</given-names></name></person-group> (<publisher-loc>Manchester</publisher-loc>: <publisher-name>University of Huddersfield</publisher-name>).</citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Erev</surname> <given-names>I.</given-names></name> <name><surname>Barron</surname> <given-names>G.</given-names></name></person-group> (<year>2005</year>). <article-title>On adaptation, maximization and reinforcement learning among cognitive strategies</article-title>. <source>Psychol. Rev.</source> <volume>112</volume>, <fpage>912</fpage>&#x02013;<lpage>931</lpage>.<pub-id pub-id-type="doi">10.1037/0033-295X.112.4.912</pub-id><pub-id pub-id-type="pmid">16262473</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Erev</surname> <given-names>I.</given-names></name> <name><surname>Ert</surname> <given-names>E.</given-names></name> <name><surname>Roth</surname> <given-names>A. E.</given-names></name></person-group> (<year>2010a</year>). <article-title>A choice prediction competition for market entry games: an introduction</article-title>. <source>Games</source> <volume>1</volume>, <fpage>117</fpage>&#x02013;<lpage>136</lpage>.<pub-id pub-id-type="doi">10.3390/g1020117</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Erev</surname> <given-names>I.</given-names></name> <name><surname>Ert</surname> <given-names>E.</given-names></name> <name><surname>Roth</surname> <given-names>A. E.</given-names></name> <name><surname>Haruvy</surname> <given-names>E.</given-names></name> <name><surname>Herzog</surname> <given-names>S. M.</given-names></name> <name><surname>Hau</surname> <given-names>R.</given-names></name> <name><surname>Hertwig</surname> <given-names>R.</given-names></name> <name><surname>Stewart</surname> <given-names>T.</given-names></name> <name><surname>West</surname> <given-names>R.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2010b</year>). <article-title>A choice prediction competition: choices from experience and from description</article-title>. <source>J. Behav. Decis. Mak.</source> <volume>23</volume>, <fpage>15</fpage>&#x02013;<lpage>47</lpage>.<pub-id pub-id-type="doi">10.1002/bdm.683</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Estes</surname> <given-names>W. K.</given-names></name> <name><surname>Maddox</surname> <given-names>W. T.</given-names></name></person-group> (<year>2005</year>). <article-title>Risks of drawing inferences about cognitive processes from model fits to individual versus average performance</article-title>. <source>Psychon. Bull. Rev.</source> <volume>12</volume>, <fpage>403</fpage>&#x02013;<lpage>408</lpage>.<pub-id pub-id-type="doi">10.3758/BF03193784</pub-id><pub-id pub-id-type="pmid">16235625</pub-id></citation></ref>
<ref id="B18"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Gladwell</surname> <given-names>M.</given-names></name></person-group> (<year>2007</year>). <source>Blink</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Back Bay Books</publisher-name>.</citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Best</surname> <given-names>B. J.</given-names></name> <name><surname>Healy</surname> <given-names>A. F.</given-names></name> <name><surname>Bourne</surname> <given-names>L. E.</given-names> <suffix>Jr.</suffix></name> <name><surname>Kole</surname> <given-names>J. A.</given-names></name></person-group> (<year>2010</year>). <article-title>A cognitive modeling account of simultaneous learning and fatigue effects</article-title>. <source>Cogn. Syst. Res.</source> <volume>12</volume>, <fpage>19</fpage>&#x02013;<lpage>32</lpage>.<pub-id pub-id-type="doi">10.1016/j.cogsys.2010.06.004</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Dutt</surname> <given-names>V.</given-names></name></person-group> (<year>2011</year>). <article-title>Instance-based learning: integrating sampling and repeated decisions from experience</article-title>. <source>Psychol. Rev.</source> <volume>118</volume>, <fpage>523</fpage>&#x02013;<lpage>551</lpage>.<pub-id pub-id-type="doi">10.1037/a0024558</pub-id><pub-id pub-id-type="pmid">21806307</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Dutt</surname> <given-names>V.</given-names></name> <name><surname>Lejarraga</surname> <given-names>T.</given-names></name></person-group> (<year>2011</year>). <article-title>A loser can be a winner: comparison of two instance-based learning models in a market entry competition</article-title>. <source>Games</source> <volume>2</volume>, <fpage>136</fpage>&#x02013;<lpage>162</lpage>.<pub-id pub-id-type="doi">10.3390/g2010136</pub-id></citation></ref>
<ref id="B22"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2005</year>). <article-title>&#x0201C;Instance-based cognitive models of decision making,&#x0201D;</article-title> in <source>Transfer of Knowledge in Economic Decision-Making</source>, eds <person-group person-group-type="editor"><name><surname>Zizzo</surname> <given-names>D.</given-names></name> <name><surname>Courakis</surname> <given-names>A.</given-names></name></person-group> (<publisher-loc>New York</publisher-loc>: <publisher-name>Palgrave Macmillan</publisher-name>), <fpage>148</fpage>&#x02013;<lpage>165</lpage>.</citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Lerch</surname> <given-names>F. J.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2003</year>). <article-title>Instance-based learning in real-time dynamic decision making</article-title>. <source>Cogn. Sci.</source> <volume>27</volume>, <fpage>591</fpage>&#x02013;<lpage>635</lpage>.<pub-id pub-id-type="doi">10.1207/s15516709cog2704_2</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hertwig</surname> <given-names>R.</given-names></name> <name><surname>Erev</surname> <given-names>I.</given-names></name></person-group> (<year>2009</year>). <article-title>The description-experience gap in risky choice</article-title>. <source>Trends Cogn. Sci. (Regul. Ed.)</source> <volume>13</volume>, <fpage>517</fpage>&#x02013;<lpage>523</lpage>.<pub-id pub-id-type="doi">10.1016/j.tics.2009.09.004</pub-id><pub-id pub-id-type="pmid">19836292</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hodgkinson</surname> <given-names>G. P.</given-names></name></person-group> (<year>1997</year>). <article-title>Cognitive inertia in a turbulent market: the case of UK residential estate agents</article-title>. <source>J. Manage. Stud.</source> <volume>34</volume>, <fpage>921</fpage>&#x02013;<lpage>945</lpage>.<pub-id pub-id-type="doi">10.1111/1467-6486.00078</pub-id></citation></ref>
<ref id="B26"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Kline</surname> <given-names>R. B.</given-names></name></person-group> (<year>2004</year>). <source>Principles and Practice of Structural Equation Modeling (Methodology in the Social Sciences)</source>, <edition>2nd Edn</edition>. <publisher-loc>New York</publisher-loc>: <publisher-name>The Guilford Press</publisher-name>.</citation></ref>
<ref id="B27"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>1999</year>). <article-title>&#x0201C;Blending,&#x0201D;</article-title> in <conf-name>Proceedings of the Sixth ACT-R Workshop</conf-name> (<conf-loc>Fairfax, VA</conf-loc>: <conf-sponsor>George Mason University</conf-sponsor>).</citation></ref>
<ref id="B28"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Lebiere</surname> <given-names>C.</given-names></name> <name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Martin</surname> <given-names>M.</given-names></name></person-group> (<year>2007</year>). <article-title>&#x0201C;Instance-based decision making model of repeated binary choice,&#x0201D;</article-title> in <conf-name>Proceedings of the 8th International Conference on Cognitive Modeling</conf-name>, <conf-loc>Ann Arbor, MI</conf-loc>.</citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lejarraga</surname> <given-names>T.</given-names></name> <name><surname>Dutt</surname> <given-names>V.</given-names></name> <name><surname>Gonzalez</surname> <given-names>C.</given-names></name></person-group> (<year>2012</year>). <article-title>Instance-based learning: a general model of repeated binary choice</article-title>. <source>J. Behav. Decis. Mak.</source> <volume>25</volume>, <fpage>143</fpage>&#x02013;<lpage>153</lpage>.<pub-id pub-id-type="doi">10.1002/bdm.722</pub-id></citation></ref>
<ref id="B30"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Loehlin</surname> <given-names>J. C.</given-names></name></person-group> (<year>2003</year>). <source>Latent Variable Models: An Introduction to Factor, Path, and Structural Equation Analysis</source>. <publisher-loc>Mahwah, NJ</publisher-loc>: <publisher-name>Lawrence Erlbaum</publisher-name>.</citation></ref>
<ref id="B31"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Martin</surname> <given-names>M. K.</given-names></name> <name><surname>Gonzalez</surname> <given-names>C.</given-names></name> <name><surname>Lebiere</surname> <given-names>C.</given-names></name></person-group> (<year>2004</year>). <article-title>&#x0201C;Learning to make decisions in dynamic environments: ACT-R plays the beer game,&#x0201D;</article-title> in <conf-name>Proceedings of the Sixth International Conference on Cognitive Modeling</conf-name> (<conf-loc>Pittsburgh, PA</conf-loc>: <conf-sponsor>Carnegie Mellon University</conf-sponsor>), <fpage>178</fpage>&#x02013;<lpage>183</lpage>.</citation></ref>
<ref id="B32"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Maruyama</surname> <given-names>G. M.</given-names></name></person-group> (<year>1997</year>). <source>Basics of Structural Equation Modeling</source>. <publisher-loc>Thousand Oaks, CA</publisher-loc>: <publisher-name>Sage Publications, Inc</publisher-name>.</citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nevo</surname> <given-names>I.</given-names></name> <name><surname>Erev</surname> <given-names>I.</given-names></name></person-group> (<year>2012</year>). <article-title>On surprise, change, and the effect of recent outcomes</article-title>. <source>Front. Psychol.</source> <volume>3</volume>:<fpage>24</fpage>.<pub-id pub-id-type="doi">10.3389/fpsyg.2012.00024</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pitt</surname> <given-names>M. A.</given-names></name> <name><surname>Myung</surname> <given-names>I. J.</given-names></name></person-group> (<year>2002</year>). <article-title>When a good fit can be bad</article-title>. <source>Trends Cogn. Sci. (Regul. Ed.)</source> <volume>6</volume>, <fpage>421</fpage>&#x02013;<lpage>425</lpage>.<pub-id pub-id-type="doi">10.1016/S1364-6613(02)01964-2</pub-id><pub-id pub-id-type="pmid">12413575</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rapoport</surname> <given-names>A.</given-names></name> <name><surname>Erev</surname> <given-names>I.</given-names></name> <name><surname>Abraham</surname> <given-names>E. V.</given-names></name> <name><surname>Olson</surname> <given-names>D. E.</given-names></name></person-group> (<year>1997</year>). <article-title>Randomization and adaptive learning in a simplified poker game</article-title>. <source>Organ. Behav. Hum. Decis. Process.</source> <volume>69</volume>, <fpage>31</fpage>&#x02013;<lpage>49</lpage>.<pub-id pub-id-type="doi">10.1006/obhd.1996.2670</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Reger</surname> <given-names>R. K.</given-names></name> <name><surname>Palmer</surname> <given-names>T. B.</given-names></name></person-group> (<year>1996</year>). <article-title>Managerial categorization of competitors: using old maps to navigate new environments</article-title>. <source>Organ. Sci.</source> <volume>7</volume>, <fpage>22</fpage>&#x02013;<lpage>39</lpage>.<pub-id pub-id-type="doi">10.1287/orsc.7.1.22</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Samuelson</surname> <given-names>L.</given-names></name></person-group> (<year>1994</year>). <article-title>Stochastic stability in games with alternative best replies</article-title>. <source>J. Econ. Theory</source> <volume>64</volume>, <fpage>35</fpage>&#x02013;<lpage>65</lpage>.<pub-id pub-id-type="doi">10.1006/jeth.1994.1053</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sandri</surname> <given-names>S.</given-names></name> <name><surname>Schade</surname> <given-names>C.</given-names></name> <name><surname>Mu&#x000DF;hoff</surname> <given-names>O.</given-names></name> <name><surname>Odening</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Holding on for too long? An experimental study on inertia in entrepreneurs&#x02019; and non-entrepreneurs&#x02019; disinvestment choices</article-title>. <source>J. Econ. Behav. Organ.</source> <volume>76</volume>, <fpage>30</fpage>&#x02013;<lpage>44</lpage>.<pub-id pub-id-type="doi">10.1016/j.jebo.2010.02.011</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stone</surname> <given-names>M.</given-names></name></person-group> (<year>1977</year>). <article-title>Asymptotics for and against cross-validation</article-title>. <source>Biometrika</source> <volume>64</volume>, <fpage>29</fpage>&#x02013;<lpage>35</lpage>.<pub-id pub-id-type="doi">10.1093/biomet/64.1.29</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tripsas</surname> <given-names>M.</given-names></name> <name><surname>Gavetti</surname> <given-names>G.</given-names></name></person-group> (<year>2000</year>). <article-title>Capabilities, cognition, and inertia: evidence from digital imaging</article-title>. <source>Strateg. Manage. J.</source> <volume>21</volume>, <fpage>1147</fpage>&#x02013;<lpage>1161</lpage>.<pub-id pub-id-type="doi">10.1002/1097-0266(200010/11)21:10/11&#x0003C;1147::AID-SMJ128&#x0003E;3.3.CO;2-I</pub-id></citation></ref>
</ref-list>
</back>
</article>