<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychol.</journal-id>
<journal-title>Frontiers in Psychology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychol.</abbrev-journal-title>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyg.2021.667802</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Three Sample Estimates of Fraction of Missing Information From Full Information Maximum Likelihood</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Chen</surname> <given-names>Lihan</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1291519/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Savalei</surname> <given-names>Victoria</given-names></name>
<uri xlink:href="http://loop.frontiersin.org/people/192294/overview"/>
</contrib>
</contrib-group>
<aff><institution>Psychology Department, University of British Columbia</institution>, <addr-line>Vancouver, BC</addr-line>, <country>Canada</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Prathiba Natesan Batley, Brunel University London, United Kingdom</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Alexander Robitzsch, IPN&#x02014;Leibniz Institute for Science and Mathematics Education, Germany; Zhenqiu Laura Lu, University of Georgia, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Lihan Chen <email>bill.lihan&#x00040;psych.ubc.ca</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Quantitative Psychology and Measurement, a section of the journal Frontiers in Psychology</p></fn></author-notes>
<pub-date pub-type="epub">
<day>26</day>
<month>08</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>12</volume>
<elocation-id>667802</elocation-id>
<history>
<date date-type="received">
<day>14</day>
<month>02</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>27</day>
<month>07</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Chen and Savalei.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Chen and Savalei</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract><p>In missing data analysis, the reporting of missing rates is insufficient for the readers to determine the impact of missing data on the efficiency of parameter estimates. A more diagnostic measure, the fraction of missing information (FMI), shows how the standard errors of parameter estimates increase from the information loss due to ignorable missing data. FMI is well-known in the multiple imputation literature (Rubin, <xref ref-type="bibr" rid="B16">1987</xref>), but it has only been more recently developed for full information maximum likelihood (Savalei and Rhemtulla, <xref ref-type="bibr" rid="B20">2012</xref>). Sample FMI estimates using this approach have since then been made accessible as part of the <italic>lavaan</italic> package (Rosseel, <xref ref-type="bibr" rid="B14">2012</xref>) in the <italic>R</italic> statistical programming language. However, the properties of FMI estimates at finite sample sizes have not been the subject of comprehensive investigation. In this paper, we present a simulation study on the properties of three sample FMI estimates from FIML in two common models in psychology, regression and two-factor analysis. We summarize the performance of these FMI estimates and make recommendations on their application.</p></abstract>
<kwd-group>
<kwd>missing data</kwd>
<kwd>full information maximum likelihood</kwd>
<kwd>regression</kwd>
<kwd>factor analysis</kwd>
<kwd>fraction of missing information</kwd>
</kwd-group>
<contract-sponsor id="cn001">Natural Sciences and Engineering Research Council of Canada<named-content content-type="fundref-id">10.13039/501100000038</named-content></contract-sponsor>
<counts>
<fig-count count="9"/>
<table-count count="6"/>
<equation-count count="6"/>
<ref-count count="26"/>
<page-count count="16"/>
<word-count count="9702"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Missing data can often occur in psychological research, whether due to dropouts in longitudinal studies, skipped questions in surveys, or equipment limitations (e.g., eye-trackers failing to capture certain eye movements). Missing data can also result from the strategic use of a planned missing data design (Graham et al., <xref ref-type="bibr" rid="B5">2006</xref>). For example, in studies with prohibitively lengthy questionnaires, researchers may want to administer only a subset of the questions to each participant, and treat the rest as missing, in order to reduce survey fatigue. Traditionally, psychological researchers who encounter missing data would most typically apply <italic>ad-hoc</italic> missing data techniques, such as <italic>listwise deletion</italic> or <italic>pairwise deletion</italic>. These methods are advantageous in their expedience, but will often lead to inconsistent estimates and power loss, resulting in bad inference. In the past decade or so, modern missing data techniques with better statistical properties, such as <italic>full information maximum likelihood</italic> (FIML; Allison, <xref ref-type="bibr" rid="B1">1987</xref>; Muth&#x000E9;n et al., <xref ref-type="bibr" rid="B10">1987</xref>; Arbuckle, <xref ref-type="bibr" rid="B2">1996</xref>) and <italic>multiple imputation</italic> (MI; Rubin, <xref ref-type="bibr" rid="B16">1987</xref>), have become increasingly accessible via computer programs like <italic>lavaan</italic> (Rosseel, <xref ref-type="bibr" rid="B14">2012</xref>) and <italic>mice</italic> (van Buuren and Groothuis-Oudshoorn, <xref ref-type="bibr" rid="B24">2011</xref>), available in R (R Core Team, <xref ref-type="bibr" rid="B12">2019</xref>).</p>
<p>Rubin (<xref ref-type="bibr" rid="B15">1976</xref>) defined three types of missing data mechanisms: <italic>missing completely at random</italic> (MCAR), when the reason data are missing is independent of any variable in the dataset, whether missing or observed; <italic>missing at random</italic> (MAR), when the reason data are missing is dependent only on observed data known as <italic>conditioning variables</italic>; and <italic>missing not at random</italic> (MNAR), when the reason data are missing is dependent on data that are not observed, even after conditioning on observed data. When data are missing in psychological studies, researchers typically assume ignorable missing data, which refers to MCAR or MAR missing mechanism, plus an additional technical assumption that the parameters describing the missing mechanism are independent of the model parameters (Little and Rubin, <xref ref-type="bibr" rid="B9">2002</xref>). This article will focus on the impact of ignorable MAR mechanism on FIML parameter estimates. When describing the possible impact of missing data, researchers typically report only the rate of missing data per variable, in the dataset as a whole, or the proportion of incomplete cases. However, the missing rates alone are insufficient to determine the efficiency of the estimates under a general MAR mechanism.</p>
<p>The MAR assumption is very general, and there is a wide range of specific missing mechanisms that can generate MAR data. For example, the relationship between the conditioning variable and the missing probability can be monotonic or not monotonic (linear vs nonlinear MAR; see Collins et al., <xref ref-type="bibr" rid="B4">2001</xref>; Yoo, <xref ref-type="bibr" rid="B25">2009</xref>; Savalei and Rhemtulla, <xref ref-type="bibr" rid="B21">2017</xref>). Further, the missing mechanism can be more deterministic or more probabilistic (strong vs weak MAR; see Yucel et al., <xref ref-type="bibr" rid="B26">2011</xref>; Sullivan et al., <xref ref-type="bibr" rid="B23">2018</xref>; Chen et al., <xref ref-type="bibr" rid="B3">2020</xref>). When MAR mechanisms differ, the statistical properties of the FIML parameter estimates can differ even when the missing rates are the same. Although MCAR, a special case of MAR, does not involve any conditioning variables, thus avoiding some of these issues, it could still lead to parameter estimates with varying efficiency when the number of missing data patterns differ (e.g., Savalei and Bentler, <xref ref-type="bibr" rid="B18">2005</xref>; Savalei and Falk, <xref ref-type="bibr" rid="B19">2014</xref>).</p>
<p>Differential performance of FIML estimates under different MAR mechanisms can translate into differences in efficiency, and therefore different power when used to perform null hypothesis tests. In light of the replication crisis, there has been a renewed call for researchers to better understand the power of their analyses. The impact of missing data on power can only be understood if we adopt a more informative diagnostic measure than missingness rates. In this article, we argue that the appropriate diagnostic measure is a quantity known as the <italic>fraction of missing information</italic> (FMI). We evaluate the performance of several estimates of FMI in the context of FIML estimation within the structural equation modeling (SEM) framework. This framework is quite general and includes regression models, path analysis, confirmatory factor analysis, and general linear models involving any combination of observed and latent variables.</p>
<p>SEMs are usually fit as <italic>mean and covariance structure models</italic>. Given a set of <italic>p</italic> variables, a particular model implies that the mean and covariance structure of these variables is given by &#x003BC;(&#x003B8;) and &#x003A3;(&#x003B8;), where &#x003B8; is a <italic>q</italic> &#x000D7; 1 vector of model parameters. Under the assumptions that the missing data mechanism is MAR and that data are multivariate normally distributed, the FIML parameter estimates <inline-formula><mml:math id="M1"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> can be obtained by maximizing the observed data normal-theory log-likelihood log<italic>L</italic>(&#x003B8;|<italic>Y</italic>), given by (Little and Rubin, <xref ref-type="bibr" rid="B8">1987</xref>):</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x003C0;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>&#x003A3;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:msubsup><mml:mrow><mml:mo>&#x003A3;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>N</italic> is the sample size, <italic>p</italic><sub><italic>i</italic></sub> is the number of variables observed for case <italic>i</italic>. &#x003BC;<sub><italic>i</italic></sub>(&#x003B8;) is the <italic>p</italic><sub><italic>i</italic></sub> &#x000D7; 1 model-implied vector of means, and &#x003A3;<sub><italic>i</italic></sub>(&#x003B8;) is the <italic>p</italic><sub><italic>i</italic></sub> &#x000D7; <italic>p</italic><sub><italic>i</italic></sub> model-implied covariance matrix, both for the variables that are not missing in case <italic>i</italic>.</p>
<p>Parameter estimates <inline-formula><mml:math id="M4"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> are asymptotically fully efficient if the distributional assumptions are met. However, they are not as efficient as they would have been had there been no missing data. If for the same sample of <italic>N</italic> observations, we write <italic>X</italic> &#x0003D; (<italic>Y, Z</italic>), where <italic>Y</italic> is the observed data, <italic>Z</italic> is the missing data, and <italic>X</italic> is the complete data, then the hypothetical complete data parameter estimates <inline-formula><mml:math id="M5"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">ML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> obtained by maximizing <italic>L</italic>(&#x003B8;|<italic>X</italic>) would be more efficient than the FIML parameter estimates <inline-formula><mml:math id="M6"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> to the extent that the missing data <italic>Z</italic> contains information about &#x003B8;.</p>
<sec>
<title>1.1. Fraction of Missing Information</title>
<p>The FMI measure quantifies the amount of information missing in the estimation of a particular parameter (i.e., element of &#x003B8;) by considering the drop in efficiency when that parameter is estimated from the observed data <italic>Y</italic> rather than from the hypothetical complete data <italic>X</italic>. Although the FMI computation was originally introduced in the context of MI (Rubin, <xref ref-type="bibr" rid="B16">1987</xref>), its theoretical roots can be traced back to the <italic>missing information principle</italic>, laid out by Orchard and Woodbury (<xref ref-type="bibr" rid="B11">1972</xref>). This principle states that the likelihood of the complete data, <italic>L</italic>(&#x003B8;|<italic>X</italic>), can be factored into the likelihood of the observed data, <italic>L</italic>(&#x003B8;|<italic>Y</italic>), and the density of the missing data given the observed data, <italic>f</italic>(<italic>Z</italic>|&#x003B8;, <italic>Y</italic>):</p>
<disp-formula id="E3"><label>(2)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi><mml:mo>|</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>It follows that the first derivative of the complete data log-likelihood, or the <italic>score</italic> vector, can be written as</p>
<disp-formula id="E4"><label>(3)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi><mml:mo>|</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The information matrix can be defined as the covariance matrix of the score vector (Rao, <xref ref-type="bibr" rid="B13">1973</xref>). Let <inline-formula><mml:math id="M9"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext></mml:mstyle><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M10"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext></mml:mstyle><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B8;</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, and <inline-formula><mml:math id="M11"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi><mml:mo>|</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, the missing information principle can then be stated as</p>
<disp-formula id="E5"><label>(4)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mo>|</mml:mo><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>When the model is correctly specified and the multivariate normality assumption is met, it is the case that <inline-formula><mml:math id="M13"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>&#x1D53C;</mml:mi><mml:mstyle mathsize="1.19em"><mml:mrow></mml:mrow></mml:mstyle><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x02202;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext></mml:mstyle><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>|</mml:mo><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mi>&#x003B8;</mml:mi><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mstyle mathsize="1.19em"><mml:mrow></mml:mrow></mml:mstyle></mml:math></inline-formula>, i.e., the information matrix for complete data is also the negative expectation of the second derivative of the log-likelihood of the data (Rao, <xref ref-type="bibr" rid="B13">1973</xref>). Similarly, <inline-formula><mml:math id="M14"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>&#x1D53C;</mml:mi><mml:mstyle mathsize="1.19em"><mml:mrow></mml:mrow></mml:mstyle><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x02202;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">log</mml:mtext></mml:mstyle><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi><mml:mo>|</mml:mo><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mi>&#x003B8;</mml:mi><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mstyle mathsize="1.19em"><mml:mrow></mml:mrow></mml:mstyle></mml:math></inline-formula> is the information matrix for observed data. The quantity <italic>J</italic><sub><italic>X</italic>|<italic>Y</italic></sub> is known as <italic>missing information</italic>. Equation (4) states that complete information can be written as a sum of observed information and missing information.</p>
<p>When the distributional assumptions are met and the model is correct, the asymptotic covariance matrix of the FIML estimates <inline-formula><mml:math id="M15"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> (multiplied by <inline-formula><mml:math id="M16"><mml:msqrt><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msqrt></mml:math></inline-formula> to stabilize the asymptotic distribution<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref>) is given by <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>, whereas the asymptotic covariance matrix of the hypothetical ML estimates <inline-formula><mml:math id="M19"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">ML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> from complete data is given by <inline-formula><mml:math id="M20"><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>. The increase in the sampling variability of the <italic>j</italic>th parameter estimate (i.e., the <italic>j</italic>th element of <inline-formula><mml:math id="M21"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula>) due to missing data is therefore <inline-formula><mml:math id="M22"><mml:msub><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, where the subscript <italic>jj</italic> indicates <italic>j</italic>th diagonal element of a matrix. The FMI for the <italic>j</italic>th parameter estimate <inline-formula><mml:math id="M23"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, can then be defined as the proportion of its sampling variability that is due to missing data (Savalei and Rhemtulla, <xref ref-type="bibr" rid="B20">2012</xref>):</p>
<disp-formula id="E70"><label>(5)</label><mml:math id="M200"><mml:mrow><mml:msub><mml:mi>&#x003B4;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mi>Y</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mi>X</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mi>Y</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mi>X</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mi>Y</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:mi>S</mml:mi><mml:msubsup><mml:mi>&#x1D53C;</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mrow><mml:mi>S</mml:mi><mml:msubsup><mml:mi>&#x1D53C;</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>O</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where <italic>SE</italic><sub><italic>j, O</italic></sub> is the standard error of <inline-formula><mml:math id="M25"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <italic>SE</italic><sub><italic>j, C</italic></sub> is what the standard error would have been under complete data (i.e., the standard error of the hypothetical estimate <inline-formula><mml:math id="M26"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">ML</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> that would have been obtained from complete data).</p>
<p>Savalei and Rhemtulla (<xref ref-type="bibr" rid="B20">2012</xref>) gave an applied interpretation of the FMI in terms of its relationship to the <italic>width inflation factor</italic> (WIF). The WIF is the increase in the standard error and the width of the confidence interval due to missing information. Define the WIF for the <italic>j</italic>th parameter as <inline-formula><mml:math id="M27"><mml:msub><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">WIF</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>O</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:mfrac></mml:math></inline-formula>. Then, from Equation (5), we have the relationship <inline-formula><mml:math id="M28"><mml:msub><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">WIF</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:math></inline-formula>. This equation shows that there is a one-to-one correspondence between the FMI and the impact missing data have on the property of the parameter estimate. For example, when &#x003B4;<sub><italic>j</italic></sub> &#x0003D;.75, WIF<sub><italic>j</italic></sub> &#x0003D; 2, which means the standard error of the <italic>j</italic>th parameter is twice as large as it would have been under complete data. Consequently, the confidence interval of that parameter estimate also becomes twice as wide. When &#x003B4;<sub><italic>j</italic></sub> &#x0003D;.5, the same computation gives us WIF<sub><italic>j</italic></sub> &#x0003D; 1.5, which shows the standard error and the CI are 50% larger due to missing data.</p></sec>
<sec>
<title>1.2. Sample Estimates of FMI</title>
<p>To obtain sample estimates of FMIs following FIML estimation from Equation (5), which provides the population definition, we need standard error estimates corresponding to complete and incomplete data. However, as we only have incomplete data, it may appear that we only have standard error estimates based on incomplete data. To bypass this problem, Savalei and Rhemtulla (<xref ref-type="bibr" rid="B20">2012</xref>) proposed to estimate what the complete data standard errors for each parameter would have been by evaluating the theoretical formulae for ML standard errors under complete data at the FIML parameter estimates. To do so, the model-implied means and covariances <inline-formula><mml:math id="M29"><mml:mi>&#x003BC;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M30"><mml:mo>&#x003A3;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">FIML</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, are first obtained by analyzing the data via FIML. SEM softwares are then able to use these means and covariances to perform a complete data estimation routine, which produces standard errors to be used as complete data standard error estimates in the computation of FMI. This computation appeared to work well, and has since been automated in <italic>lavaan</italic>. However, the initial implementation had problems, in that negative estimates of FMIs were often produced, particularly in small samples. The reason for this poor performance was that by default, SEM software packages compute so-called &#x0201C;expected&#x0201D; standard errors for complete data and &#x0201C;observed&#x0201D; standard errors for incomplete data (Savalei, <xref ref-type="bibr" rid="B17">2010</xref>), and the estimates proposed and studied in Savalei and Rhemtulla (<xref ref-type="bibr" rid="B20">2012</xref>) relied on these defaults. However, in small samples, the differences between these types of standard errors can be large enough to create problems, resulting in negative FMI estimates. Starting with <italic>lavaan</italic> version 0.6-7, the package now uses the same type of information matrix, leading to the same type of standard error estimates, when estimating FMIs using Equation (5)<xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>. Because standard errors for incomplete MAR data are not consistent unless they are based on observed information (Little and Rubin, <xref ref-type="bibr" rid="B9">2002</xref>), we recommend always using observed information to compute FMIs; this is in fact the default in <italic>lavaan</italic>.</p>
<p>There are three computational variations that fall under the broad label of &#x0201C;observed&#x0201D; standard errors, all of which remain consistent when the missing mechanism is MAR and the model is true. These three computational variations will be identical for saturated models (i.e., models that have zero degrees of freedom, such as regression models), but they will generally differ for structured models, i.e., models that have positive degrees of freedom and that impose some constraints on the means and/or the covariance matrix of the data. These variations are described in <xref ref-type="supplementary-material" rid="SM2">Appendix B</xref>, and they are implemented in <italic>lavaan</italic> version 0.6-7 (for further details, see Savalei and Rosseel, ress). These variations lead to three different computational estimates of FMI, which we will call <inline-formula><mml:math id="M31"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M32"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="M33"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Briefly, the estimate <inline-formula><mml:math id="M34"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> uses the numeric Hessian (the second derivative of the log-likelihood) in both complete and incomplete standard error estimates. Standard errors based on the Hessian are the default in <italic>lavaan</italic> when the FIML estimator is requested, and these will be the default FMI estimates. The second estimate, <inline-formula><mml:math id="M35"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, is obtained analytically by approximating the second derivative of the log-likelihood with the first (dominant) term and dropping the second term that depends on the second derivatives of the model (this second smaller term is cumbersome to obtain analytically). The third estimate, <inline-formula><mml:math id="M36"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, has the same analytic form as <inline-formula><mml:math id="M37"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, but uses unstructured estimates of means and covariance matrix in all parts of the computation that rely on these estimates.</p>
<p>When the proposed structured model is true, the three estimates are consistent estimates of the same quantity and therefore asymptotically equivalent (see <xref ref-type="supplementary-material" rid="SM2">Appendix B</xref>; Savalei and Rosseel, ress). However, they will diverge when the model is false, and only the estimate <inline-formula><mml:math id="M38"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, based on the Hessian, will remain consistent. As this is the first study of these estimates, we only evaluate the scenarios where the proposed model is either saturated or true. In such situations, the estimate <inline-formula><mml:math id="M39"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> may have the disadvantage that it is more complicated to compute and may thus lead to less accurate FMI estimates in small samples, but this is a prediction that needs to be empirically evaluated.</p></sec></sec>
<sec id="s2">
<title>2. Simulation Study</title>
<p>While the idea for how to obtain sample FMIs following FIML estimation was published a number of years ago (Savalei and Rhemtulla, <xref ref-type="bibr" rid="B20">2012</xref>), we are not aware of any simulation studies that have been conducted to evaluate the performance of sample FMIs. Furthermore, it was not until recently that FMI estimates with better properties (using the same type of standard errors in both numerator and denominator), have become automated in <italic>lavaan</italic>. In order to evaluate the performance of these sample FMI estimates, we conducted a simulation study with two commonly used models in psychology, a simple regression model and a two-factor model. The main objective was to determine whether the population FMI values can be estimated with reasonable accuracy under a realistic sample size. The three estimates, <inline-formula><mml:math id="M40"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M41"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="M42"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, were expected to be practically equal in the saturated regression model. It was unclear, however, how they would differ in the two-factor model. Therefore, another objective for the study was to compare their performance.</p>
<sec>
<title>2.1. Method</title>
<p>The study was conducted in <italic>R</italic> v3.6.1 (R Core Team, <xref ref-type="bibr" rid="B12">2019</xref>). All FMIs were computed from <italic>lavaan</italic> 0.6-7 (Rosseel, <xref ref-type="bibr" rid="B14">2012</xref>), using options specified in <xref ref-type="supplementary-material" rid="SM1">Supplementary Table 1</xref>. Example code of how to use <italic>lavaan</italic> to compute the three FMI estimates is given in <xref ref-type="supplementary-material" rid="SM2">Appendix A</xref>. Simulations were conducted with 1,000 replications in each cell. An example of the simulation code is available on OSF at <ext-link ext-link-type="uri" xlink:href="https://osf.io/xyzt8/">https://osf.io/xyzt8/</ext-link>.</p>
<sec>
<title>2.1.1. Model 1: Simple Regression</title>
<p>The simple regression model simulations contained 4 (sample size) &#x000D7; 3 (missing mechanisms) &#x000D7; 3 (missing rates) = 12 conditions. For each condition, we examined the three sample FMIs (<inline-formula><mml:math id="M44"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M45"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M46"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>) of the regression coefficient (&#x003B2;). Data were simulated from the path model <italic>Y</italic> &#x0003D; &#x003B2;<italic>X</italic>&#x0002B;<italic>E</italic>, where <italic>X</italic> and <italic>Y</italic> followed standard normal distributions (i.e., means of 0 and standard deviations of 1), &#x003B2; was always 0.4. <italic>E</italic> was therefore a normal distribution with a mean of 0 and standard deviation of <inline-formula><mml:math id="M47"><mml:msqrt><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msqrt><mml:mo>=</mml:mo><mml:mo>.</mml:mo><mml:mn>92</mml:mn></mml:math></inline-formula>. The sample sizes were <italic>N</italic>=50, 100, 200, and 500. In order to compute the bias of the sample FMI estimates, pseudo-population values of the FMIs were computed from an additional single replication simulation with <italic>N</italic> &#x0003D; 1, 000, 000.</p>
<p>The missing data mechanism conditions were MCAR, linear MAR (MAR-L), and nonlinear MAR (MAR-NL), with values missing on <italic>X</italic> at population rates of 0.2, 0.4, and 0.6. To avoid the NMAR mechanism, missingness was only assigned to <italic>X</italic>, with missing rates &#x003C0;<sub><italic>mis</italic></sub> = 20, 40, and 60%. Since missingness was only assigned to one of the two variables, the overall missing rates were therefore 10, 20, and 30%, respectively. In the special case of MCAR, each value of <italic>X</italic> was assigned an independent &#x003C0;<sub><italic>mis</italic></sub> chance of missing. For MAR conditions, we specifically chose very strong selection mechanisms in the MAR conditions, as weak MAR mechanisms tend to behave similarly to each other and to MCAR. The choice of strong selection mechanisms allows us to contrast the differences between the different MAR conditions. For these MAR conditions, the missingness in <italic>X</italic> was conditioned on <italic>Y</italic>.</p>
<p>Under MAR-L, values of <italic>X</italic> were set to be missing with a high probability when the corresponding values of <italic>Y</italic> were above a certain cutoff. A single cutoff rule was used to implement this mechanism. When <italic>Y</italic> was above the percentile cutoff value <italic>y</italic><sub><italic>cut</italic></sub>, the missing rate on <italic>X</italic> was &#x003C0;<sub>1</sub> = 90%; below that cutoff, the missing rate on <italic>X</italic> was &#x003C0;<sub>2</sub> = 10%. In each condition, the value <italic>y</italic><sub><italic>cut</italic></sub> was computed from the theoretical &#x003C0;<sub><italic>c</italic></sub> percentile of the population distribution of <italic>Y</italic>, where &#x003C0;<sub><italic>c</italic></sub> &#x0003D; (&#x003C0;<sub><italic>mis</italic></sub>&#x02212;&#x003C0;<sub>1</sub>)/(&#x003C0;<sub>2</sub>&#x02212;&#x003C0;<sub>1</sub>)= 0.875, 0.625, and 0.375. This produces a linear MAR mechanism with per variable missing rates of &#x003C0;<sub><italic>mis</italic></sub> = 20, 40, and 60%, respectively. Since <italic>Y</italic> followed a standard normal distribution, the cutoffs were, for each missing rate condition respectively, <italic>y</italic><sub><italic>cut</italic></sub>= 1.15, 0.32, and &#x02212;0.32.</p>
<p>MAR-NL was simulated using two cutoffs, <italic>y</italic><sub><italic>cut</italic>1</sub> and <italic>y</italic><sub><italic>cut</italic>2</sub>. When <italic>Y</italic> was either above <italic>y</italic><sub><italic>cut</italic>1</sub> or below <italic>y</italic><sub><italic>cut</italic>2</sub>, the missing rate was &#x003C0;<sub>1</sub> = 90%; between those two cutoffs, the missing rate was &#x003C0;<sub>2</sub> = 10%. The cutoffs were the theoretical 50%&#x000B1;0.5&#x003C0;<sub><italic>c</italic></sub> percentiles of the population distribution of <italic>Y</italic>, where 0.5&#x003C0;<sub><italic>c</italic></sub>= 0.4375, 0.3125, and 0.1875. Since <italic>Y</italic> followed a standard normal distribution, the cutoffs were, for each condition respectively, <italic>y</italic><sub><italic>cut</italic>1</sub>= 1.53, 0.89, and 0.49 and <italic>y</italic><sub><italic>cut</italic>2</sub>= &#x02212;1.53, &#x02212;0.89, and &#x02212;0.49. This type of nonlinear MAR resulted in missing values on both ends of the distribution of <italic>X</italic>, which would theoretically lead to more information loss.</p></sec>
<sec>
<title>2.1.2. Model 2</title>
<p>The two-factor model simulations contained 4 (sample size) &#x000D7; 3 (missing mechanisms) &#x000D7; 3 (missing rates) = 12 conditions. For each condition, we investigated the properties of the three sample FMIs (<inline-formula><mml:math id="M75"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M76"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M77"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>) for the factor correlation (&#x003D5;) and factor loadings (&#x003BB;s). The two latent factors were <italic>F</italic><sub>1</sub> and <italic>F</italic><sub>2</sub>, and each factor included with four indicators: <italic>X</italic><sub>1</sub>, <italic>X</italic><sub>2</sub>, <italic>X</italic><sub>3</sub>, <italic>X</italic><sub>4</sub> measured <italic>F</italic><sub>1</sub>, and <italic>Y</italic><sub>1</sub>, <italic>Y</italic><sub>2</sub>, <italic>Y</italic><sub>3</sub>, <italic>Y</italic><sub>4</sub> measured <italic>F</italic><sub>2</sub>. All indicators, factors, and measurement errors followed standard normal distributions. The two factors were correlated at &#x003D5; &#x0003D; 0.4. The measurement model was set to have equal loadings, with all loading values set to &#x003BB; &#x0003D; 0.49 (see <xref ref-type="supplementary-material" rid="SM1">Supplementary Table 2</xref> for a summary of the study design). As factor analysis is typically conducted with a larger sample size than regression, the sample sizes were set to <italic>N</italic>=100, 200, 500, and 1,000, excluding the <italic>N</italic> &#x0003D; 50 condition and adding the <italic>N</italic>=1,000 condition. The &#x0201C;population&#x0201D;-level FMIs were once again computed from a single sample with <italic>N</italic> &#x0003D; 1, 000, 000.</p>
<p>Three missing data mechanism conditions, MCAR, MAR-L, and MAR-NL, were included in the two-factor model simulations. Half the variables were assigned to contain missing values: <italic>X</italic><sub>2</sub>, <italic>X</italic><sub>4</sub>, <italic>Y</italic><sub>2</sub>, and <italic>Y</italic><sub>4</sub>. The variables <italic>X</italic><sub>1</sub>, <italic>X</italic><sub>3</sub>, <italic>Y</italic><sub>1</sub>, and <italic>Y</italic><sub>3</sub> were fully observed. Where the missing data mechanism was not MCAR, missingness on <italic>X</italic><sub>2</sub> and <italic>X</italic><sub>4</sub> was conditioned on the value of <italic>C</italic><sub><italic>X</italic></sub> &#x0003D; <italic>X</italic><sub>1</sub>&#x0002B;<italic>X</italic><sub>3</sub>. Similarly, missingness on <italic>Y</italic><sub>2</sub> and <italic>Y</italic><sub>4</sub> was conditioned on the value of <italic>C</italic><sub><italic>Y</italic></sub> &#x0003D; <italic>Y</italic><sub>1</sub>&#x0002B;<italic>Y</italic><sub>3</sub>. The cutoffs for MAR-L and MAR-NL were based on the standardized values of <italic>C</italic><sub><italic>X</italic></sub> and <italic>C</italic><sub><italic>Y</italic></sub>, and were the same as those in the regression model. For each variable containing missing values, the missing rate on that variable was &#x003C0;<sub><italic>mis</italic></sub> = 20, 40, and 60%. Since only half the variables contained missingness, these per variable missing rates corresponded to overall missing rates of 10, 20, and 30%.</p></sec>
<sec>
<title>2.1.3. Evaluation Criteria</title>
<p>In order to assess the properties of the sample estimates of FMIs, we computed the raw bias, root mean squared error (RMSE), and 95% equal-tailed interval (ETI) width based on the 1,000 replications. The raw bias was computed as <inline-formula><mml:math id="M78"><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, where <inline-formula><mml:math id="M79"><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:math></inline-formula> was the mean of the sample FMI estimates produced in all replications, and <italic>e</italic>= 1, 2, or 3, as the evaluation criteria were computed in the same when for all three FMI estimates. <inline-formula><mml:math id="M80"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was the pseudo-population FMI value generated from the additional single replication with <italic>N</italic>= 1,000,000. The raw bias indicated how much the sample FMI deviated from the population on average, in the long run. We expected these estimates to be largely unbiased, and therefore we considered a bias of more than 0.05 (&#x003B4; ranged from 0 to 1) to be notable.</p>
<p>The RMSE was computed as <inline-formula><mml:math id="M81"><mml:msqrt><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:msubsup><mml:mrow><mml:mo>&#x003A3;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msqrt></mml:math></inline-formula>, where <italic>N</italic><sub><italic>rep</italic></sub> was the number of replications and <inline-formula><mml:math id="M82"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was the sample FMI estimate on the <italic>i</italic>th replication. The RMSE captured the bias and variability of the estimate from run to run, and could be roughly interpreted as the average difference from the true population value. Given that we expected the estimates to be largely unbiased, the RMSE was mainly used as a measure of efficiency.</p>
<p>As the RMSE is not very easily interpretable, we included an additional efficiency measure, the 95% ETI width, defined as the difference between the 2.5th quantile and the 97.5th quantile of the observed <inline-formula><mml:math id="M83"><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> values across all replications. The ETI width therefore indicated where most of the sample estimates lied across all replications, and a low ETI width suggests the parameter estimate is more precise. In the case of a symmetrical sampling distribution, 95% of the estimates in all replications fall within an interval of <inline-formula><mml:math id="M84"><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:math></inline-formula> plus and minute half the ETI width. For example, if the mean of a symmetrical FMI sampling distribution is 0.5, an ETI width of 0.2 shows that 95% of the sample FMIs fall within, 0.5&#x000B1;0.2/2, or [0.4, 0.6].</p></sec></sec>
<sec>
<title>2.2. Results</title>
<p>In this study, we examine the FMIs of three parameters, namely, the regression coefficient (&#x003B2;), the factor loading (&#x003BB;), and the factor correlation (&#x003D5;). For simplicity of notation, the numeric subscript <italic>j</italic> is substituted by the name of the parameter when a specific FMI is referenced. For example, the numeric Hessian FMI estimate <inline-formula><mml:math id="M85"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> for the regression coefficient &#x003B2; will be denoted as <inline-formula><mml:math id="M86"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. The result summary of each simulation condition is available on OSF at <ext-link ext-link-type="uri" xlink:href="https://osf.io/xyzt8/">https://osf.io/xyzt8/</ext-link>.</p>
<sec>
<title>2.2.1. Population Values</title>
<p>As was theoretically expected, the three FMI computations were identical at <italic>N</italic>=1,000,000. Therefore, for the population values, we do not distinguish between &#x003B4;<sub>1, <italic>j</italic></sub>, &#x003B4;<sub>2, <italic>j</italic></sub>, and &#x003B4;<sub>3, <italic>j</italic></sub>, and simply refer to the value as &#x003B4;<sub><italic>pop, j</italic></sub>.</p>
<p>In Model 1 (regression), the FMIs of the regression coefficient (&#x003B2;) under per variable missing rates of &#x003C0;<sub><italic>mis</italic></sub>=0.2, 0.4, 0.6, respectively, were &#x003B4;<sub><italic>pop, j</italic></sub>=0.15, 0.33, and 0.52 under MCAR. Under MAR-L, they were &#x003B4;<sub><italic>pop</italic>, &#x003B2;</sub>=0.29, 0.51, 0.63, and MAR-NL, they were &#x003B4;<sub><italic>pop</italic>, &#x003B2;</sub>=0.35, 0.65, 0.79 (as seen in <xref ref-type="table" rid="T1">Table 1</xref>). As expected, the population FMIs increased as the missing rate increased in all three missing data mechanism conditions. However, the missing rates did not capture all the information loss. When the missing rates were equal, the FMIs could vary greatly across the three missing data mechanism conditions. Most notably, due the loss of data at the tail ends of the distribution in the nonlinear MAR condition, this missing data mechanism was associated with the highest FMIs.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>The population FMIs.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Parameter</bold></th>
<th valign="top" align="center"><bold>&#x003C0;<sub><italic>mis</italic></sub></bold></th>
<th valign="top" align="center" colspan="3" style="border-bottom: thin solid #000000;"><bold>&#x003B4;<sub><italic>pop, j</italic></sub></bold></th>
</tr>
<tr>
<th/>
<th/>
<th/>
<th valign="top" align="center"><bold>MCAR</bold></th>
<th valign="top" align="center"><bold>MAR-L</bold></th>
<th valign="top" align="center"><bold>MAR-NL</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Regression</td>
<td valign="top" align="center">&#x003B2;</td>
<td valign="top" align="center">0.2</td>
<td valign="top" align="center">0.15</td>
<td valign="top" align="center">0.29</td>
<td valign="top" align="center">0.35</td>
</tr>
<tr>
<td valign="top" align="left">Regression</td>
<td valign="top" align="center">&#x003B2;</td>
<td valign="top" align="center">0.4</td>
<td valign="top" align="center">0.33</td>
<td valign="top" align="center">0.51</td>
<td valign="top" align="center">0.65</td>
</tr>
<tr>
<td valign="top" align="left">Regression</td>
<td valign="top" align="center">&#x003B2;</td>
<td valign="top" align="center">0.6</td>
<td valign="top" align="center">0.52</td>
<td valign="top" align="center">0.63</td>
<td valign="top" align="center">0.79</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003D5;</td>
<td valign="top" align="center">0.2</td>
<td valign="top" align="center">0.15</td>
<td valign="top" align="center">0.29</td>
<td valign="top" align="center">0.35</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003D5;</td>
<td valign="top" align="center">0.4</td>
<td valign="top" align="center">0.33</td>
<td valign="top" align="center">0.51</td>
<td valign="top" align="center">0.65</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003D5;</td>
<td valign="top" align="center">0.6</td>
<td valign="top" align="center">0.52</td>
<td valign="top" align="center">0.63</td>
<td valign="top" align="center">0.79</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003BB;</td>
<td valign="top" align="center">0.2</td>
<td valign="top" align="center">0.15</td>
<td valign="top" align="center">0.29</td>
<td valign="top" align="center">0.35</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003BB;</td>
<td valign="top" align="center">0.4</td>
<td valign="top" align="center">0.33</td>
<td valign="top" align="center">0.51</td>
<td valign="top" align="center">0.65</td>
</tr>
<tr>
<td valign="top" align="left">Two-factor</td>
<td valign="top" align="center">&#x003BB;</td>
<td valign="top" align="center">0.6</td>
<td valign="top" align="center">0.52</td>
<td valign="top" align="center">0.63</td>
<td valign="top" align="center">0.79</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Numbers computed from N=1,000,000. &#x003C0;<sub>mis</sub> is the probability of missing value in each variable with missingness. The overall missing rate is &#x003C0;<sub>mis</sub>/2. &#x003B2;, the regression coefficient. &#x003D5;, the factor correlation. &#x003BB;, the factor loading. The population FMIs of the four factor loadings corresponding to variables containing missingness are all within 0.01 of each other, and therefore only the FMIs of the &#x003BB; for X<sub>2</sub> is reported</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>In Model 2 (two-factor model), there were 4 factor loadings associated with variables that contained missing data, namely the factor loadings of <italic>X</italic><sub>2</sub>, <italic>X</italic><sub>4</sub> on <italic>F</italic><sub>1</sub>, and the factor loadings of <italic>Y</italic><sub>2</sub>, <italic>Y</italic><sub>4</sub> on <italic>F</italic><sub>2</sub>. As these loadings were of the same size, and the missing rate was the same for all four variables, only the FMI on the loading of <italic>X</italic><sub>2</sub> is reported here. Under MCAR, the FMIs of the factor loading (&#x003BB;) were &#x003B4;<sub><italic>pop</italic>, &#x003BB;</sub>=0.20, 0.40, and 0.60 for per-variable missing rates of &#x003C0;<sub><italic>mis</italic></sub>=0.2, 0.4, and 0.6, respectively. Under MAR-L, the FMIs were &#x003B4;<sub><italic>pop</italic>, &#x003BB;</sub>=0.37, 0.60, 0.71. Under MAR-NL, they were &#x003B4;<sub><italic>pop</italic>, &#x003BB;</sub>=0.44, 0.71, 0.83. Once again, we observed that in each condition, the FMI increased as the missing rates increased, and the MAR-NL condition lead to the highest FMIs, followed by MAR-L, and then MCAR.</p>
<p>The pattern of differences between the missing data mechanisms, however, did not hold for the FMIs of the factor correlation (&#x003D5;) in Model 2. The FMIs under per variable missing rates of &#x003C0;<sub><italic>mis</italic></sub>=0.2, 0.4, and 0.6, respectively, were &#x003B4;<sub><italic>pop</italic>, &#x003D5;</sub>=0.13, 0.26, 0.38 in MCAR, &#x003B4;<sub><italic>pop</italic>, &#x003D5;</sub>=0.14, 0.27, 0.38 in MAR-L, and &#x003B4;<sub><italic>pop</italic>, &#x003D5;</sub>=0.14, 0.28, and 0.39 in MAR-NL. The FMIs of the factor correlation, unlike the FMIs of other parameters, did not differ notably across the missing mechanisms. Even within the same model, parameters affected by the same missing rates and the same missing data mechanism could see drastically different amounts of information loss. For instance, under &#x003C0;<sub><italic>mis</italic></sub> &#x0003D; 0.2 and MAR-NL, the FMI on &#x003D5; was &#x003B4;<sub><italic>pop</italic>, &#x003D5;</sub> &#x0003D; 0.14, but the FMI on &#x003BB; in the same model was &#x003B4;<sub><italic>pop</italic>, &#x003D5;</sub> &#x0003D; 0.44. These results emphasize the point that information loss is not predictable from the missing rates and whether the missing data mechanism is MCAR or MAR. Information loss can only be quantified by computing the FMI for a particular mechanism and parameter of interest.</p></sec>
<sec>
<title>2.2.2. Model 1</title>
<p>For the regression coefficient &#x003B2;, the three sample FMI estimates, <inline-formula><mml:math id="M87"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="M88"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="M89"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were identical, due to the regression model being saturated. We refer to the resulting estimate simply as <inline-formula><mml:math id="M90"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in <xref ref-type="table" rid="T2">Table 2</xref>. As we can see in the tables, the sample FMI estimates were largely unbiased, with the long run means of <inline-formula><mml:math id="M91"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> over the 1,000 replications falling within 0.05 of the population value in practically all conditions. Notable bias (i.e., &#x0003E;0.05) only arose when the missing data mechanism was NL-MAR, the sample size was small, and the missing rate was high, where the sample FMI could underestimate the population value by as much as 0.07 (near the top of <xref ref-type="table" rid="T2">Table 2</xref> under the MAR-NL columns).</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>The bias, RMSE, and the 95% equal-tailed interval width of the regression coefficient FMI estimate (<inline-formula><mml:math id="M43"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>).</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>&#x003C0;<sub><italic>mis</italic></sub></bold></th>
<th valign="top" align="left"><bold>N</bold></th>
<th valign="top" align="left" colspan="3"><bold>MCAR</bold></th>
<th valign="top" align="left" colspan="3"><bold>MAR-L</bold></th>
<th valign="top" align="left" colspan="3"><bold>MAR-NL</bold></th>
</tr>
<tr>
<th/>
<th/>
<th valign="top" align="left"><bold>Bias</bold></th>
<th valign="top" align="left"><bold>RMSE</bold></th>
<th valign="top" align="left"><bold>ETI</bold></th>
<th valign="top" align="left"><bold>Bias</bold></th>
<th valign="top" align="left"><bold>RMSE</bold></th>
<th valign="top" align="left"><bold>ETI</bold></th>
<th valign="top" align="left"><bold>Bias</bold></th>
<th valign="top" align="left"><bold>RMSE</bold></th>
<th valign="top" align="left"><bold>ETI</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.45</bold></td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left"><bold>0.14</bold></td>
<td valign="top" align="left"><bold>0.52</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left"><bold>0.39</bold></td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left"><bold>0.15</bold></td>
<td valign="top" align="left"><bold>0.54</bold></td>
<td valign="top" align="left"><bold>&#x02212;0.05</bold></td>
<td valign="top" align="left"><bold>0.19</bold></td>
<td valign="top" align="left"><bold>0.64</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.43</bold></td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left"><bold>0.14</bold></td>
<td valign="top" align="left"><bold>0.52</bold></td>
<td valign="top" align="left"><bold>&#x02212;0.07</bold></td>
<td valign="top" align="left"><bold>0.19</bold></td>
<td valign="top" align="left"><bold>0.57</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.09</bold></td>
<td valign="top" align="left"><bold>0.34</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left"><bold>0.39</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.43</bold></td>
<td valign="top" align="left">&#x02212;0.03</td>
<td valign="top" align="left"><bold>0.15</bold></td>
<td valign="top" align="left"><bold>0.56</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.42</bold></td>
<td valign="top" align="left"><bold>&#x02212;0.06</bold></td>
<td valign="top" align="left"><bold>0.16</bold></td>
<td valign="top" align="left"><bold>0.53</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.13</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.25</bold></td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.09</bold></td>
<td valign="top" align="left"><bold>0.35</bold></td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.45</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.22</bold></td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.31</bold></td>
<td valign="top" align="left">&#x02212;0.03</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.44</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.09</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.16</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.19</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.12</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.22</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.29</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.14</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.21</bold></td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>&#x003C0;<sub>mis</sub> is the probability of missing value in each variable with missingness. The overall missing rate is &#x003C0;<sub>mis</sub>/2. ETI denotes the 95% ETI width, which is the distance between the 2.5 and 97.5 percentile of the sampling distributions. Bias values with magnitudes &#x0003E;0.05, RMSEs &#x0003E;0.05, and ETI widths &#x0003E;0.20 are indicated with bold fonts. The conditional formatting was applied prior to rounding</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>The sample FMI estimates were not particularly efficient at the smaller sample sizes typical of regression analyses, and would produce highly variable results from run to run. At <italic>N</italic> &#x0003D; 50 and 100, the RMSEs were around 0.10 or higher across nearly all missing rates and missing data mechanisms. At <italic>N</italic> = 100, the estimate was the most precise estimate under MCAR with &#x003C0;<sub><italic>mis</italic></sub> &#x0003D; 0.2, with RMSE of 0.05 and 95% ETI width of 0.19. For larger sample sizes <italic>N</italic> &#x0003D; 200 and 500, the performance of <inline-formula><mml:math id="M92"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was overall acceptable under MCAR, but not for MAR-L and MAR-NL. Under both MAR conditions, good performance was only achieved when <italic>N</italic> &#x0003D; 500 and &#x003C0;<sub><italic>mis</italic></sub> &#x0003D; 0.2.</p>
<p>For a more intuitive illustration of bias and variability, the smoothed densities of the 1,000 replications in each condition are shown in <xref ref-type="fig" rid="F1">Figures 1</xref>&#x02013;<xref ref-type="fig" rid="F3">3</xref>. We can see that almost all sampling distributions centered around the population value, with the exception of the top right panels of <xref ref-type="fig" rid="F3">Figure 3</xref>, which correspond to the small sample size, high missing rate, nonlinear MAR conditions. These distributions, however, do not pack very tightly around the population value, except when the sample size is 500, and only when either the missing rate was low, or the missing mechanism was MCAR.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The sampling distribution of FMI for regression coefficients in MCAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line. The sampling distributions of the three estimates are virtually identical in simple regression.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0001.tif"/>
</fig>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>The sampling distribution of FMI for regression coefficients in linear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line. The sampling distributions of the three estimates are virtually identical in simple regression.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0002.tif"/>
</fig>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>The sampling distribution of FMI for regression coefficients in nonlinear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line. The sampling distributions of the three estimates are virtually identical in simple regression.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0003.tif"/>
</fig></sec>
<sec>
<title>2.2.3. Model 2</title>
<p>In the two-factor model, not all replications were usable. In some cases, the model would fail to converge. In other cases, when FMIs were requested, <italic>lavaan</italic> would sometimes produce an error, leading to NAs or negative values for the FMI estimates. These issues were exacerbated by higher missing rates, and were particularly pronounced in <inline-formula><mml:math id="M93"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> when the sample size was small. Under <italic>N</italic> &#x0003D; 100 and &#x003C0;<sub><italic>mis</italic></sub>=0.6 (overall missing rate 0.3), <inline-formula><mml:math id="M94"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> would encounter more issues in more than 30% of the runs. In this regard, the best performing estimate was <inline-formula><mml:math id="M95"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, which almost never produced negative values, and had the lowest rate of NA occurrences. However, even for <inline-formula><mml:math id="M96"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the rate of failed or improper estimates was often close to or above 10% at <italic>N</italic> &#x0003D; 100. At <italic>N</italic> &#x0003D; 200, <inline-formula><mml:math id="M97"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> would encounter this issue at most 3% of the time, even when the per variable missing rate was 0.6. See <xref ref-type="supplementary-material" rid="SM1">Supplementary Tables 3</xref>, <xref ref-type="supplementary-material" rid="SM1">4</xref> for the rate of failed or improper estimates for the factor correlations and factor loadings, respectively. The following results were obtained by excluding all occurrences of NAs and negative values when aggregating the sample FMI estimates.</p>
<p>The sample FMIs of factor correlation (&#x003D5;) were largely unbiased in MCAR, or when the mechanism was MAR but the sample size was 200 or above (see <xref ref-type="table" rid="T3">Tables 3</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref>). Bias was also generally small at <italic>N</italic> &#x0003D; 100 when the missing data mechanism was MCAR, but notable bias could occur in MAR-L and MAR-NL. Among the three estimates, <inline-formula><mml:math id="M98"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> showed the largest amount of bias overall, severely overestimating the FMI (0.17 above the true value on average) when the missing data mechanism was nonlinear MAR and the per variable missing rate was &#x003C0;<sub><italic>mis</italic></sub>=0.6. <inline-formula><mml:math id="M99"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> also showed some notable bias under <italic>N</italic> &#x0003D; 100, but only up to 0.11 in the worst case. <inline-formula><mml:math id="M100"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> showed the least amount of bias, producing raw bias values very close to 0.05 or below across all the conditions.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>The bias, RMSE, and the 95% equal-tailed interval width of factor correlation FMIs under MCAR.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>&#x003C0;<sub><italic>mis</italic></sub></bold></th>
<th valign="top" align="left"><bold>N</bold></th>
<th valign="top" align="left" colspan="3"><bold>Bias</bold></th>
<th valign="top" align="left" colspan="3"><bold>RMSE</bold></th>
<th valign="top" align="left" colspan="3"><bold>95% ETI width</bold></th>
</tr>
<tr>
<th/>
<th/>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M48"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M49"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M50"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M51"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M52"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M53"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M54"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M55"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M56"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.31</bold></td>
<td valign="top" align="left"><bold>0.23</bold></td>
<td valign="top" align="left"><bold>0.24</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left"><bold>0.16</bold></td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.67</bold></td>
<td valign="top" align="left"><bold>0.44</bold></td>
<td valign="top" align="left"><bold>0.39</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left"><bold>0.25</bold></td>
<td valign="top" align="left"><bold>0.17</bold></td>
<td valign="top" align="left"><bold>0.16</bold></td>
<td valign="top" align="left"><bold>0.93</bold></td>
<td valign="top" align="left"><bold>0.66</bold></td>
<td valign="top" align="left"><bold>0.58</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.17</td>
<td valign="top" align="left">0.16</td>
<td valign="top" align="left">0.15</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.09</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.34</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
<td valign="top" align="left"><bold>0.29</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">&#x02212;0.02</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left"><bold>0.15</bold></td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.56</bold></td>
<td valign="top" align="left"><bold>0.46</bold></td>
<td valign="top" align="left"><bold>0.43</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.10</td>
<td valign="top" align="left">0.10</td>
<td valign="top" align="left">0.10</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.20</td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.18</td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.07</td>
<td valign="top" align="left">0.07</td>
<td valign="top" align="left">0.07</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.13</td>
<td valign="top" align="left">0.13</td>
<td valign="top" align="left">0.13</td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.20</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>&#x003C0;<sub>mis</sub> is the probability of missing value in each variable with missingness. The overall missing rate is &#x003C0;<sub>mis</sub>/2. Distribution width is the distance between the 2.5 and 97.5 percentile in the sampling distributions. Bias values with magnitudes &#x0003E;0.05, RMSEs &#x0003E;0.05, and ETI widths &#x0003E;0.20 are indicated with bold fonts. The conditional formatting was applied prior to rounding</italic>.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>The bias, RMSE, and the 95% equal-tailed interval width of factor correlation FMIs under linear MAR.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>&#x003C0;<sub><italic>mis</italic></sub></bold></th>
<th valign="top" align="left"><bold>N</bold></th>
<th valign="top" align="left" colspan="3"><bold>Bias</bold></th>
<th valign="top" align="left" colspan="3"><bold>RMSE</bold></th>
<th valign="top" align="left" colspan="3"><bold>95% ETI width</bold></th>
</tr>
<tr>
<th/>
<th/>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M57"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M58"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M59"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M60"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M61"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M62"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M63"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M64"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M65"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.38</bold></td>
<td valign="top" align="left"><bold>0.26</bold></td>
<td valign="top" align="left"><bold>0.25</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left"><bold>0.20</bold></td>
<td valign="top" align="left"><bold>0.13</bold></td>
<td valign="top" align="left"><bold>0.13</bold></td>
<td valign="top" align="left"><bold>0.87</bold></td>
<td valign="top" align="left"><bold>0.49</bold></td>
<td valign="top" align="left"><bold>0.45</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left"><bold>0.19</bold></td>
<td valign="top" align="left"><bold>0.17</bold></td>
<td valign="top" align="left"><bold>0.91</bold></td>
<td valign="top" align="left"><bold>0.73</bold></td>
<td valign="top" align="left"><bold>0.61</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left"><bold>0.20</bold></td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.18</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left"><bold>0.09</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.38</bold></td>
<td valign="top" align="left"><bold>0.33</bold></td>
<td valign="top" align="left"><bold>0.31</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">&#x02212;0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left"><bold>0.16</bold></td>
<td valign="top" align="left"><bold>0.13</bold></td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.61</bold></td>
<td valign="top" align="left"><bold>0.52</bold></td>
<td valign="top" align="left"><bold>0.46</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.12</td>
<td valign="top" align="left">0.11</td>
<td valign="top" align="left">0.11</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.21</bold></td>
<td valign="top" align="left"><bold>0.21</bold></td>
<td valign="top" align="left">0.20</td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.32</bold></td>
<td valign="top" align="left"><bold>0.29</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.08</td>
<td valign="top" align="left">0.08</td>
<td valign="top" align="left">0.07</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.14</td>
<td valign="top" align="left">0.14</td>
<td valign="top" align="left">0.14</td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left">0.20</td>
<td valign="top" align="left">0.20</td>
<td valign="top" align="left"><bold>0.20</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>&#x003C0;<sub>mis</sub> is the probability of missing value in each variable with missingness. The overall missing rate is &#x003C0;<sub>mis</sub>/2. Distribution width is the distance between the 2.5 and 97.5 percentile in the sampling distributions. Bias values with magnitudes &#x0003E;0.05, RMSEs &#x0003E;0.05, and ETI widths &#x0003E;.20 are indicated with bold fonts. The conditional formatting was applied prior to rounding</italic>.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>The bias, RMSE, and the 95% equal-tailed interval width of factor correlation FMIs under nonlinear MAR.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>&#x003C0;<sub><italic>mis</italic></sub></bold></th>
<th valign="top" align="left"><bold>N</bold></th>
<th valign="top" align="left" colspan="3"><bold>Bias</bold></th>
<th valign="top" align="left" colspan="3"><bold>RMSE</bold></th>
<th valign="top" align="left" colspan="3"><bold>95% ETI width</bold></th>
</tr>
<tr>
<th/>
<th/>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M66"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M67"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M68"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M69"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M70"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M71"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M72"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M73"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="left"><bold><inline-formula><mml:math id="M74"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left"><bold>0.12</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.42</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left"><bold>0.26</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.23</bold></td>
<td valign="top" align="left"><bold>0.15</bold></td>
<td valign="top" align="left"><bold>0.14</bold></td>
<td valign="top" align="left"><bold>0.90</bold></td>
<td valign="top" align="left"><bold>0.54</bold></td>
<td valign="top" align="left"><bold>0.47</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">100</td>
<td valign="top" align="left"><bold>0.17</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.30</bold></td>
<td valign="top" align="left"><bold>0.20</bold></td>
<td valign="top" align="left"><bold>0.19</bold></td>
<td valign="top" align="left"><bold>0.90</bold></td>
<td valign="top" align="left"><bold>0.76</bold></td>
<td valign="top" align="left"><bold>0.59</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left"><bold>0.22</bold></td>
<td valign="top" align="left">0.19</td>
<td valign="top" align="left">0.18</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left"><bold>0.13</bold></td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.09</bold></td>
<td valign="top" align="left"><bold>0.48</bold></td>
<td valign="top" align="left"><bold>0.40</bold></td>
<td valign="top" align="left"><bold>0.34</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">200</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.19</bold></td>
<td valign="top" align="left"><bold>0.15</bold></td>
<td valign="top" align="left"><bold>0.14</bold></td>
<td valign="top" align="left"><bold>0.82</bold></td>
<td valign="top" align="left"><bold>0.56</bold></td>
<td valign="top" align="left"><bold>0.47</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.03</td>
<td valign="top" align="left">0.12</td>
<td valign="top" align="left">0.12</td>
<td valign="top" align="left">0.11</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.05</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left"><bold>0.26</bold></td>
<td valign="top" align="left"><bold>0.20</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">500</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left"><bold>0.11</bold></td>
<td valign="top" align="left"><bold>0.10</bold></td>
<td valign="top" align="left"><bold>0.08</bold></td>
<td valign="top" align="left"><bold>0.43</bold></td>
<td valign="top" align="left"><bold>0.37</bold></td>
<td valign="top" align="left"><bold>0.32</bold></td>
</tr>
<tr>
<td valign="top" align="left">0.2</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.02</td>
<td valign="top" align="left">0.08</td>
<td valign="top" align="left">0.08</td>
<td valign="top" align="left">0.07</td>
</tr>
<tr>
<td valign="top" align="left">0.4</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.05</td>
<td valign="top" align="left">0.04</td>
<td valign="top" align="left">0.18</td>
<td valign="top" align="left">0.18</td>
<td valign="top" align="left">0.15</td>
</tr>
<tr>
<td valign="top" align="left">0.6</td>
<td valign="top" align="left">1000</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.00</td>
<td valign="top" align="left">0.01</td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.07</bold></td>
<td valign="top" align="left"><bold>0.06</bold></td>
<td valign="top" align="left"><bold>0.28</bold></td>
<td valign="top" align="left"><bold>0.26</bold></td>
<td valign="top" align="left"><bold>0.21</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>&#x003C0;<sub>mis</sub> is the probability of missing value in each variable with missingness. The overall missing rate is &#x003C0;<sub>mis</sub>/2. Distribution width is the distance between the 2.5 and 97.5 percentile in the sampling distributions. Bias values with magnitudes &#x0003E;0.05, RMSEs &#x0003E;0.05, and ETI widths &#x0003E;0.20 are indicated with bold fonts. The conditional formatting was applied prior to rounding</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Although the population FMIs of the factor correlations were quite low, they proved difficult to estimate precisely. As illustrated in <xref ref-type="fig" rid="F4">Figures 4</xref>&#x02013;<xref ref-type="fig" rid="F6">6</xref>, the sampling distributions of the FMI estimates were very wide, and showed considerable bias at small sample sizes. Although the bias was absent at larger sample sizes, the estimates would only fall closely around the population value when the missing rate was low. As seen in <xref ref-type="table" rid="T3">Tables 3</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref>, when the per variable missing rate was &#x003C0;<sub><italic>mis</italic></sub>=0.4, a sample size of 500 is required for <inline-formula><mml:math id="M101"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> to provide a precise estimate, with <inline-formula><mml:math id="M102"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="M103"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> giving worse performances. Under the high missing rate of &#x003C0;<sub><italic>mis</italic></sub> &#x0003D; 0.6, a sample size of 1,000 was required. Overall, <inline-formula><mml:math id="M104"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> gave the best performance, showing the lowest RMSE and 95% ETI width.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>The sampling distribution of FMI for factor correlations in MCAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0004.tif"/>
</fig>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>The sampling distribution of FMI for factor correlations in linear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>The sampling distribution of FMI for factor correlations in nonlinear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0006.tif"/>
</fig>
<p>Compared to the factor correlation FMIs, the factor loading FMIs performed much better in Model 2. Similar to the population simulation, here we report on the FMIs of the <italic>X</italic><sub>2</sub> loading on <italic>F</italic>1. The results for the other loadings <italic>X</italic><sub>4</sub>, <italic>Y</italic><sub>2</sub>, and <italic>Y</italic><sub>4</sub> are largely identical. Overall, as illustrated in <xref ref-type="fig" rid="F7">Figures 7</xref>&#x02013;<xref ref-type="fig" rid="F9">9</xref>, the sample FMIs of factor loadings showed little bias and would typically fall much closer to the population values. As shown in <xref ref-type="supplementary-material" rid="SM1">Supplementary Tables 5</xref>&#x02013;<xref ref-type="supplementary-material" rid="SM1">7</xref>, the bias, RMSE, and 95% ETI width were satisfactory for all three estimates when the sample size was <italic>N</italic> &#x0003D; 500 or greater. At <italic>N</italic> &#x0003D; 200, the FMIs only performed well when the missing mechanism was MCAR and the per variable missing rate was 0.2. At <italic>N</italic> &#x0003D; 100, all three estimates performed poorly, but <inline-formula><mml:math id="M105"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003BB;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was closer to acceptable performance, producing 95% ETI widths of 0.30 when <inline-formula><mml:math id="M106"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003BB;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> would produce widths close to 0.50, and <inline-formula><mml:math id="M107"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003BB;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> would produce widths close to 0.70. Once again, <inline-formula><mml:math id="M108"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003BB;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> showed the best performance overall, with the lowest RMSE and 95% ETI width, and a similar bias to <inline-formula><mml:math id="M109"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>&#x003BB;</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>The sampling distribution of FMI for factor loadings in MCAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0007.tif"/>
</fig>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>The sampling distribution of FMI for factor loadings in linear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0008.tif"/>
</fig>
<fig id="F9" position="float">
<label>Figure 9</label>
<caption><p>The sampling distribution of FMI for factor loadings in nonlinear MAR. The panel rows correspond to sample sizes of <italic>N</italic> &#x0003D; 50, 100, 200, and 500. The panel columns correspond to per variable missing rates of 20, 40, and 60%, or overall population missing rates of 10, 20, and 30%. The population FMI value in each panel is given as a vertical black dotted line.</p></caption>
<graphic xlink:href="fpsyg-12-667802-g0009.tif"/>
</fig></sec></sec></sec>
<sec id="s3">
<title>3. Example Analysis</title>
<p>Here we provide an example on how to obtain FMIs from the <italic>lavaan</italic> package (version 0.6-7 and up) in <italic>R</italic>, using the Holzinger and Swineford (<xref ref-type="bibr" rid="B7">1939</xref>) dataset and simulated MCAR missing data. The example is adapted from Savalei and Rosseel (ress). The dataset, available through the <italic>lavaan</italic> package, contains cognitive performance test scores from 301 school children. The data for this example can be loaded into the <italic>R</italic> workspace using the following code.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0001.tif"/></p>
<p>For the purpose of this demonstration, we will conduct a confirmatory factor analysis with three correlated factors: visual skills, verbal skills, and mental speed. Each factor is measured by three tests, with variable names <monospace>x1</monospace> to <monospace>x9</monospace> in the datasets, following a model given by the <italic>lavaan</italic> model syntax below.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0002.tif"/></p>
<p>As the dataset does not contain missing data, we will introduce MCAR missingness by randomly removing 61 values from each variable independently, resulting in an overall missing rate of 20%.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0003.tif"/></p>
<p>With the example data and model syntax prepared, we fit the data to the CFA model and obtain the FMIs with the following code.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0004.tif"/></p>
<p>The option <monospace>std.lv = TRUE</monospace> fixes all variances of the latent factors to 1, allowing all loadings to be freely estimated, while the <monospace>missing = &#x0007E;ml&#x0007E;</monospace> asks <italic>lavaan</italic> to handle the missing data using FIML. The function <monospace>parameterEstimates</monospace> extracts the results from the model fit, where the option <monospace>fmi = TRUE</monospace> requests FMI estimates from <italic>lavaan</italic> alongside the parameter estimates. The <monospace>remove.nonfree = TRUE</monospace> option omits parameters that are not freely estimated from the output&#x02014;In this case, the latent factor variances are not printed in the output table, as they were fixed to 1. By default, <italic>lavaan</italic> uses Hessian numeric estimation of the observed information matrix, which yields <inline-formula><mml:math id="M110"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> for the FMI estimates. To produce <inline-formula><mml:math id="M111"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, the <monospace>observed.information</monospace> input must be specified to request the analytic approximation of the information.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0005.tif"/></p>
<p>By default, the analytic approximation is based on structured information, to obtain <inline-formula><mml:math id="M112"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, the <monospace>h1.information</monospace> input must be provided to request the unstructured information.</p>
<p><inline-graphic xlink:href="fpsyg-12-667802-i0006.tif"/></p>
<p>The three FMI estimates provide largely similar values (see <xref ref-type="table" rid="T6">Table 6</xref>). They agree on which loading estimates have the highest FMIs, namely the loadings of <italic>X</italic><sub>1</sub> (<inline-formula><mml:math id="M113"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>29</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M114"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>27</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M115"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>29</mml:mn></mml:math></inline-formula>), <italic>X</italic><sub>2</sub> (<inline-formula><mml:math id="M116"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>26</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M117"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>25</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M118"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula>), <italic>X</italic><sub>3</sub> (<inline-formula><mml:math id="M119"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M120"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M121"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>31</mml:mn></mml:math></inline-formula>), and <italic>X</italic><sub>7</sub> (<inline-formula><mml:math id="M122"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>27</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M123"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>27</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M124"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula>). The three estimates also agree on which factor correlation has the highest FMI, namely the correlation between visual skill and mental speed, <inline-formula><mml:math id="M125"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>23</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M126"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>24</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M127"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>29</mml:mn></mml:math></inline-formula>. The FMI of the factor correlation between visual skill and mental speed also shows the largest difference among the FMI estimates, between 0.23 and 0.29. The three estimates disagree on which factor correlation has the lowest FMI, but the differences are small: For the correlation between visual and verbal skills, <inline-formula><mml:math id="M128"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>18</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M129"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>15</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M130"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>18</mml:mn></mml:math></inline-formula>; for the correlation between verbal skill and mental speed, <inline-formula><mml:math id="M131"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>18</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M132"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>18</mml:mn></mml:math></inline-formula>, <inline-formula><mml:math id="M133"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>20</mml:mn></mml:math></inline-formula>.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Results from the example analysis.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Parameter</bold></th>
<th valign="top" align="left"><bold>Variables</bold></th>
<th valign="top" align="center"><bold><inline-formula><mml:math id="M140"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="center"><bold><inline-formula><mml:math id="M141"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="center"><bold><inline-formula><mml:math id="M142"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Visual, X1</td>
<td valign="top" align="center">0.29</td>
<td valign="top" align="center">0.27</td>
<td valign="top" align="center">0.29</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Visual, X2</td>
<td valign="top" align="center">0.26</td>
<td valign="top" align="center">0.25</td>
<td valign="top" align="center">0.28</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Visual, X3</td>
<td valign="top" align="center">0.28</td>
<td valign="top" align="center">0.28</td>
<td valign="top" align="center">0.31</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Verbal, X4</td>
<td valign="top" align="center">0.24</td>
<td valign="top" align="center">0.22</td>
<td valign="top" align="center">0.24</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Verbal, X5</td>
<td valign="top" align="center">0.17</td>
<td valign="top" align="center">0.17</td>
<td valign="top" align="center">0.19</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Verbal, X6</td>
<td valign="top" align="center">0.19</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.19</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Speed, X7</td>
<td valign="top" align="center">0.27</td>
<td valign="top" align="center">0.27</td>
<td valign="top" align="center">0.28</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Speed, X8</td>
<td valign="top" align="center">0.23</td>
<td valign="top" align="center">0.24</td>
<td valign="top" align="center">0.26</td>
</tr>
<tr>
<td valign="top" align="left">Loading</td>
<td valign="top" align="left">Speed, X9</td>
<td valign="top" align="center">0.19</td>
<td valign="top" align="center">0.22</td>
<td valign="top" align="center">0.22</td>
</tr>
<tr>
<td valign="top" align="left">Factor correlation</td>
<td valign="top" align="left">Visual, verbal</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.15</td>
<td valign="top" align="center">0.18</td>
</tr>
<tr>
<td valign="top" align="left">Factor correlation</td>
<td valign="top" align="left">Visual, speed</td>
<td valign="top" align="center">0.23</td>
<td valign="top" align="center">0.24</td>
<td valign="top" align="center">0.29</td>
</tr>
<tr>
<td valign="top" align="left">Factor correlation</td>
<td valign="top" align="left">Verbal, speed</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.20</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Overall, the highest FMI estimate in the model is 0.31. The associated WIF is <inline-formula><mml:math id="M134"><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:msqrt><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>31</mml:mn></mml:mrow></mml:msqrt><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>.</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>, which indicates an estimated 20% increase of the width of the confidence interval due to missing data. Reporting these FMIs alongside analyses of empirical data will provide readers with a better sense of how much the presence of missing data has affected the efficiency of the parameter estimates, especially when comparing across studies where FMIs may differ even under the same missing rates. The full <italic>R</italic> code of this example is provided in <xref ref-type="supplementary-material" rid="SM2">Appendix A</xref>, and a summary of the lavaan options of the three estimates is given in <xref ref-type="supplementary-material" rid="SM1">Supplementary Table 1</xref>.</p></sec>
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>The current simulation study suggests that a relatively large sample size may be necessary for the estimation of the FMIs. Sample FMI estimates were largely unbiased, even in very small samples with <italic>N</italic> &#x0003D; 50, which are typical of regression analyses. However, at such a small sample size, the estimates were imprecise, varying greatly from sample to sample, especially when the missing rate was high. When the missing rates were reasonably low (&#x003C0;<sub><italic>mis</italic></sub> &#x0003D; 0.2, 0.4; corresponding to 10&#x02013;20% overall missing rate), sample sizes of several hundreds, which are typical in structural equation models, were able to produce reasonably efficient estimates. However, at an overall missing rate of 30%, it would require sample sizes exceeding 1,000 to produce precise estimates. It is worth noting that we used very strong selection mechanisms in the MAR-L and MAR-NL conditions to contrast the results from MCAR. In applied settings, the MAR selection mechanism would typically be weaker, which would lead to sample FMI estimate performances that are closer to better performances we saw in the MCAR conditions.</p>
<p>The three estimates are identical for saturated models, such as regression. However, the choice makes a difference when the model is not saturated. In the two-factor model, <inline-formula><mml:math id="M135"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the estimate via numeric Hessian, showed a distinct disadvantage as it was more likely break down when the sample size was small, or when the missing rate was high. In contrast, <inline-formula><mml:math id="M136"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the analytic estimate based on the unstructured model, was much less likely to break down in all cases, and was more precise than <inline-formula><mml:math id="M137"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Although <inline-formula><mml:math id="M138"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> occasionally showed a slightly higher bias than <inline-formula><mml:math id="M139"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, which was based on the structured model, its performance was overall more favorable in these simulations.</p>
<p>The simulation study investigated FMI in FIML, but the results should generalize to FMI computed from MI with a large number of imputations. In MI, the FMI is conceptually given by the ratio of the between-imputation variance over the sum of the within- and between- imputation variances. As the number of imputations approaches infinity, this ratio becomes equivalent to the ratio of variance increase due to missing data over variance in the observed data as estimated from FIML. For simulation studies, FMI can be more computationally expensive in MI, as the estimate is produced in the final pooling stage of the analysis, and often requires a large number of imputations (more than 100) to achieve an acceptable level of accuracy (Harel, <xref ref-type="bibr" rid="B6">2007</xref>). For substantive research, the researcher may simply choose between FIML and MI as the estimation method of FMI based on the missing data technique they are already using to produce the estimates of the model parameters.</p>
<p>As far as we are aware, this study is the first to look at the properties of sample FMIs computed using FIML. As such, we focused on two relatively simple and commonly used models, with three missing data mechanisms selected to contrast the impact of the specific mechanism on the FMI values and to stress that these values are not the same as the rates of missing data. Future research may wish to expand on the study conditions, for example, by controlling for the number of missing patterns, examining how changing the values of parameters in the model (such as the regression coefficient) would change the properties of the FMI estimates. It would also be worthwhile to investigate the relative performance of the three FMI estimates under incorrect models. When the model is wrong, the Hessian-based estimate, <inline-formula><mml:math id="M143"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, is theoretically superior, as it is the only consistent estimate. However, whether this theoretical advantage would translate into a practical advantage needs to be examined in simulation studies. It would also be helpful to develop bootstrap SE/CI for the sample FMIs, so that researchers would have a better sense of the precision of the FMI estimates in their particular sample.</p>
<p>While our focus was on evaluating the properties of sample FMI estimates in terms how well they served as estimates of the corresponding population FMIs, the properties of population FMIs themselves may be of interest, and is a topic we are exploring in other work. In this ongoing work, we are finding that information loss can occur in unintuitive and unpredictable ways, and patterns in population FMIs observed in one context do not always generalize to other context. For example, based on the population-level FMI values we obtained for the conditions in this study, one may be tempted to conclude that the population FMIs of factor correlations are, in general, insensitive to missing data mechanisms (e.g., the middle rows of <xref ref-type="table" rid="T1">Table 1</xref>). However, this was not always the case. In conditions not reported here, when the indicators of <italic>F</italic><sub>1</sub> were all completely observed, and the indicators of <italic>F</italic><sub>2</sub> contained missingness conditioned on the indicators of <italic>F</italic><sub>1</sub>, the FMI of the factor correlation became more sensitive to the missing data mechanism. Although the population quantities estimated by the sample FMIs may exhibit different patterns, we do not expect the sample FMI estimates themselves to show drastically different properties in these alternative scenarios.</p>
<p>The FMI estimates via FIML are available in most recent releases of <italic>lavaan</italic> (version 0.6-7 and up), and example <italic>R</italic> code for how to retrieve them is given in <xref ref-type="supplementary-material" rid="SM2">Appendix A</xref>. We recommend empirical researchers to routinely examine and report FMIs for key parameters in substantive analysis. FMIs capture the complex interplay between numerous factors such as missing rates, missing data mechanisms, and model parameters, sometimes in unintuitive ways. They can provide critical additional insights into how the standard errors, confidence intervals, and hypothesis tests may have been impacted by the presence of missing data. For methodologists conducting simulation studies with missing data, the population FMIs in the different study design conditions should be computed and reported, in order to provide better context of the performance of missing data techniques being studied.</p></sec>
<sec sec-type="data-availability-statement" id="s5">
<title>Data Availability Statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: <ext-link ext-link-type="uri" xlink:href="https://osf.io/xyzt8/">https://osf.io/xyzt8/</ext-link>.</p></sec>
<sec id="s6">
<title>Author Contributions</title>
<p>VS developed the equations, statistical properties, and details of implementation for FMI estimates. LC designed and carried out the simulation study and wrote the initial draft of the manuscript. All authors contributed to the revisions of the manuscript.</p></sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
<sec sec-type="disclaimer" id="s7">
<title>Publisher&#x00027;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec> </body>
<back>
<sec sec-type="supplementary-material" id="s8">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fpsyg.2021.667802/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fpsyg.2021.667802/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_2.pdf" id="SM2" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/></sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Allison</surname> <given-names>P. D.</given-names></name></person-group> (<year>1987</year>). <article-title>Estimation of linear models with incomplete data</article-title>. <source>Sociol. Methodol</source>. <volume>17</volume>, <fpage>71</fpage>&#x02013;<lpage>103</lpage>. <pub-id pub-id-type="doi">10.2307/271029</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Arbuckle</surname> <given-names>J. L.</given-names></name></person-group> (<year>1996</year>). <source>Full Information Estimation in the Presence of Incomplete Data</source>. <publisher-loc>Mahwah, NJ</publisher-loc>: <publisher-name>Lawrence Erlbaum Associates</publisher-name>.<pub-id pub-id-type="pmid">22411044</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>L.</given-names></name> <name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Rhemtulla</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Two-stage maximum likelihood approach for item-level missing data in regression</article-title>. <source>Behav. Res. Methods</source> <volume>52</volume>, <fpage>2306</fpage>&#x02013;<lpage>2323</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-020-01355-x</pub-id><pub-id pub-id-type="pmid">32333330</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Collins</surname> <given-names>L. M.</given-names></name> <name><surname>Schafer</surname> <given-names>J. L.</given-names></name> <name><surname>Kam</surname> <given-names>C.-M.</given-names></name></person-group> (<year>2001</year>). <article-title>A comparison of inclusive and restrictive strategies in modern missing data procedures</article-title>. <source>Psychol. Methods</source> <volume>6</volume>, <fpage>330</fpage>&#x02013;<lpage>351</lpage>. <pub-id pub-id-type="doi">10.1037/1082-989X.6.4.330</pub-id><pub-id pub-id-type="pmid">11778676</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Graham</surname> <given-names>J. W.</given-names></name> <name><surname>Taylor</surname> <given-names>B. J.</given-names></name> <name><surname>Olchowski</surname> <given-names>A. E.</given-names></name> <name><surname>Cumsille</surname> <given-names>P. E.</given-names></name></person-group> (<year>2006</year>). <article-title>Planned missing data designs in psychological research</article-title>. <source>Psychol. Methods</source> <volume>11</volume>, <fpage>323</fpage>&#x02013;<lpage>343</lpage>. <pub-id pub-id-type="doi">10.1037/1082-989X.11.4.323</pub-id><pub-id pub-id-type="pmid">17154750</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Harel</surname> <given-names>O.</given-names></name></person-group> (<year>2007</year>). <article-title>Inferences on missing information under multiple imputation and two-stage multiple imputation</article-title>. <source>Stat. Methodol</source>. <volume>4</volume>, <fpage>75</fpage>&#x02013;<lpage>89</lpage>. <pub-id pub-id-type="doi">10.1016/j.stamet.2006.03.002</pub-id><pub-id pub-id-type="pmid">23712802</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Holzinger</surname> <given-names>K. J.</given-names></name> <name><surname>Swineford</surname> <given-names>F.</given-names></name></person-group> (<year>1939</year>). <article-title>A study in factor analysis: the stability of a bi-factor solution</article-title>. <source>Suppl. Educ. Monogr</source>. <volume>48</volume>:<fpage>xi &#x0002B; 91</fpage>.</citation>
</ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Little</surname> <given-names>R. J. A.</given-names></name> <name><surname>Rubin</surname> <given-names>D. B.</given-names></name></person-group> (<year>1987</year>). <source>Statistical Analysis With Missing Data</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Wiley</publisher-name>.</citation>
</ref>
<ref id="B9">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Little</surname> <given-names>R. J. A.</given-names></name> <name><surname>Rubin</surname> <given-names>D. B.</given-names></name></person-group> (<year>2002</year>). <source>Statistical Analysis With Missing Data, 2nd Edn</source>. <publisher-loc>Wiley series in probability and statistics. Hoboken, NJ</publisher-loc>: <publisher-name>Wiley</publisher-name>. <pub-id pub-id-type="doi">10.1002/9781119013563</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Muth&#x000E9;n</surname> <given-names>B.</given-names></name> <name><surname>Kaplan</surname> <given-names>D.</given-names></name> <name><surname>Hollis</surname> <given-names>M.</given-names></name></person-group> (<year>1987</year>). <article-title>On structural equation modeling with data that are not missing completely at random</article-title>. <source>Psychometrika</source> <volume>52</volume>, <fpage>431</fpage>&#x02013;<lpage>462</lpage>. <pub-id pub-id-type="doi">10.1007/BF02294365</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Orchard</surname> <given-names>T.</given-names></name> <name><surname>Woodbury</surname> <given-names>M. A.</given-names></name></person-group> (<year>1972</year>). <article-title>&#x0201C;A missing information principle: theory and applications,&#x0201D;</article-title> in <source>Proceedings of the Sixth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Theory of Statistics</source> (<publisher-loc>Berkeley, CA</publisher-loc>: <publisher-name>University of California Press</publisher-name>), <fpage>697</fpage>&#x02013;<lpage>715</lpage>. <pub-id pub-id-type="doi">10.1525/9780520325883-036</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><collab>R Core Team</collab></person-group> (<year>2019</year>). <source>R: A Language and Environment for Statistical Computing</source>. <publisher-loc>Vienna</publisher-loc>: <publisher-name>R Foundation for Statistical Computing</publisher-name>.</citation>
</ref>
<ref id="B13">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Rao</surname> <given-names>C. R.</given-names></name></person-group> (<year>1973</year>). <source>Linear Statistical Inference and Its Applications</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>John Wiley &#x00026; Sons</publisher-name>. <pub-id pub-id-type="doi">10.1002/9780470316436</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rosseel</surname> <given-names>Y.</given-names></name></person-group> (<year>2012</year>). <article-title>lavaan: An R package for structural equation modeling</article-title>. <source>J. Stat. Softw</source>. <volume>48</volume>, <fpage>1</fpage>&#x02013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v048.i02</pub-id><pub-id pub-id-type="pmid">25601849</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rubin</surname> <given-names>D. B.</given-names></name></person-group> (<year>1976</year>). <article-title>Inference and missing data</article-title>. <source>Biometrika</source> <volume>63</volume>, <fpage>581</fpage>&#x02013;<lpage>592</lpage>. <pub-id pub-id-type="doi">10.1093/biomet/63.3.581</pub-id></citation>
</ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Rubin</surname> <given-names>D. B.</given-names></name></person-group> (<year>1987</year>). <source>Multiple Imputation for Nonresponse in Surveys</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>John Wiley &#x00026; Sons</publisher-name>. <pub-id pub-id-type="doi">10.1002/9780470316696</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name></person-group> (<year>2010</year>). <article-title>Expected versus observed information in sem with incomplete normal and nonnormal data</article-title>. <source>Psychol. Methods</source> <volume>15</volume>, <fpage>352</fpage>&#x02013;<lpage>367</lpage>. <pub-id pub-id-type="doi">10.1037/a0020143</pub-id><pub-id pub-id-type="pmid">20853954</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Bentler</surname> <given-names>P. M.</given-names></name></person-group> (<year>2005</year>). <article-title>A statistically justified pairwise ml method for incomplete nonnormal data: a comparison with direct ml and pairwise ADF</article-title>. <source>Struct. Equat. Model. Multidisc. J</source>. <volume>12</volume>, <fpage>183</fpage>&#x02013;<lpage>214</lpage>. <pub-id pub-id-type="doi">10.1207/s15328007sem1202_1</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Falk</surname> <given-names>C. F.</given-names></name></person-group> (<year>2014</year>). <article-title>Robust two-stage approach outperforms robust full information maximum likelihood with incomplete nonnormal data</article-title>. <source>Struct. Equat. Model. Multidisc. J</source>. <volume>21</volume>, <fpage>280</fpage>&#x02013;<lpage>302</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2014.882692</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Rhemtulla</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>On obtaining estimates of the fraction of missing information from full information maximum likelihood</article-title>. <source>Struct. Equat. Model. Multidisc. J</source>. <volume>19</volume>, <fpage>477</fpage>&#x02013;<lpage>494</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2012.687669</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Rhemtulla</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Normal theory two-stage ML estimator when data are missing at the item level</article-title>. <source>J. Educ. Behav. Stat</source>. <volume>42</volume>, <fpage>405</fpage>&#x02013;<lpage>431</lpage>. <pub-id pub-id-type="doi">10.3102/1076998617694880</pub-id><pub-id pub-id-type="pmid">29276371</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savalei</surname> <given-names>V.</given-names></name> <name><surname>Rosseel</surname> <given-names>Y.</given-names></name></person-group> (in press). <article-title>Computational options for standard errors test statistics with incomplete nonnormal data</article-title>. <source>Struct. Equat. Model.</source></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sullivan</surname> <given-names>T. R.</given-names></name> <name><surname>White</surname> <given-names>I. R.</given-names></name> <name><surname>Salter</surname> <given-names>A. B.</given-names></name> <name><surname>Ryan</surname> <given-names>P.</given-names></name> <name><surname>Lee</surname> <given-names>K. J.</given-names></name></person-group> (<year>2018</year>). <article-title>Should multiple imputation be the method of choice for handling missing data in randomized trials?</article-title> <source>Stat. Methods Med. Res</source>. <volume>27</volume>, <fpage>2610</fpage>&#x02013;<lpage>2626</lpage>. <pub-id pub-id-type="doi">10.1177/0962280216683570</pub-id><pub-id pub-id-type="pmid">28034175</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Buuren</surname> <given-names>S.</given-names></name> <name><surname>Groothuis-Oudshoorn</surname> <given-names>K.</given-names></name></person-group> (<year>2011</year>). <article-title>mice: Multivariate imputation by chained equations in R</article-title>. <source>J. Stat. Softw</source>. <volume>45</volume>, <fpage>1</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v045.i03</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yoo</surname> <given-names>J. E.</given-names></name></person-group> (<year>2009</year>). <article-title>The effect of auxiliary variables and multiple imputation on parameter estimation in confirmatory factor analysis</article-title>. <source>Educ. Psychol. Measure</source>. <volume>69</volume>, <fpage>929</fpage>&#x02013;<lpage>947</lpage>. <pub-id pub-id-type="doi">10.1177/0013164409332225</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yucel</surname> <given-names>R. M.</given-names></name> <name><surname>He</surname> <given-names>Y.</given-names></name> <name><surname>Zaslavsky</surname> <given-names>A. M.</given-names></name></person-group> (<year>2011</year>). <article-title>Gaussian-based routines to impute categorical variables in health surveys</article-title>. <source>Stat. Med</source>. <volume>30</volume>, <fpage>3447</fpage>&#x02013;<lpage>3460</lpage>. <pub-id pub-id-type="doi">10.1002/sim.4355</pub-id><pub-id pub-id-type="pmid">21976366</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>As the sample size approaches infinity, all parameter estimates approach their true values, and therefore all standard errors approach 0. Multiplying the standard errors by <inline-formula><mml:math id="M17"><mml:msqrt><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msqrt></mml:math></inline-formula> makes them converge to nonzero numbers.</p></fn>
<fn id="fn0002"><p><sup>2</sup>See <xref ref-type="supplementary-material" rid="SM2">Appendix A</xref> for how to request FMI estimates in <italic>lavaan</italic>. We would like to thank Yves Rossel, the developer of <italic>lavaan</italic>, for implementing these variations in the computation of FMI, permitting us to carry out this study.</p></fn>
</fn-group>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This research was supported by grant RGPIN-2015-05251 from the Natural Sciences and Engineering Research Council of Canada (NSERC) to VS.</p></fn>
</fn-group>
</back>
</article> 