<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Appl. Math. Stat.</journal-id>
<journal-title>Frontiers in Applied Mathematics and Statistics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Appl. Math. Stat.</abbrev-journal-title>
<issn pub-type="epub">2297-4687</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fams.2023.1095769</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Applied Mathematics and Statistics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A maximum likelihood approach for asymmetric non-normal data using a transformational measurement model</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Schweizer</surname> <given-names>Karl</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/433057/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>DiStefano</surname> <given-names>Christine</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/608150/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>French</surname> <given-names>Brian</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/100720/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Institute of Psychology, Goethe University Frankfurt</institution>, <addr-line>Frankfurt</addr-line>, <country>Germany</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Educational Studies, University of South Carolina</institution>, <addr-line>Columbia, MO</addr-line>, <country>United States</country></aff>
<aff id="aff3"><sup>3</sup><institution>College of Education, Washington State University</institution>, <addr-line>Pullman, WA</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Han-Ying Liang, Tongji University, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Steffen Zitzmann, University of T&#x000FC;bingen, Germany; Irini Moustaki, London School of Economics and Political Science, United Kingdom</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Karl Schweizer <email>k.schweizer&#x00040;psych.uni-frankfurt.de</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Statistics and Probability, a section of the journal Frontiers in Applied Mathematics and Statistics</p></fn>
<fn fn-type="other" id="fn002"><p>&#x02020;ORCID: Karl Schweizer <ext-link ext-link-type="uri" xlink:href="https://orcid.org/0000-0002-3143-2100">orcid.org/0000-0002-3143-2100</ext-link></p></fn></author-notes>
<pub-date pub-type="epub">
<day>28</day>
<month>03</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>9</volume>
<elocation-id>1095769</elocation-id>
<history>
<date date-type="received">
<day>11</day>
<month>11</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>13</day>
<month>03</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Schweizer, DiStefano and French.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Schweizer, DiStefano and French</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>A transformational measurement model for structural equation modeling (SEM) of asymmetric non-normal data is proposed. This measurement model aligns with the expectation-maximization (EM) algorithm of the maximum likelihood estimation (MLE) method, creating adaptability to data that deviate from normality. Distinctive properties of the connection of the measurement model and EM algorithm are maintenance of the normality assumption, which is at the core of EM algorithm, and applicability to asymmetric non-normality of observed data mediated by distortion coefficients. An evaluation using a mixture of normal and severely asymmetric non-normal data analyzed by MLE for asymmetric non-normal data (MLE for ASN) demonstrated efficiency of the model. Comparisons with robust DWLS and WLS yielded better fit results under MLE for ASN estimation.</p></abstract>
<kwd-group>
<kwd>maximum likelihood estimation</kwd>
<kwd>confirmatory factor analysis</kwd>
<kwd>EM algorithm</kwd>
<kwd>DWLS</kwd>
<kwd>WLS</kwd>
</kwd-group>
<counts>
<fig-count count="0"/>
<table-count count="3"/>
<equation-count count="15"/>
<ref-count count="29"/>
<page-count count="9"/>
<word-count count="7210"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>A popular method for parameter estimation in structural equation modeling (SEM) is the maximum likelihood estimation (MLE) method which was introduced in the late 1960&#x00027;s [<xref ref-type="bibr" rid="B1">1</xref>]. A major component of this estimation technique is the maximum likelihood function derived from the density function of a multivariate normal distribution. Recently, the long-standing popularity of this estimation method has been hampered, as the estimator has come under scrutiny because the validity of MLE outcomes is impaired when investigating non-normal data [<xref ref-type="bibr" rid="B2">2</xref>&#x02013;<xref ref-type="bibr" rid="B4">4</xref>]. These outcomes suggest the restriction of MLE to investigations of data which closely follows a normal distribution.</p>
<p>Since much of the data investigated in applied situation are non-normally distributed, replacement and modification of MLE are options to explore. There are multiple ways to conduct alternative MLE including modified fitting functions [<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B6">6</xref>] and <italic>post-hoc</italic> corrections which take data characteristics into consideration (e.g., Satorra-Bentler method [<xref ref-type="bibr" rid="B7">7</xref>]). Furthermore, estimation methods based on a least-square approach are available [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B9">9</xref>]. Recent studies compare two popular approaches: robust MLE with diagonally weighted least square estimation [<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>]. These methods perform equally well in the presence of data commonly encountered by researchers such as: number of ordinal categories between four and ten, sample sizes between 200 and 1,000, and symmetric as well as asymmetric distributions. But, aside from the overall similarity in many situations, when MLE was applied to non-normal (asymmetrically distributed) data, parameter estimates by MLE yielded larger bias.</p>
<p>This essay reports on another possible modification of MLE for the purpose of improving its performance when analyzing data showing various degrees of non-normality. It takes the individual deviation of each observed variable into consideration and is not restricted to estimates of model fit. This modification is accomplished by a transformational measurement model that exerts influence on parameter estimation by use of the expectation-maximization (EM) algorithm [<xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B13">13</xref>]. We characterize it as transformational because it is expected to shift the range of applicability from normally distributed data to asymmetric non-normal data.</p>
<p>In a way, the modification brought about by the transformational measurement model narrows the gap between MLE in factor analysis (including SEM) and outside of factor analysis where normality is rarely observed [<xref ref-type="bibr" rid="B14">14</xref>]. While in the context of factor analysis, MLE expects normally distributed data; however, in other applications MLE is recommended for dealing with non-normality [<xref ref-type="bibr" rid="B15">15</xref>]. An elaboration in line with this recommendation, for example, is an adaptation of MLE to the Rayleigh distribution [<xref ref-type="bibr" rid="B16">16</xref>].</p>
<p>For establishing the transformational measurement model as source of influence on EM algorithm, components of the model of the covariance matrix of MLE have to be related to corresponding components of the EM algorithm. This means re-parametrization of EM algorithm [<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B18">18</xref>]. In the course of re-parametrization it is necessary to assure that the performance of the algorithm is not impaired and that at the same time deviation from normality is taken into consideration.</p>
<p>Since non-normality as descriptive term applies to a variety of different distributions that may require different treatments, we restrict our research to a subset of non-normal distributions. The focus is on distributions that can be perceived as derived from a normal distribution by distortion (e.g., due to a scale restriction) in such a way that some properties are retained: the continuous scale, the existence of a major (i.e., global) peak as well as a mostly monotone increase up to this peak and a mostly monotone decrease following the peak. A major difference between original normal and non-normal data distributions is lack of symmetry. We characterize such distributions as <italic>asymmetric non-normal</italic>, abbreviated as <italic>asn</italic> (see <xref ref-type="table" rid="T1">Tables 1</xref>&#x02013;<bold>3</bold>). The version of MLE described in this paper is optimized for investigating asn data.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Mean fit results and SDs (in italics) observed in investigating data showing (no) asymmetric non-normality (ASN) by one-factor models using regular MLE and MLE for ASN (<italic>N</italic> = 500).</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496;color:#ffffff">
<th valign="top" align="left"><bold>Characteristics of data (and estimation)</bold></th>
<th/>
<th valign="top" align="center"><bold>&#x003C7;<sup>2</sup></bold></th>
<th valign="top" align="center"><bold>RMSEA</bold></th>
<th valign="top" align="center"><bold>SRMR</bold></th>
<th valign="top" align="center"><bold>NNFI</bold></th>
<th valign="top" align="center"><bold>CFI</bold></th>
<th valign="top" align="center"><bold>AIC</bold></th>
</tr>
</thead>
<tbody>
<tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="8"><bold>Expected first-level factor loadings (0.40)</bold></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">No ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">258.0</td>
<td valign="top" align="center">0.021</td>
<td valign="top" align="center">0.057</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">300.0</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>78.3</italic></td>
<td valign="top" align="center"><italic>0.018</italic></td>
<td valign="top" align="center"><italic>0.016</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>78.3</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">300.3</td>
<td valign="top" align="center">0.034</td>
<td valign="top" align="center">0.069</td>
<td valign="top" align="center">0.95</td>
<td valign="top" align="center">0.95</td>
<td valign="top" align="center">342.3</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>33.0</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>33.0</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLE for ASN)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">229.9</td>
<td valign="top" align="center">0.019</td>
<td valign="top" align="center">0.054</td>
<td valign="top" align="center">0.98</td>
<td valign="top" align="center">0.98</td>
<td valign="top" align="center">271.9</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>25.7</italic></td>
<td valign="top" align="center"><italic>0.008</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>25.7</italic></td>
</tr> <tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="8"><bold>Expected second-level factor loadings (0.35)</bold></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">No ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">235.5</td>
<td valign="top" align="center">0.018</td>
<td valign="top" align="center">0.052</td>
<td valign="top" align="center">0.96</td>
<td valign="top" align="center">0.96</td>
<td valign="top" align="center">277.5</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>56.6</italic></td>
<td valign="top" align="center"><italic>0.015</italic></td>
<td valign="top" align="center"><italic>0.011</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>56.6</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">262.7</td>
<td valign="top" align="center">0.027</td>
<td valign="top" align="center">0.059</td>
<td valign="top" align="center">0.94</td>
<td valign="top" align="center">0.94</td>
<td valign="top" align="center">304.7</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>27.9</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>27.9</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLE for ASN)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">216.0</td>
<td valign="top" align="center">0.015</td>
<td valign="top" align="center">0.049</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">258.0</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>24.0</italic></td>
<td valign="top" align="center"><italic>0.008</italic></td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>24.0</italic></td>
</tr> <tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="8"><bold>Expected third-level factor loadings (0.30)</bold></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">No ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">217.6</td>
<td valign="top" align="center">0.014</td>
<td valign="top" align="center">0.048</td>
<td valign="top" align="center">0.96</td>
<td valign="top" align="center">0.95</td>
<td valign="top" align="center">259.6</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>39.8</italic></td>
<td valign="top" align="center"><italic>0.012</italic></td>
<td valign="top" align="center"><italic>0.007</italic></td>
<td valign="top" align="center"><italic>0.06</italic></td>
<td valign="top" align="center"><italic>0.05</italic></td>
<td valign="top" align="center"><italic>39.8</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLE)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">234.7</td>
<td valign="top" align="center">0.021</td>
<td valign="top" align="center">0.052</td>
<td valign="top" align="center">0.93</td>
<td valign="top" align="center">0.93</td>
<td valign="top" align="center">276.7</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>24.2</italic></td>
<td valign="top" align="center"><italic>0.006</italic></td>
<td valign="top" align="center"><italic>0.003</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>24.2</italic></td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">ASN (MLA for ASN)</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">205.6</td>
<td valign="top" align="center">0.012</td>
<td valign="top" align="center">0.046</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">247.6</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>22.2</italic></td>
<td valign="top" align="center"><italic>0.008</italic></td>
<td valign="top" align="center"><italic>0.003</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>22.2</italic></td>
</tr></tbody>
</table>
</table-wrap>
<p>The specification for asymmetric non-normal data is a characteristic that distinguishes the proposed version of MLE from other versions that are considered for non-normal data in general. Another characteristic is that parameters and fit statistics are estimated in the framework of a model reflecting the deviation of data from normality. This means that there is no <italic>post-hoc</italic> correction as found in other MLE versions. We refer to the proposed version of MLE as MLE for ASN.</p>
<p>The research reported in this paper was guided by two aims: (1) conceptualization of a version of MLE for asymmetric non-normal data within the framework of regular MLE for SEM by combining it with a transformational measurement model, and (2) demonstrating the efficiency of this version in investigating the structure of asymmetric non-normal data. MLE for ASN was also applied to asymmetric non-normal data with one-dimensional and two-dimensional underlying structures and compared with weighted least squares and diagonally weighted least squares methods.</p>
<p>The theoretical argument shows the following structure: at first, the model of the covariance matrix of MLE [<xref ref-type="bibr" rid="B19">19</xref>] is related to corresponding components of EM algorithm [<xref ref-type="bibr" rid="B13">13</xref>]. Subsequently, changes are introduced to fit the model to asymmetric non-normal data. Finally, the transformational measurement model is extracted.</p>
<sec>
<title>The outset</title>
<p>We begin this investigation by considering the MLE fitting function, F, that is proposed for fitting the parameters of the theoretical model to data <italic>via</italic> the maximum likelihood criterion [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B9">9</xref>, <xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B17">17</xref>]. This outset is selected since it provides information on what must be taken into consideration in order to demonstrate acceptable model-data fit. Let <bold>S</bold> (<bold>S</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>) be the <italic>p</italic> &#x000D7; <italic>p</italic> empirical covariance matrix, <bold>&#x003A3;</bold> (<bold>&#x003A3;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>) the <italic>p</italic> &#x000D7; <italic>p</italic> model-implied covariance matrix that is specified by assigning value(s) to parameter(s), &#x003D1;, and F() the MLE fitting function for model evaluation. MLE by definition refers to an iterative search for &#x003D1; which provides the best fit of <bold>&#x003A3;</bold> to <bold>S</bold> on the basis of</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mtext>F</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003D1;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Whereas <bold>S</bold> is given and <bold>&#x003A3;</bold> reflects the assumptions of the tested model, &#x003D1; is estimated while minimizing F in a succession of cycles of EM algorithm [<xref ref-type="bibr" rid="B12">12</xref>].</p>
<p>The model-implied covariance matrix, <bold>&#x003A3;</bold>(<bold>&#x003A3;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>), can show different degrees of complexity. For the purpose of the present study, we concentrate on the basic version with one latent variable for centered data associated with the congeneric measurement model [<xref ref-type="bibr" rid="B20">20</xref>] typically used with confirmatory factor analysis and structural equation modeling. This model includes two components that are expected to account for common and unique variation of data, respectively:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mo>&#x003A3;</mml:mo><mml:mo>=</mml:mo><mml:mo>&#x003BB;</mml:mo><mml:mi>&#x003D5;</mml:mi><mml:msup><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x00398;</mml:mi><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>with <italic>p</italic> &#x000D7; 1 vector <bold>&#x003BB;</bold> (<bold>&#x003BB;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;1</sup>) of factor loadings on the latent variable of the corresponding measurement model, &#x003BE; [&#x003BE; &#x0007E; <italic>N</italic>(0, &#x003C3;&#x0002A;)], variance parameter &#x003C6;(&#x003C6; &#x02208; &#x0211C;<sup>&#x0002B;</sup>), and <italic>p</italic> &#x000D7; <italic>p</italic> diagonal matrix of residual variances <bold>&#x00398;</bold>(<bold>&#x00398;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>). Since <bold>&#x003BB;</bold> and &#x003D5; cannot be estimated at the same time, scaling is necessary [<xref ref-type="bibr" rid="B21">21</xref>]. This is frequently done by setting &#x003D5; = 1.</p>
<p><bold>&#x003A3;</bold> can also be viewed as a matrix composed of variances and covariances. Since variances are the more complex entries, which include two components, it is sufficient to concentrate the discussion on these components. The variance of <italic>i</italic>th manifest variable, &#x003C3;<sub><italic>ii</italic></sub> <inline-formula><mml:math id="M3"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002B;</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>), is defined as</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mi>&#x003D5;</mml:mi><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003BB;<sub><italic>i</italic></sub> (&#x003BB;<sub><italic>i</italic></sub> &#x02208; &#x0211C;) is the factor loading of the <italic>i</italic>th manifest variable on latent variable &#x003BE;, &#x003D5;(&#x003C6; &#x02208; &#x0211C;<sup>&#x0002B;</sup>) the variance parameter and &#x003B8;<sub><italic>i</italic></sub> <inline-formula><mml:math id="M5"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002B;</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> the residual variance. In this model, &#x003D5; measures the dispersion of the latent variable that is common to all manifest variables and, &#x003B8;<sub><italic>i</italic></sub>, the dispersion that is unique for the <italic>i</italic>th manifest variable.</p>
<p>In MLE the estimation of parameters is conducted by the EM algorithm. This method is a tool for the iterative estimation of parameters according to the maximum likelihood principle [<xref ref-type="bibr" rid="B22">22</xref>]. It is employed in situations such as missing information under the normal distribution assumption and mixture distributions [<xref ref-type="bibr" rid="B23">23</xref>]. The EM algorithm is considered important throughout statistics. Of importance for the present research is its application to factor analysis [<xref ref-type="bibr" rid="B13">13</xref>]. Employed as part of MLE, EM performs parameter estimation under the assumption of the normal distribution. Therefore, in the present study the consequences of deviations from normality reflected by the transformational measurement model for parameter estimation by EM algorithm need to be analyzed in some detail.</p>
<p>As the EM algorithm for factor analysis plays a major role in MLE for ASN, we proceed by recognizing the relationship between the parameters of Equation 3 and the parameters of EM algorithm. Furthermore, distributional assumptions need to be taken into consideration. EM-theory for factor analysis [<xref ref-type="bibr" rid="B13">13</xref>] assumes latent variables (i.e., factors) established by unobservable factor scores, which are symbolized as <italic>Z</italic> (<italic>Z</italic> &#x02208; &#x0211C;<sup><italic>n</italic>&#x000D7;<italic>p</italic></sup>). These factor scores are assumed to follow a normal distribution. This means that latent variables can be considered as normal through factor scores. EM theory also considers observed scores that are symbolized by <italic>Y</italic> (<italic>Y</italic> &#x02208; &#x0211C;<sup><italic>n</italic>&#x000D7;<italic>q</italic></sup>). In MLE they are also assumed to follow a normal distribution [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B19">19</xref>].</p>
<p>Adaptation of the model described by Equation 3 to the notation of EM algorithm requires linking parameters &#x003BB;, &#x003B8; and &#x003D5; of <bold>&#x003BB;</bold>, &#x003D5; and <bold>&#x00398;</bold> to parameters of &#x003B2;, &#x003C4;<sup>2</sup> and <italic>R</italic> [<xref ref-type="bibr" rid="B13">13</xref>] in corresponding order. Matrix &#x003B2; is defined as the &#x0201C;regression coefficient matrix&#x0201D; that &#x0201C;is commonly called factor-loading matrix&#x0201D; (p. 70), and matrix &#x003C4;<sup>2</sup> as matrix &#x0201C;of residual variances &#x02026; commonly called the uniquenesses&#x0201D; (p. 70). Furthermore, matrix <italic>R</italic> is defined as correlation matrix of the unobservable factor scores. We assume dimensionalities and data types for the EM matrices corresponding to the matrices of Equation 2. The symbol equivalent to &#x003D5; is <italic>R</italic><sup>&#x0002A;</sup> since in a one-factor model <italic>R</italic> reduces to a 1 &#x000D7; 1 matrix, i.e., a scalar that is signified by the added star. It can be set free for estimation (see [<xref ref-type="bibr" rid="B13">13</xref>], Case 3). After a few EM cycles, this free parameter is likely to no more demonstrate the property of a (Pearson) correlation but may serve as estimate of common latent variation. Finally, &#x003C3;<sub><italic>ii</italic></sub> (Equation 3) is re-written using parameters of &#x003B2;, &#x003C4;<sup>2</sup> and <italic>R</italic> as</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mi>R</mml:mi><mml:mo>*</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
</sec>
<sec>
<title>The preparation of EM algorithm for data non-normality</title>
<p>In this section, the model for normally distributed data is slightly modified to make it suitable for investigating asymmetric non-normal data while the basic definitions and assumptions of EM algorithm except of the ones regarding <italic>Y</italic> are retained.</p>
<p>The investigation of asymmetric non-normal data requires the establishment of a relationship between normally distributed data and asymmetric non-normal data. Recall at this point that asymmetric non-normal data are assumed to originate from normally distributed data by distortion (see Introductory section). We consider such a relationship at the level of variances (and covariances) since variances (and covariances) serve as input to SEM. We assume that the effect of distortion is reflected by distortion coefficient <inline-formula><mml:math id="M7"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> <inline-formula><mml:math id="M8"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002B;</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) linking the variance of the manifest variable representing asymmetric non-normal data [&#x003C3;<sub><sub><italic>ii</italic></sub>(asn)</sub>] to the variance of the (corresponding) manifest variable representing normally distributed data [&#x003C3;<sub><sub><italic>ii</italic></sub>(n)</sub>]:
<disp-formula id="E5"><label>(5)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
In the following several transformations are described that are necessary for performing parameter estimation that is in line with Equation 5 by available SEM software including EM algorithm. In the first step, &#x003C3;<sub><sub><italic>ii</italic></sub>(n)</sub> is replaced by its constituents (see Equation 3) so that its common and residual components can be treated separately. In the next step, the distortion coefficient is associated with each one of the two components so that</p>
<disp-formula id="E6"><label>(6)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mi>&#x003D5;</mml:mi><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In the following steps, <inline-formula><mml:math id="M11"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> and &#x003B8;<sub><italic>i</italic></sub> are merged to give &#x003B8;<sub><sub><italic>i</italic></sub>(asn)</sub> on one hand and on the other hand scalar <inline-formula><mml:math id="M12"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> <inline-formula><mml:math id="M13"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002B;</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is subdivided and used for multiplying the common component included in parentheses from the left-hand side and the right-hand side. Since &#x003D5; represents the dispersion of normally distributed latent variable &#x003BE;, this is signified by adding a subscript; it is written as &#x003D5;<sub>(n)</sub>:</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>asn</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>n</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mo>&#x003BB;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>asn</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Finally, we introduce an assumption that enables equal treatments of all manifest variables. It is assumed that the latent variable equally contributes to all manifest variables. This is a useful assumption for simulation studies and can also apply to empirical data under appropriate conditions. It allows setting &#x003BB;<sub><italic>i</italic></sub> = 1 (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) so that the product of variance parameter and factor loadings, &#x003BB;<sub><italic>i</italic></sub> &#x003D5; &#x003BB;<sub><italic>i</italic></sub>, can be replaced by &#x003D5;<sub>common(n)</sub> (= 1&#x003D5;<sub>(n)</sub>1) that is the same for all manifest variables (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>):</p>
<disp-formula id="E8"><label>(8)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>asn</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mtext>common</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>n</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>asn</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>This expression includes a parameter that reflects normality of the latent variable, &#x003D5;<sub>common(n)</sub>, and distinguishes parts that relate to asymmetric non-normality, &#x003B1;<sub><italic>i</italic></sub>, or are influenced by it, &#x003B8;<sub><sub><italic>i</italic></sub>(asn)</sub>. It implies re-scaling: &#x003D5; is set free while the replacement of the factor loading is fixed.</p>
<p>Following the final transformation we note the compatibility with EM algorithm. According to Case 3 of EM algorithm [<xref ref-type="bibr" rid="B13">13</xref>], &#x003B2;<sub><italic>i</italic></sub> of &#x003B2;<sub><italic>i</italic></sub><inline-formula><mml:math id="M16"><mml:msup><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msup><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (see Equation 4) can be fixed while <italic>R</italic><sup>&#x0002A;</sup> be free. Therefore, we set &#x003B2;<sub><italic>i</italic></sub> equal to &#x003B1;<sub><italic>i</italic></sub> (Equation 8) and <italic>R</italic><sup>&#x0002A;</sup> equal to &#x003D5;<sub>common(n)</sub> (Equation 8). Furthermore, since the description of EM algorithm does not includes a distributional assumption regarding &#x003C4;<sup>2</sup>, &#x003C4;<sup>2</sup> is set equal to &#x003B8;<sub><sub><italic>i</italic></sub>(asn)</sub>. Writing Equation 8 accordingly gives</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>asn</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>EM</mml:mtext><mml:mo>,</mml:mo><mml:mtext>fixed</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>EM</mml:mtext><mml:mo>,</mml:mo><mml:mtext>estimated</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>EM</mml:mtext><mml:mo>,</mml:mo><mml:mtext>fixed</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>EM</mml:mtext><mml:mo>,</mml:mo><mml:mtext>estimated</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>that is in line with Equation 4. Subscripts added; parentheses indicate a role in EM algorithm and whether the element is estimated or fixed.</p>
<p>Parameter estimation by EM algorithm occurs in cycles. Within each cycle <italic>R</italic><sup>&#x0002A;</sup> under the condition of Case 3 is estimated by means of a regression coefficient, &#x003B4;, applied to the empirical covariance matrix. Here, &#x003B4; is determined by &#x003B2; as well as the previous estimate of <italic>R</italic><sup>&#x0002A;</sup> [<xref ref-type="bibr" rid="B13">13</xref>]. Furthermore, residual matrix, &#x00394;, with &#x003B2; and the previous estimate of <italic>R</italic><sup>&#x0002A;</sup> as components also exerts influence on revised <italic>R</italic><sup>&#x0002A;</sup> [<xref ref-type="bibr" rid="B13">13</xref>]. This means that the fixed &#x003B2; (i.e., &#x003B1;<sup>2</sup>) parameter upholds the relationship between what is expected to follow the normal distribution and what appears as distributional distortion.</p>
</sec>
<sec>
<title>Quantification of deviation from normality</title>
<p>This section presents a description of quantifying distortion coefficients, <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) (see Equation 5). Distortion coefficients are assumed to reflect the effect of distorting normally distributed data regarding their variances. Since &#x003C3;<sub><sub><italic>ii</italic></sub>(n)</sub> and &#x003C3;<sub><sub><italic>ii</italic></sub>(asn)</sub> of Equation 5 are unknown, <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> is not obtainable by rearranging the ingredients of Equation 5. Here, what is possible is using the possible proportionality of <inline-formula><mml:math id="M20"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) and &#x003C3;<sub><sub><italic>ii</italic></sub>(asn)</sub> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>). For this purpose all &#x003B1;s and &#x003C3;<sub>(asn)</sub>s are arranged as <italic>p</italic> &#x000D7; 1 vectors. They show proportionality if &#x003C3;<sub><sub><italic>ii</italic></sub>(n)</sub> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) can be assumed to be constant:</p>
<disp-formula id="E10"><label>(10)</label><mml:math id="M21"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x022EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x0221D;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>11</mml:mn><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>22</mml:mn><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>33</mml:mn><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x022EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>p</mml:mi><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The equality assumption of factor loadings leading to Equation 8 justifies Equation 10.</p>
<p>Although &#x003C3;<sub>(asn)</sub>s are unknown, they demonstrate a property that can be employed for arriving at values that approximate &#x003B1;s. This property is that each &#x003C3;<sub><sub><italic>ii</italic></sub>(asn)</sub> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) is expected to reflect the observed variance of corresponding asymmetric non-normal data, <italic>s</italic><sub><italic>ii</italic></sub> (<italic>i</italic> = 1, &#x02026;, <italic>p</italic>) of <bold>S</bold>. This expectation implies that &#x003C3;<sub><sub><italic>ii</italic></sub>(asn)</sub> shows a size similar to the size of <italic>s</italic><sub><italic>ii</italic></sub>:</p>
<disp-formula id="E11"><label>(11)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi><mml:mtext>(asn)</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02248;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Together Equations 10 and 11 suggest a vector of variances obtained from distorted data for adjusting the statistical model to the effect of distortion. Furthermore, Equations 8 and 9 suggest the replacement of &#x003B1;<sup>2</sup>s by their square roots as well as <italic>s</italic>s by their square roots: let <inline-formula><mml:math id="M23"><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> <inline-formula><mml:math id="M24"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> be the <italic>p</italic> &#x000D7; 1 vector of observation-based distortion coefficients corresponding to the standard deviations observed in distorted data. Then</p>
<disp-formula id="E12"><label>(12)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>.</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>.</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>.</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The hat (<sup>&#x02227;</sup>) on top of <bold>&#x003B1;</bold> and &#x003B1;s is added to signify that observations serve as source.</p>
</sec>
<sec>
<title>The transformational measurement model for asymmetric non-normal data</title>
<p>Finally, the transformational measurement model is achieved by determining the measurement model associated with Equation 8 (i.e., finding the measurement model giving rise to Equation 8) with deviation coefficients according to Equation 12:</p>
<disp-formula id="E13"><label>(13)</label><mml:math id="M26"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle><mml:mo>=</mml:mo><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mi>&#x003BE;</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mstyle mathvariant="bold"><mml:mi>&#x003B4;</mml:mi></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>including <italic>p</italic> &#x000D7; 1 vector <bold>x</bold> (<bold>x</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;1</sup>) of centered manifest variables, <italic>p</italic> &#x000D7; 1 vector <inline-formula><mml:math id="M27"><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula><inline-formula><mml:math id="M28"><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211C;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> of distortion-reflecting fixations replacing fixed factor loadings (see Equation 12), latent variable &#x003BE; and <italic>p</italic> &#x000D7; 1 vector <bold>&#x003B4;</bold> (<bold>&#x003B4;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;1</sup>)of residual variables. This model is proposed for investigating the hypothesis that the systematic variation of data is due to one underlying dimension (i.e., structural validity) if data show asymmetric non-normality.</p>
</sec>
<sec>
<title>Checking the compatibility with fitting function F</title>
<p>This section addresses the question whether fitting function F (Equation 1) is appropriate for asymmetric non-normal data. There is the original version of F derived from the density function of normally distributed data [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B15">15</xref>]. Furthermore, there is an extended version that is implemented in current SEM software. It is an extended version that shows a special characteristic which broadens the field of application. It is defined as</p>
<disp-formula id="E14"><label>(14)</label><mml:math id="M29"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mtext>F</mml:mtext><mml:mo>=</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mo>|</mml:mo><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle><mml:mo>|</mml:mo><mml:mo>&#x0002B;</mml:mo><mml:mtext>tr</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mo>|</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle><mml:mo>|</mml:mo><mml:mo>-</mml:mo><mml:mi>p</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <bold>&#x003A3;</bold> (<bold>&#x003A3;</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>) and <bold>S</bold> (<bold>S</bold> &#x02208; &#x0211C;<sup><italic>p</italic>&#x000D7;<italic>p</italic></sup>) represent the <italic>p</italic> &#x000D7; <italic>p</italic> model-implied and empirical covariance matrices in corresponding order and positive integer <italic>p</italic> the number of manifest variables [<xref ref-type="bibr" rid="B9">9</xref>, <xref ref-type="bibr" rid="B17">17</xref>].</p>
<p>The special characteristic of this version is that in the case of a correct model (including correct parameter estimates), <bold>&#x003A3;</bold><sub>c</sub>, and data (an empirical covariance matrix) <italic>free</italic> of random influences, <bold>S</bold><sub>f</sub>: log|<bold>&#x003A3;</bold><sub>c</sub>| = log|<bold>S</bold><sub>f</sub>|, so that their difference is zero. Furthermore, in this case: <bold>S</bold><sub>f</sub> &#x000D7; <bold>&#x003A3;</bold><sub>c</sub><sup><bold>&#x02212;1</bold></sup> = I and tr(I) = <italic>p</italic>. Moreover, the difference of the trace of <bold>S</bold><sub>f</sub> &#x000D7; <bold>&#x003A3;</bold> <sub>c</sub><sup><bold>&#x02212;1</bold></sup> and <italic>p</italic> is also zero so that,</p>
<disp-formula id="E15"><label>(15)</label><mml:math id="M30"><mml:mtable class="eqnarray" columnalign="center"><mml:mtr><mml:mtd><mml:mtext>F</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>c</mml:mtext></mml:mrow></mml:msub><mml:mtext>,</mml:mtext><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>f</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>c</mml:mtext></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:mo>&#x0002B;</mml:mo><mml:mtext>tr</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>f</mml:mtext></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x003A3;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>c</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>S</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext>f</mml:mtext></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:mo>-</mml:mo><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In sum, in the case a correct model and data with no error influence, F can be expected to arrive at a value of zero, and it is not necessary to make assumptions regarding the distribution of data (but, the matrices involved in the equation must allow for the computation of a determinant and inverse).</p>
<p>The presence of random influences may not invalidate the previous reasoning. In this case the model has to correctly account for the systematic variation of data so that only unsystematic variation due to random influences remains that is reflected by the outcome of F. If there is reason for assuming that deviations from expectations due to random influences follow normal distributions, the outcome can be considered as a &#x003C7;<sup>2</sup> statistic, as is in investigating normally distributed data.</p>
<p>Although the fitting function, F, was developed for normally distributed data, its extended version appears to be applicable to a wider range of data types if the described characteristics hold. This means that there is the possibility that <bold>&#x003A3;</bold> accounts for systematic variation of <bold>S</bold> computed from asymmetric non-normal data because of the adjustment by <inline-formula><mml:math id="M31"><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>&#x003B1;</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula>. When there are also remainders due to random influences, the outcome of computing F may follow a &#x003C7;<sup>2</sup> distribution.</p>
<p>It should be noted that the F-based &#x003C7;<sup>2</sup> is currently only of importance as a component for the calculation of some fit indices. Furthermore, model fit may also be evaluated by fit indices that do not include &#x003C7;<sup>2</sup> as a primary ingredient (e.g., SRMR).</p></sec>
</sec>
<sec id="s2">
<title>An evaluation using simulated data</title>
<p>The major aim of the evaluation was to demonstrate that the described measurement model in combination with MLE will improve performance in investigating the latent structure of data in the presence of asymmetric non-normality. This premise gave rise to two hypotheses: first, we hypothesized that when analyzing data illustrating asymmetric non-normality MLE for ASN would lead to better overall model fit compared to regular MLE. This hypothesis is based on the expectation that only MLE for ASN, but not regular MLE, could account for all systematic variation in distorted data. Second, we hypothesized that MLE for ASN would not provide an advantage over regular MLE in accounting for systematic variation unrelated to asymmetric non-normality. This hypothesis aimed at systematic variation that was neither reached by an incorrect model nor related to distributional deviation. MLE for ASN was not expected to account for this kind of systematic variation. Another aim was the comparison of MLE for ASN with two established estimation methods suggested to account to non-normality: robust DWLS and WLS.</p>
<p>To evaluate our claims, we conducted a simulation study. Structured random data with and without skewness manipulation were investigated by a measurement model corresponding to the model for data generation, that is a model assuming equal contributions of a common source to all items. For this purpose we generated normally distributed random data showing an underlying structure and manipulated their distribution for obtaining asymmetric non-normal data. In most investigations the structure of data was unidimensional. In additional investigations with incorrect (and corresponding correct) models the underlying structure was two-dimensional.</p>
<sec>
<title>Method</title>
<p>Data generation started with three sets of 500 &#x000D7; 20 matrices of continuous and normally distributed data [<italic>N</italic>(0,1)] using a generation method described by [<xref ref-type="bibr" rid="B24">24</xref>]. Either one or two unrelated underlying dimensions characterized the matrices. The expected factor loading of the first set was 0.40, in the second set 0.35 and in the third set 0.30. We refer to these as data with <italic>expected first-level factor loadings, expected second-level factor loadings</italic> and <italic>expected third-level factor loadings</italic> in corresponding order. The expected residuals were set to the difference between 1 and the common variation of items (the squares of the corresponding expected factor loadings). Data matrices with two underlying dimensions were realized as the combination of two subsets of columns. In this case, columns 1&#x02013;10 of a matrix were associated with the first dimension and columns 11&#x02013;20 with the second one.</p>
<p>When analyzing real data outside of the current research project, the typical observation was that the items&#x00027; deviation from a symmetric distribution varied. We attempted to simulate this characteristic of real data by restricting deviation from symmetry to 10 of the 20 columns and generated different degrees of deviation. This transformation from symmetric to asymmetric was conducted according to the following formula: <italic>x</italic><sub>asn</sub> = (<italic>x</italic> &#x000D7; <italic>k</italic>)<sup>1/2</sup>/<italic>k</italic>, where <italic>x</italic> &#x0003E; 0. That is, that the data of a column were multiplied by integer <italic>k</italic> in the first step. Then the square root was computed and the result divided by <italic>k</italic>. The integer, <italic>k</italic>, was set to 4 in transforming two columns of a matrix, to 16 in another two, to 64 in further two, and also to 256 and 1,024 in two plus two more columns such that the amount of non-normality varied across columns.</p>
<p>The measurement model for the statistical investigation included one latent variable with fixed factor loadings (see Equation 13) whereas the variance parameter of the corresponding model-implied covariance matrix was set free for estimation, as also were the parameters included in the main diagonal of <bold>&#x00398;</bold> (see Equations 8, 9). The factor loadings were fixed according to Equation. 12.</p>
<p>Parameter estimation was conducted by means of the MLE option in LISREL [<xref ref-type="bibr" rid="B25">25</xref>] (regular MLE and MLE for ASN). In addition to MLE, DWLS and WLS were employed for parameter estimation. The LISREL version of DWLS additionally performed robust estimation according to a Satorra-Bentler <italic>post-hoc</italic> correction method.</p>
</sec>
<sec>
<title>Results</title>
<p>In the following the results of investigating model fit are reported for the different data conditions (data with no ASN and data with ASN) using regular MLE and MLE for ASN (first section), for correct and incorrect models (second section), and for different estimation methods (MLE for ASN, DWLS, and WLS) (third section).</p>
</sec>
<sec>
<title>Effects of MLE versions in normal and non-normal data (hypothesis 1)</title>
<p>Fit results for normally distributed and asymmetric non-normal data investigated by regular MLE and MLE for ASN are included in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<p>The three sections of this table comprise fit results observed when investigating data with an expected factor loading of 0.40 (expected first-level factor loadings), of 0.35 (expected second-level factor loadings) and of 0.30 (expected third-level factor loadings).</p>
<p>The first row of the upper section includes the mean &#x003C7;<sup>2</sup>, RMSEA, SRMR, NNFI, CFI, and AIC for the one-factor model with regular MLE applied to normally distributed data and the second row corresponding standard deviations (SD). Using commonly accepted guidelines (e.g., RMSEA &#x02264; 0.6, SRMR &#x02264; 0.8, NNFI &#x02265; 0.95, and CFI &#x02265; 0.95), all model fit indices indicated good model fit [<xref ref-type="bibr" rid="B26">26</xref>]. The third row provides the mean results observed in investigating asymmetric non-normal data also using regular MLE and the fourth row corresponding SDs. All statistics showed a numeric impairment in comparison to values presented in the first row; however, only NNFI and CFI values were down to the cutoff for good model fit. The mean results obtained by MLE for ASN applied to asymmetric non-normal data are included in the fifth row and the corresponding SDs in the sixth rows. All fit indices obtained by MLE for ASN indicated good model fit (i.e., were lower respectively larger than the corresponding cutoffs).</p>
<p>The fit results reported in the second and third sections (expected second-level and third-level factor loadings) of <xref ref-type="table" rid="T1">Table 1</xref> showed similar patterns as in the first one although numerically slightly worse degrees of model fit were signified. In these sections, asymmetric non-normal data led to the switch from good to acceptable model fit according to NNFI and CFI when there was no MLE for ASN. In contrast, MLE for ASN yielded good model fit.</p>
<p>In sum, investigations by regular MLE revealed the expected decrease in model fit in the presence of asymmetric non-normality while adjustment to asymmetric non-normality resulted in the return to the original degree of model fit (confirmation of hypothesis 1).</p>
</sec>
<sec>
<title>Effects of MLE versions in unrelated systematic variation (hypothesis 2)</title>
<p>Fit results for correct and incorrect models used in investigating normally distributed and asymmetric non-normal data with regular MLE and MLE for ASN are reported in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Mean fit results and SDs (in italics) observed in investigating data showing (no) asymmetric non-normality (ASN) by one-factor and two-factor models using regular MLE and MLE for ASN (<italic>N</italic> = 500).</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496;color:#ffffff">
<th valign="top" align="left"><bold>Characteristics of model</bold></th>
<th valign="top" align="center" colspan="2">&#x003C7;<sup><bold>2</bold></sup></th>
<th valign="top" align="center"><bold>RMSEA</bold></th>
<th valign="top" align="center"><bold>SRMR</bold></th>
<th valign="top" align="center"><bold>NNFI</bold></th>
<th valign="top" align="center"><bold>CFI</bold></th>
<th valign="top" align="center"><bold>AIC</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Two factors, MLE</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">266.2</td>
<td valign="top" align="center">0.029</td>
<td valign="top" align="center">0.055</td>
<td valign="top" align="center">0.86</td>
<td valign="top" align="center">0.86</td>
<td valign="top" align="center">312.2</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>26.6</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.003</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>26.6</italic></td>
</tr> <tr>
<td valign="top" align="left">Two factors, MLE for ASN</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">236.8</td>
<td valign="top" align="center">0.022</td>
<td valign="top" align="center">0.051</td>
<td valign="top" align="center">0.91</td>
<td valign="top" align="center">0.91</td>
<td valign="top" align="center">282.8</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>25.2</italic></td>
<td valign="top" align="center"><italic>0.007</italic></td>
<td valign="top" align="center"><italic>0.003</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>0.04</italic></td>
<td valign="top" align="center"><italic>25.2</italic></td>
</tr> <tr>
<td valign="top" align="left">One factor, MLE</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">534.3</td>
<td valign="top" align="center">0.06</td>
<td valign="top" align="center">0.07</td>
<td valign="top" align="center">0.59</td>
<td valign="top" align="center">0.59</td>
<td valign="top" align="center">576.3</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>71</italic></td>
<td valign="top" align="center"><italic>0.006</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.05</italic></td>
<td valign="top" align="center"><italic>0.05</italic></td>
<td valign="top" align="center"><italic>71</italic></td>
</tr> <tr>
<td valign="top" align="left">One factor, MLE for ASN</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">524.8</td>
<td valign="top" align="center">0.059</td>
<td valign="top" align="center">0.069</td>
<td valign="top" align="center">0.62</td>
<td valign="top" align="center">0.62</td>
<td valign="top" align="center">566.8</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>72.5</italic></td>
<td valign="top" align="center"><italic>0.006</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.05</italic></td>
<td valign="top" align="center"><italic>0.05</italic></td>
<td valign="top" align="center"><italic>72.5</italic></td>
</tr></tbody>
</table>
</table-wrap>
<p>The first to fourth rows of this table include fit results for the correctly specified model and the remaining rows for the misspecified (i.e., incorrectly specified) model. The fit results for the correct model, but with no adaptation to asymmetric non-normality (first row), signified good model fit according to RMSEA and SRMR whereas NNFI and CFI illustrated model misfit. The report for the correct model with adaptation to asymmetric non-normality (third row) revealed improvement regarding NNFI and CFI to the level of acceptable model fit. All fit results for the incorrect models (see fifth to eights rows) were numerically worse than the fit results for the correct models (see first to fourth rows). The impairment in model fit was very obvious in NNFI and CFI results. These values were far below of what was considered as acceptable. Furthermore, in misspecified models, there was virtually no fit improvement due to adaptation to asymmetric non-normality, as was suggested by the second hypothesis.</p>
<p>In sum, MLE for ASN did improve model fit when the model was correct. In contrast, in incorrect models there was virtually no improvement (results in line with hypothesis 2).</p>
</sec>
<sec>
<title>Comparison of estimation methods</title>
<p>Finally, when analyzing asymmetric non-normal data, model fit was compared under MLE for ASN and alternative estimation methods, DWLS and WLS. In order to establish comparability, the alternative estimation methods were applied in combination with the one-factor model showing factor loadings fixed to the value of one, i.e., factor loadings reflecting the assumption of equal contributions of the latent source to all manifest variables.</p>
<p><xref ref-type="table" rid="T3">Table 3</xref> provides the mean fit results.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Mean fit results and SDs (in italics) observed for MLE for ASN, DWLS and WLS in investigating data showing asymmetric non-normality (ASN) (<italic>N</italic> = 500).</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496;color:#ffffff">
<th valign="top" align="left" colspan="2"><bold>Estimation method</bold></th>
<th valign="top" align="center"><bold>NT</bold></th>
<th valign="top" align="center"><bold>SB</bold></th>
<th valign="top" align="center"><bold>RMSEA</bold></th>
<th valign="top" align="center"><bold>SRMR</bold></th>
<th valign="top" align="center"><bold>NNFI</bold></th>
<th valign="top" align="center"><bold>CFI</bold></th>
<th valign="top" align="center"><bold>AIC</bold></th>
</tr>
<tr>
<th valign="top" align="center" colspan="2"></th>
<th valign="top" align="center"><bold>&#x003C7;<sup>2</sup></bold></th>
<th valign="top" align="center"><bold>&#x003C7;<sup>2</sup></bold></th>
<th/>
<th/>
<th/>
<th/>
<th/>
</tr>
</thead>
<tbody>
<tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="9"><bold>Expected first-level factor loadings (0.40)</bold></td>
</tr> <tr>
<td valign="top" align="left">MLE for ASN</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">229.4</td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">0.019</td>
<td valign="top" align="center">0.054</td>
<td valign="top" align="center">0.98</td>
<td valign="top" align="center">0.98</td>
<td valign="top" align="center">271.4</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>26.1</italic></td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center"><italic>0.008</italic></td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>26.1</italic></td>
</tr> <tr>
<td valign="top" align="left">DWLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">300.7</td>
<td valign="top" align="center">288.5</td>
<td valign="top" align="center">0.032</td>
<td valign="top" align="center">0.066</td>
<td valign="top" align="center">0.95</td>
<td valign="top" align="center">0.95</td>
<td valign="top" align="center">330.5</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>30.2</italic></td>
<td valign="top" align="center"><italic>28.8</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>0.01</italic></td>
<td valign="top" align="center"><italic>28.8</italic></td>
</tr> <tr>
<td valign="top" align="left">WLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">428.7<xref ref-type="table-fn" rid="TN1"><sup>a</sup></xref></td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">0.050</td>
<td valign="top" align="center">0.126</td>
<td valign="top" align="center">0.38</td>
<td valign="top" align="center">0.38</td>
<td valign="top" align="center">470.7</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>44.5</italic></td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.015</italic></td>
<td valign="top" align="center"><italic>0.08</italic></td>
<td valign="top" align="center"><italic>0.08</italic></td>
<td valign="top" align="center"><italic>44.5</italic></td>
</tr> <tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="9"><bold>Expected second-level factor loadings (0.35)</bold></td>
</tr> <tr>
<td valign="top" align="left">MLE for ASN</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">216.0</td>
<td/>
<td valign="top" align="center">0.015</td>
<td valign="top" align="center">0.049</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">258.0</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>24.0</italic></td>
<td/>
<td valign="top" align="center"><italic>0.008</italic></td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>24.0</italic></td>
</tr> <tr>
<td valign="top" align="left">DWLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">262.6</td>
<td valign="top" align="center">255.7</td>
<td valign="top" align="center">0.026</td>
<td valign="top" align="center">0.057</td>
<td valign="top" align="center">0.94</td>
<td valign="top" align="center">0.94</td>
<td valign="top" align="center">297.7</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>26.4</italic></td>
<td valign="top" align="center"><italic>26.0</italic></td>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.004</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>0.02</italic></td>
<td valign="top" align="center"><italic>26.0</italic></td>
</tr> <tr>
<td valign="top" align="left">WLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">400.0<xref ref-type="table-fn" rid="TN1"><sup>a</sup></xref></td>
<td/>
<td valign="top" align="center">0.047</td>
<td valign="top" align="center">0.101</td>
<td valign="top" align="center">0.40</td>
<td valign="top" align="center">0.40</td>
<td valign="top" align="center">442.0</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>45.1</italic></td>
<td/>
<td valign="top" align="center">0.005</td>
<td valign="top" align="center">0.012</td>
<td valign="top" align="center">0.09</td>
<td valign="top" align="center">0.09</td>
<td valign="top" align="center">45.1</td>
</tr> <tr style="background-color:#dee1e1;">
<td valign="top" align="left" colspan="9"><bold>Expected third-level factor loadings (0.30)</bold></td>
</tr> <tr>
<td valign="top" align="left">MLE for ASN</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">205.4</td>
<td/>
<td valign="top" align="center">0.011</td>
<td valign="top" align="center">0.046</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">0.97</td>
<td valign="top" align="center">247.4</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center">22.9</td>
<td/>
<td valign="top" align="center">0.008</td>
<td valign="top" align="center">0.003</td>
<td valign="top" align="center">0.04</td>
<td valign="top" align="center">0.03</td>
<td valign="top" align="center">22.9</td>
</tr> <tr>
<td valign="top" align="left">DWLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">235.5</td>
<td valign="top" align="center">231.8</td>
<td valign="top" align="center">0.020</td>
<td valign="top" align="center">0.051</td>
<td valign="top" align="center">0.93</td>
<td valign="top" align="center">0.93</td>
<td valign="top" align="center">273.5</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>23.1</italic></td>
<td valign="top" align="center"><italic>23.0</italic></td>
<td valign="top" align="center"><italic>0.006</italic></td>
<td valign="top" align="center"><italic>0.003</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>0.03</italic></td>
<td valign="top" align="center"><italic>20.0</italic></td>
</tr> <tr>
<td valign="top" align="left">WLS</td>
<td valign="top" align="left">Mean</td>
<td valign="top" align="center">373.7<xref ref-type="table-fn" rid="TN1"><sup>a</sup></xref></td>
<td/>
<td valign="top" align="center">0.044</td>
<td valign="top" align="center">0.086</td>
<td valign="top" align="center">0.40</td>
<td valign="top" align="center">0.40</td>
<td valign="top" align="center">415.7</td>
</tr>
 <tr>
<td valign="top" align="left"><italic>SD</italic></td>
<td valign="top" align="center"><italic>39.3</italic></td>
<td/>
<td valign="top" align="center"><italic>0.005</italic></td>
<td valign="top" align="center"><italic>0.009</italic></td>
<td valign="top" align="center"><italic>0.09</italic></td>
<td valign="top" align="center"><italic>0.09</italic></td>
<td valign="top" align="center"><italic>39.3</italic></td>
</tr></tbody>
</table>
<table-wrap-foot>
<fn id="TN1"><label>a</label><p>LISREL reports minimum fit function &#x003C7;<sup>2</sup> for WLS.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>Results presented in <xref ref-type="table" rid="T3">Table 3</xref> follow the same structure as <xref ref-type="table" rid="T1">Table 1</xref>. There are three parts according to the expected factor loadings considered in data generation: expected factor loading of 0.40 (expected first-level factor loadings), of 0.35 (expected second-level factor loadings) and of 0.30 (expected third-level factor loadings). In each part the first and second rows repeat fit results of <xref ref-type="table" rid="T1">Table 1</xref> for MLE for ASN, followed by DWLS and WLS results in corresponding order.</p>
<p>In each part, the fit results observed in investigating the data by MLE for ASN signified the numerically best model fit, DWLS the second best fit and the one obtained by WLS, ranked third. Furthermore, we investigated the difference between the statistics using the reported SDs. In data characterized by expected first-level factor loadings the difference between MLE for ASN and DWLS was larger than two SDs for &#x003C7;<sup>2</sup>, NNFI, CFI and AIC, and it was exactly two SDs for SRMR in data characterized by expected second-order factor loadings. WLS differed from MLE for ASN and from DWLS by more than two SDs in all comparison. Moreover, comparisons by CFI difference test [<xref ref-type="bibr" rid="B27">27</xref>] revealed substantial differences between all estimation methods for all data types.</p></sec>
</sec>
<sec sec-type="discussion" id="s3">
<title>Discussion</title>
<p>The maximum likelihood estimation method for factor analysis is closely linked to the concept of a multivariate normal distribution. The density function of a multivariate normal distribution provided the outset for creating the maximum likelihood fitting function [<xref ref-type="bibr" rid="B1">1</xref>], and unobservable data as the core component of parameter estimation by EM algorithm are assumed to follow a normal distribution [<xref ref-type="bibr" rid="B13">13</xref>]. Furthermore, normality of the remainders of fitting <bold>&#x003A3;</bold> to <bold>S</bold> is necessary to justify the assumption that the output of the maximum likelihood fitting function follows a &#x003C7;<sup>2</sup> distribution [<xref ref-type="bibr" rid="B19">19</xref>]. Therefore, it is no surprise that analyzing non-normal data by regular MLE may not yield expected fit results.</p>
<p>Although the primary output of MLE, the &#x003C7;<sup>2</sup> statistic, is currently rarely used in evaluating model fit, many alternative fit indices that are preferred instead [<xref ref-type="bibr" rid="B26">26</xref>] include the &#x003C7;<sup>2</sup> statistic as ingredient in their calculations. The fit indices with cutoffs used in evaluating model fit show different degrees of sensitivity for deviation from a normal distribution in the reported study. NNFI and CFI showed considerable sensitivity for deviation whereas RMSEA and SRMR indicated good model fit under all conditions. The replacement of MLE by the other estimation methods revealed that SRMR was also to some degree sensitive to deviation from normality. Although RMSEA showed numeric impairment due to deviation from normality, there was no substantial difference according to the corresponding cutoff [<xref ref-type="bibr" rid="B27">27</xref>]. The mean RMSEA always signified good model fit, even in incorrect models.</p>
<p>The modification of MLE by using the transformational measurement model for the investigation of asymmetric non-normal data led to the expected improvement in model fit over regular MLE. When analyzing data in a variety of data types, the CFI difference [<xref ref-type="bibr" rid="B27">27</xref>] signified fit improvement. A substantial improvement was also observed by the RMSEA difference [<xref ref-type="bibr" rid="B27">27</xref>] in data with expected first-level factor loadings. Both MLE versions discriminated virtually equally well between correct and incorrect models. But there was no improvement due to MLE for ASN if systematic variation that was not otherwise reached was unrelated to distributional deviation.</p>
<p>To arrive at general statements regarding the performances of the estimation methods, we base the comparisons among these methods on the means reported in <xref ref-type="table" rid="T1">Tables 1</xref>&#x02013;<xref ref-type="table" rid="T3">3</xref> for one-factor models regarding the cutoffs for good model-data fit (RMSEA &#x02264; 0.6, SRMR &#x02264; 0.8, NNFI &#x02265; 0.95, and CFI &#x02265; 0.95). The percentages of reaching these cutoffs were 100 for MLE ASN, 66.6 for regular MLE, 66.6 for DWLS and 25 for WLS. The similar performances for regular MLE and DWLS are overall in line with results reported for these methods observed in investigating data showing less strong deviation from normality [<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>]. The better performance of MLE for ASN is presumably restricted to data showing considerable deviation from normality. It is likely to do less well in overall normally distributed data.</p>
<p>An important but restricting assumption of MLE for ASN is the assumption of equal contributions of the latent source to all responses that justifies setting the factor loadings to equal sizes. The assumption of one latent variable as the source of responding is characteristic of customary confirmatory factor analysis [<xref ref-type="bibr" rid="B20">20</xref>] and the version of EM algorithm for confirmatory factor analysis [<xref ref-type="bibr" rid="B13">13</xref>]. But, in applied research, equal contributions of the latent source are rarely assumed so that factor loadings are mostly free for estimation. In investigating simulated data with expected factor loadings of the same size for all manifest variables, fixed factor loadings do not mean a disadvantage in comparison to free factor loadings [<xref ref-type="bibr" rid="B28">28</xref>]. Under appropriate conditions it may also apply for empirical data. Nevertheless, a version allowing for free factor loadings is desirable that, however, would require a more basic change of EM algorithm.</p>
<p>One limitation of the present study is the restriction to data distributions derived from a normal distribution by distortion. We concentrated on such asymmetric non-normal data because of our experiences with empirical research. The reason was that we observed that problems in reaching good model fit were frequently associated with data showing asymmetric non-normality. Another limitation is that distributions related to the normal distribution like beta, binomial and gamma distributions [<xref ref-type="bibr" rid="B29">29</xref>] are not taken into consideration. A further limitation is that we concentrated on model fit and omitted investigating bias in estimates. Future studies may include different simulation conditions to examine the viability of MLE for ASN under additional distribution and model conditions while including a variety of outcomes (e.g., parameter bias).</p>
<p>In summary, we present an alternative version of the maximum likelihood estimation (MLE) method for confirmatory factor analysis which includes a transformational measurement model. This transformational measurement model enables the investigation of the latent structure of asymmetric non-normal data. Furthermore, a simulation study is reported that demonstrates the efficiency of the method. An advantage in investigating asymmetric non-normal data over otherwise efficient robust DWLS and WLS is made apparent. We look forward to future research studies addressing these strengths and limitations of MLE for ASN.</p></sec>
<sec sec-type="conclusions" id="s4">
<title>Conclusion</title>
<p>The transformational measurement model opens up maximum likelihood estimation to data that deviate from a normal distribution that is otherwise a precondition for maximum likelihood estimation in factor analysis.</p></sec>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p></sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>KS conceptualized the study, performed the computations, and contributed to the writing. CD and BF shared ideas and contributed substantially to the writing. All authors contributed to the article and approved the submitted version.</p></sec>
</body>
<back>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s7">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>J&#x000F6;reskog</surname> <given-names>KG</given-names></name></person-group>. <article-title>A general approach to confirmatory maximum likelihood factor analysis</article-title>. <source>Psychometrika.</source> (<year>1969</year>) <volume>34</volume>:<fpage>183</fpage>&#x02013;<lpage>202</lpage>. <pub-id pub-id-type="doi">10.1007/BF02289343</pub-id></citation></ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jobst</surname> <given-names>LJ</given-names></name> <name><surname>Auerswald</surname> <given-names>M</given-names></name> <name><surname>Moshagen</surname> <given-names>M</given-names></name></person-group>. <article-title>The effect of latent and error non-normality on measures of fit in structural equation modeling</article-title>. <source>Educ Psychol Meas</source>. (<year>2021</year>) <volume>82</volume>:<fpage>911</fpage>&#x02013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1177/00131644211046201</pub-id><pub-id pub-id-type="pmid">35989731</pub-id></citation></ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lai</surname> <given-names>K</given-names></name></person-group>. <article-title>Estimating standardized SEM parameters given non-normal data and incorrect model: methods and comparisons</article-title>. <source>Struct Equation Model.</source> (<year>2018</year>) <volume>25</volume>:<fpage>1</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2017.1392248</pub-id></citation></ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>West</surname> <given-names>SG</given-names></name> <name><surname>Finch</surname> <given-names>JF</given-names></name> <name><surname>Curran</surname> <given-names>PJ</given-names></name></person-group>. <article-title>Structural equation models with non-normal variables: problems and remedies</article-title>. In: <person-group person-group-type="editor"><name><surname>Hoyle</surname> <given-names>R</given-names></name></person-group>, editor. <source>Structural Equation Modeling: Concepts, Issues, and Applications.</source> <publisher-loc>Thousand Oaks, CA</publisher-loc>: <publisher-name>Sage</publisher-name> (<year>1995</year>). p. <fpage>56</fpage>&#x02013;<lpage>75</lpage>.</citation></ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>C</given-names></name> <name><surname>Muth&#x000E9;n</surname> <given-names>B</given-names></name></person-group>. <article-title>Evaluation of model fit indices for latent variable models with categorical and continuous outcomes</article-title>. In: <source>Paper presented at the Annual Meeting of the American Educational Research Association</source> (<publisher-loc>New Orleans</publisher-loc>) (<year>2002</year>).<pub-id pub-id-type="pmid">29664908</pub-id></citation></ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deng</surname> <given-names>L</given-names></name> <name><surname>Yang</surname> <given-names>M</given-names></name> <name><surname>Marcoulides</surname> <given-names>KM</given-names></name></person-group>. <article-title>Structural equation modeling with many variables: a systematic review of issues and developments</article-title>. <source>Front Psychol.</source> (<year>2018</year>) <volume>9</volume>:<fpage>1</fpage>&#x02013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2018.00580</pub-id><pub-id pub-id-type="pmid">29755388</pub-id></citation></ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="book"><person-group person-group-type="author"><collab>Satorra A and Bentler PM</collab></person-group>. <article-title>Corrections to test statistics and standard errors in covariance structure analysis</article-title>. In: <person-group person-group-type="editor"><name><surname>von Eye</surname> <given-names>A</given-names></name> <name><surname>Clogg</surname> <given-names>CC</given-names></name></person-group>, editors. <source>Latent Variables Analysis: Applications for Developmental Research.</source> <publisher-loc>Thousand Oaks</publisher-loc>: <publisher-name>Sage</publisher-name> (<year>1994</year>). p. <fpage>399</fpage>&#x02013;<lpage>419</lpage>.</citation></ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Arruda</surname> <given-names>EH</given-names></name> <name><surname>Bentler</surname> <given-names>PM</given-names></name> <name><surname>A</surname></name></person-group>. <article-title>Regularized GLS for structural equation modeling</article-title>. <source>Struct Equation Model.</source> (<year>2017</year>) <volume>24</volume>:<fpage>657</fpage>&#x02013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2017.1318392</pub-id></citation></ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Browne</surname> <given-names>MW</given-names></name></person-group>. <article-title>Asymptotically distribution-free methods for the analysis of covariance structures</article-title>. <source>Br J Math Stat Psychol.</source> (<year>1984</year>) <volume>37</volume>:<fpage>62</fpage>&#x02013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1111/j.2044-8317.1984.tb00789.x</pub-id><pub-id pub-id-type="pmid">6733054</pub-id></citation></ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>CH</given-names></name></person-group>. <article-title>Confirmatory factor analysis with ordinal data: comparing robust maximum likelihood and diagonally weighted least squares</article-title>. <source>Behav Res Methods.</source> (<year>2016</year>) <volume>48</volume>:<fpage>936</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-015-0619-7</pub-id><pub-id pub-id-type="pmid">26174714</pub-id></citation></ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>CH</given-names></name></person-group>. <article-title>Statistical estimation of structural equation models with a mixture of continuous and categorical observed variables</article-title>. <source>Behav Res Methods.</source> (<year>2021</year>) <volume>53</volume>:<fpage>2191</fpage>&#x02013;<lpage>213</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-021-01547-z</pub-id><pub-id pub-id-type="pmid">33791955</pub-id></citation></ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dempster</surname> <given-names>AP</given-names></name> <name><surname>Laird</surname> <given-names>NM</given-names></name> <name><surname>Rubin</surname> <given-names>DB</given-names></name></person-group>. <article-title>Maximum Likelihood from Incomplete Data via the EM algorithm</article-title>. <source>J Roy Stat Soc B.</source> (<year>1977</year>) <volume>39</volume>:<fpage>1</fpage>&#x02013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1111/j.2517-6161.1977.tb01600.x</pub-id></citation></ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rubin</surname> <given-names>DB</given-names></name> <name><surname>Thayer</surname> <given-names>DT</given-names></name></person-group>. <article-title>EM algorithms for ML factor analysis</article-title>. <source>Psychometrika</source>. (<year>1982</year>) <volume>47</volume>:<fpage>69</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1007/BF02293851</pub-id></citation></ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Micceri</surname> <given-names>T</given-names></name></person-group>. <article-title>The unicorn, the normal curve, and other improbable creatures</article-title>. <source>Psychol Bull.</source> (<year>1989</year>) <volume>105</volume>:<fpage>156</fpage>&#x02013;<lpage>66</lpage>. <pub-id pub-id-type="doi">10.1037/0033-2909.105.1.156</pub-id></citation></ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Myung</surname> <given-names>IJ</given-names></name></person-group>. <article-title>Tutorial on maximum likelihood estimation</article-title>. <source>J Math Psychol.</source> (<year>2003</year>) <volume>47</volume>:<fpage>90</fpage>&#x02013;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1016/S0022-2496(02)00028-7</pub-id></citation></ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lalitha</surname> <given-names>S</given-names></name> <name><surname>Mishra</surname> <given-names>A</given-names></name></person-group>. <article-title>Modified maximum likelihood estimation for Rayleigh distribution</article-title>. <source>Commun Stat A Theor.</source> (<year>2011</year>) <volume>25</volume>:<fpage>389</fpage>&#x02013;<lpage>401</lpage>. <pub-id pub-id-type="doi">10.1080/03610929608831702</pub-id><pub-id pub-id-type="pmid">33267165</pub-id></citation></ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gelman</surname> <given-names>A</given-names></name> <name><surname>Hill</surname> <given-names>J</given-names></name></person-group>. (<year>2007</year>). <source>Data Analysis Using Regression and Multilevel/Hierarchical Models.</source> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Cambridge University Press</publisher-name> (<year>2007</year>).<pub-id pub-id-type="pmid">17608921</pub-id></citation></ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hecht</surname> <given-names>M</given-names></name> <name><surname>Gische</surname> <given-names>C</given-names></name> <name><surname>Vogel</surname> <given-names>D</given-names></name> <name><surname>Zitzmann</surname> <given-names>S</given-names></name></person-group>. <article-title>Integrating out nuisance parameters for computationally more efficient Bayesian estimation&#x02014;an illustration and tutorial</article-title>. <source>Struct Equation Model.</source> (<year>2020</year>) <volume>27</volume>:<fpage>483</fpage>&#x02013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2019.1647432</pub-id></citation></ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>J&#x000F6;reskog</surname> <given-names>KG</given-names></name></person-group>. <article-title>A general method for analysis of covariance structure</article-title>. <source>Biometrika.</source> (<year>1970</year>) <volume>57</volume>:<fpage>239</fpage>&#x02013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.2307/2334833</pub-id></citation></ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Graham</surname> <given-names>JM</given-names></name></person-group>. <article-title>Congeneric and (essentially) tau-equivalent estimates of score reliability</article-title>. <source>Educ Psychol Meas.</source> (<year>2006</year>) <volume>66</volume>:<fpage>930</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1177/0013164406288165</pub-id></citation></ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schweizer</surname> <given-names>K</given-names></name> <name><surname>Troche</surname> <given-names>S</given-names></name> <name><surname>DiStefano</surname> <given-names>C</given-names></name></person-group>. <article-title>Scaling the variance of a latent variable while assuring constancy of the model</article-title>. <source>Front Psychol.</source> (<year>2019</year>) <volume>10</volume>:<fpage>887</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2019.00887</pub-id><pub-id pub-id-type="pmid">31068871</pub-id></citation></ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>McLachlan</surname> <given-names>GJ</given-names></name></person-group>. <source>Computation: Expectation-Maximization Algorithm</source>. <source>International Encyclopedia of the Social and Behavioral Sciences (Second Edition).</source> <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Elsevier.</publisher-name> (<year>2016</year>).</citation></ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gumedze</surname> <given-names>FN</given-names></name> <name><surname>Dunne</surname> <given-names>TT</given-names></name></person-group>. <article-title>Parameter estimation and inference in the linear mixed model</article-title>. <source>Linear Algebra Appl.</source> (<year>2011</year>) <volume>435</volume>:<fpage>1920</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1016/j.Iaa.2011.04.015</pub-id><pub-id pub-id-type="pmid">11746323</pub-id></citation></ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>J&#x000F6;reskog</surname> <given-names>KG</given-names></name> <name><surname>S&#x000F6;rbom</surname> <given-names>D</given-names></name></person-group>. <source>Interactive LISREL: User&#x00027;s Guide</source>. <publisher-loc>Lincolnwood, IL</publisher-loc>: <publisher-name>Scientific Software International Inc</publisher-name>. (<year>2001</year>).</citation></ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>J&#x000F6;reskog</surname> <given-names>KG</given-names></name> <name><surname>S&#x000F6;rbom</surname> <given-names>D</given-names></name></person-group>. <source>LISREL 8.80.</source> <publisher-loc>Lincolnwood, IL</publisher-loc>: <publisher-name>Scientific Software International Inc</publisher-name> (<year>2006</year>).</citation></ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>DiStefano</surname> <given-names>C</given-names></name></person-group>. <article-title>Examining fit with structural equation models</article-title>. In: <person-group person-group-type="editor"><name><surname>Schweizer</surname> <given-names>K</given-names></name> <name><surname>Distefano</surname> <given-names>C</given-names></name></person-group>, editors. <source>Principles and Methods of Test Construction</source>. <publisher-loc>G&#x000F6;ttingen</publisher-loc>: <publisher-name>Hogrefe Publishing</publisher-name> (<year>2016</year>). p. <fpage>166</fpage>&#x02013;<lpage>193</lpage>.</citation></ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cheung</surname> <given-names>GW</given-names></name> <name><surname>Rensvold</surname> <given-names>RB</given-names></name></person-group>. <article-title>Evaluating goodness-of-fit indexes for testing measurement invariance</article-title>. <source>Struct Equation Model.</source> (<year>2002</year>) <volume>9</volume>:<fpage>233</fpage>&#x02013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1207/S15328007SEM0902_5</pub-id></citation></ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schweizer</surname> <given-names>K</given-names></name> <name><surname>Ren</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>T</given-names></name> <name><surname>Zeller</surname> <given-names>F</given-names></name></person-group>. <article-title>Does the constraint of factor loadings impair model fit and accuracy in parameter estimation?</article-title> <source>Int J Stat Prob.</source> (<year>2015</year>) <volume>4</volume>:<fpage>40</fpage>&#x02013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.5539/ijsp.v4n4p40</pub-id></citation></ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Leemis</surname> <given-names>LM</given-names></name> <name><surname>McQueston</surname> <given-names>JT</given-names></name></person-group>. <article-title>Univariate distribution relationships</article-title>. <source>Am Stat.</source> (<year>2008</year>) <volume>62</volume>:<fpage>45</fpage>&#x02013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.1198/000313008X270448</pub-id></citation></ref>
</ref-list> 
</back>
</article>