<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychol.</journal-id>
<journal-title>Frontiers in Psychology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychol.</abbrev-journal-title>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyg.2021.666074</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Home Textile Pattern Emotion Labeling Using Deep Multi-View Feature Learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Yang</surname> <given-names>Juan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1003841/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Zhang</surname> <given-names>Yuanpeng</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/903164/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>School of Textile and Clothing, Nantong University</institution>, <addr-line>Nantong</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Medical Informatics, Nantong University</institution>, <addr-line>Nantong</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Yaoru Sun, Tongji University, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Chenxi Huang, Xiamen University, China; Yuexin Li, Hubei University, China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Yuanpeng Zhang <email>maxbirdzhang&#x00040;ntu.edu.cn</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Emotion Science, a section of the journal Frontiers in Psychology</p></fn></author-notes>
<pub-date pub-type="epub">
<day>19</day>
<month>04</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>12</volume>
<elocation-id>666074</elocation-id>
<history>
<date date-type="received">
<day>09</day>
<month>02</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>01</day>
<month>03</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Yang and Zhang.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Yang and Zhang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license></permissions>
<abstract><p>Different home textile patterns have different emotional expressions. Emotion evaluation of home textile patterns can effectively improve the retrieval performance of home textile patterns based on semantics. It can not only help designers make full use of existing designs and stimulate creative inspiration but also help users select designs and products that are more in line with their needs. In this study, we develop a three-stage framework for home textile pattern emotion labeling based on artificial intelligence. To be specific, first of all, three kinds of aesthetic features, i.e., shape, texture, and salient region, are extracted from the original home textile patterns. Then, a CNN (convolutional neural network)-based deep feature extractor is constructed to extract deep features from the aesthetic features acquired in the previous stage. Finally, a novel multi-view classifier is designed to label home textile patterns that can automatically learn the weight of each view. The three-stage framework is evaluated by our data and the experimental results show its promising performance in home textile patterns labeling.</p></abstract>
<kwd-group>
<kwd>home textile pattern</kwd>
<kwd>emotion labeling</kwd>
<kwd>deep learning</kwd>
<kwd>multi-view learning</kwd>
<kwd>feature selection</kwd>
</kwd-group>
<counts>
<fig-count count="4"/>
<table-count count="6"/>
<equation-count count="7"/>
<ref-count count="24"/>
<page-count count="7"/>
<word-count count="4091"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>Emotion is the spiritual essence of home textile design. Fabric pattern is an important component of home textiles, which contains rich emotional information, including aesthetics and values. Therefore, fabric patterns rich in connotation and emotion are more and more respected by designers, which can meet the multiple needs of consumers. However, the pattern materials in home textile design and production are increasing day by day, and there are tens of thousands of patterns in the sample database. It is difficult for designers to make full use of the existing rich fabric patterns for home textile design. Therefore, how to integrate the objective characteristics (color, texture, pattern, etc.) and perceptual experience of fabric patterns into a mathematical model for aesthetic evaluation, emotional classification and retrieval, and emotion labeling of fabric patterns is one of the important topics for computer vision and textile design researchers.</p>
<p>With the continuous development of computer science, AI (artificial intelligence) and CV (computer vision) provide ideas and methods to solve this problem. Gan et al. (<xref ref-type="bibr" rid="B4">2014</xref>) made use of deep self-taught learning to obtain hierarchical representations, learn the concept of facial beauty, and produce human-like predictor. Datta and Wang (<xref ref-type="bibr" rid="B2">2010</xref>) established the first image aesthetics scoring website ACQUINE in 2010. Although the accuracy of the evaluation results is not high, it has shown that calculable aesthetics is feasible. Li and Chen (<xref ref-type="bibr" rid="B10">2009</xref>) adopted the features of color and composition in artistic aesthetic features to realize the classification of high and low aesthetic sense of paintings, and achieved a classification accuracy of more than 70%. Lo et al. (<xref ref-type="bibr" rid="B11">2012</xref>) studied image aesthetic classification from the aesthetic perspective of image color, layout, edge, and other features, and the results showed that image aesthetic features could be used for image sentiment analysis. With the development of deep learning, many deep learning based methods are also used for image aesthetic classification. Lu et al. (<xref ref-type="bibr" rid="B12">2015</xref>) considered both the local and global perspectives of images, designed the CNN (convolutional neural network) model for feature learning and classifier training, and evaluated the aesthetic quality of 250,000 images from AVA database. Compared with the traditional manual feature methods, this method has a great improvement in classification performance. Dong et al. (<xref ref-type="bibr" rid="B3">2015</xref>) used the CNN model trained by ImageNet large physical classification database to extract image features and classify high and low aesthetic values. The effectiveness of this method was proved in two image quality evaluation data sets. The above CNN-based aesthetic classification methods all use the pixel values of sample images in the large database as the input of CNN, without integrating the existing mature manual features. It is often very difficult to obtain a large number of home textile design patterns. Using relatively small and limited image aesthetic data sets for training will easily to lead to problems of overfitting and unstable convergence.</p>
<p>In this study, to achieve home textile emotion labeling, we propose a multi-view learning framework that contains three main components. The first component is used to extract initial multi-view features from the shape, texture, and salient region perspectives. The second component is used to extract deep features from the initial multi-view features by CNN. The last component is used to collaboratively learn from multi-view deep features.</p></sec>
<sec id="s2">
<title>Data And Methods</title>
<sec>
<title>Data</title>
<p>We employed 20 students from the School of Textile and Clothing of Nantong University to collect 5000 home textile patterns from the Internet. All images were re-sized into 256<sup>&#x0002A;</sup>256 for further use. We also invited another 10 students to conduct subjective aesthetic evaluation on these collected patterns. Students calibrated all the images from three aspects of high aesthetic feeling, low aesthetic feeling, and uncertainty. When the emotion evaluation labels with eight or more participants were the same, this label was determined as the final label of the pattern. After removing the uncertain labels, we finally obtained labeled 4,480 patterns. Sample images and the number of sample images of high and low aesthetic are shown in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>An example of high and low aesthetic home textile patterns.</p></caption>
<graphic xlink:href="fpsyg-12-666074-g0001.tif"/>
</fig>
<p>We used image clipping and horizontal flipping to enhance the patterns, and intercepted 4 corners and the middle image for mirror processing. After data enhancement, there are 25,000 patterns in each class (high aesthetic patterns and low aesthetic patterns).</p></sec>
<sec>
<title>Ethics</title>
<p>The studies involving human participants were reviewed and approved by ethics committee of Nantong University.</p>
<p>Written informed consent to participate in this study was provided by the participants.</p></sec>
<sec>
<title>Methods</title>
<p>In this study, we construct a three-stage method for home textile emotion labeling. The first stage is to extract initial shape features, texture features, and salient region features from home textile patterns. The second stage is a CNN-based feature extractor that is used to extract deep features from different aesthetic views from original home textile patterns. With deep features, in the second stage, we design a multi-view classifier to realize emotion labeling. The three-stage framework of emotion labeling is shown in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Three-stage framework of emotion labeling.</p></caption>
<graphic xlink:href="fpsyg-12-666074-g0002.tif"/>
</fig>
<sec>
<title>Acquisition of Initial Multi-View Features</title>
<p>As shown in <xref ref-type="fig" rid="F2">Figure 2</xref>, the initial aesthetic features of home textile contain shape features, texture features, and salient region features. Shape is one of the main characters of design pattern, which can be described by the edge. The edge of an image is a collection of points where the gray value is discontinuous or the gray value changes dramatically. In this study, we use the Sobel operator (Gao et al., <xref ref-type="bibr" rid="B5">2010</xref>) to detect the edge of an image. Texture is a very important content in fabric pattern, which contains many aesthetic features that affect the sense of beauty. The frequency and direction of Gabor filter (Mehrotra et al., <xref ref-type="bibr" rid="B13">1992</xref>) are similar to that of human visual system, which is suitable for image texture feature description. Therefore, in this study, Gabor features are used to represent texture features. The saliency region of an image is the region that attracts the most visual attention and has a more significant influence on the aesthetic feeling of the image. The saliency value of a pixel is defined by the contrast between the pixel and other pixel in the image. Pixels of the same color have the same salience. In this study, we use the LC (Luminance Contrast) (Zhai and Shah, <xref ref-type="bibr" rid="B23">2006</xref>) algorithm to extract salience region features. <xref ref-type="fig" rid="F3">Figure 3</xref> illustrates an example of initial multi-view features.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>An example of original multi-view features. <bold>(A)</bold> Original pattern, <bold>(B)</bold> Feature map of shape, <bold>(C)</bold> Feature map of texture, and <bold>(D)</bold> Feature map of salient region.</p></caption>
<graphic xlink:href="fpsyg-12-666074-g0003.tif"/>
</fig></sec>
<sec>
<title>Deep Multi-View Feature Learning</title>
<p>In this study, CNN (Sainath et al., <xref ref-type="bibr" rid="B20">2013</xref>; Kalchbrenner et al., <xref ref-type="bibr" rid="B8">2014</xref>; O&#x00027;Shea and Nash, <xref ref-type="bibr" rid="B16">2015</xref>; Albawi et al., <xref ref-type="bibr" rid="B1">2017</xref>) is employed to extract deep features from home textile patterns from different views. The initial features in shape, texture, and salience region are first constructed following the methods discussed in section &#x02018;Acquisition of Initial Multi-View Features&#x02019;. The network structure of our deep feature extractor is shown in <xref ref-type="fig" rid="F4">Figure 4</xref>.</p>
<list list-type="order">
<list-item><p>Input layer: in this layer, we reduce each original three-channel image of 256<sup>&#x0002A;</sup>256<sup>&#x0002A;</sup>3 size to 10 images of 224<sup>&#x0002A;</sup>224<sup>&#x0002A;</sup>3 size by random cropping.</p></list-item>
<list-item><p>The first convolution layer (Con-1): in this layer, the input images are reduced into 55<sup>&#x0002A;</sup>55 feature maps by 48 convolutional kernels (the kernel size is 11<sup>&#x0002A;</sup>11 and the step size is 4). Because the response result of ReLU (Rectified Linear Units) is unbounded (it can be very large), normalization is required. Here, LRN (Local Response Normalization) (Robinson et al., <xref ref-type="bibr" rid="B19">2007</xref>) is used to perform local response normalization.</p></list-item>
<list-item><p>The second convolution layer (Con-2): in this layer, we use 128 convolutional kernels (the kernel size is 5<sup>&#x0002A;</sup>5 and the step size is 2) to further extract features from the 48 feature maps (the kernel size is 27<sup>&#x0002A;</sup>27) generated by the last layer.</p></list-item>
<list-item><p>The third convolution layer (Con-3): in this layer, we use 192 convolutional kernels (the kernel size is 3<sup>&#x0002A;</sup>3 and the step size is 1) to generate 192 feature maps (13<sup>&#x0002A;</sup>13).</p></list-item>
</list>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>CNN-based deep feature extractor.</p></caption>
<graphic xlink:href="fpsyg-12-666074-g0004.tif"/>
</fig>
<p>Our feature extractor uses the final output to calculate the approximation error and performs back propagation to update the network parameters during training. When the training process is done, the output of the penultimate layer is selected as the deep features learned by our feature extractor. Therefore, the deep features we extracted not only have lower dimensionality than the original features but also possess better discrimination ability to enhance the generalizability of the subsequent classification model.</p></sec>
<sec>
<title>Multi-View Learning</title>
<p>In this section, we will develop a multi-view classifier (Jiang et al., <xref ref-type="bibr" rid="B6">2016</xref>; Qian et al., <xref ref-type="bibr" rid="B17">2018</xref>) for emotion labeling based on the deep multi-view features that are extracted by the CNN-based deep feature extractor shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. The basic idea is that the Shannon Entropy is introduced to the Ridge Regression (SERR) model to automatically identify the weight of each view. With view weights, the deep shape features, deep texture features, and deep salient region features are combined to achieve collaborate learning. Suppose <italic>X</italic><sub><italic>k</italic></sub> represents the training feature set of the <italic>k</italic>th view, <italic>Y</italic> represents the emotion label of the training set, then the objective function of the new classifier SERR is formulated as follows:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>J</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>Y</mml:mi><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B4;</mml:mi><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;ln&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>s</mml:mi><mml:mo>.</mml:mo><mml:mi>t</mml:mi><mml:mo>.</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003C9;<sub><italic>k</italic></sub> represents the weight of the <italic>k</italic>th view, <italic>W</italic><sub><italic>k</italic></sub> represents the corresponding transformation matrix, and &#x003B3; and &#x003B4; are two parameters to control the penalization terms. The objective function in Equation (1) can be solved by introducing Lagrangian multipliers. To be specific, by introducing the Lagrangian multipliers &#x003C4;, the corresponding Lagrangian function can be formulated as</p>
<disp-formula id="E3"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>Y</mml:mi><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B4;</mml:mi><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;ln&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>By setting <inline-formula><mml:math id="M5"><mml:mfrac><mml:mrow><mml:mo>&#x02202;</mml:mo><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02202;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M6"><mml:mfrac><mml:mrow><mml:mo>&#x02202;</mml:mo><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02202;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>, we have</p>
<disp-formula id="E5"><label>(3)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mi>I</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:mi>Y</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(4)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>exp</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>Y</mml:mi><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>/</mml:mo><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mtext>exp</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mo stretchy="false">&#x02016;</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>Y</mml:mi><mml:msup><mml:mrow><mml:mo stretchy="false">&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>/</mml:mo><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>With <italic>W</italic><sub><italic>k</italic></sub> and &#x003C9;<sub><italic>k</italic></sub>, we can use the alternate iteration to search for the optimal solution. The detailed steps of home textile emotion labeling using deep multi-view feature learning are shown as in Algorithm 1.</p>
<table-wrap position="float">
<label>Algorithm 1</label>
<caption><p>Model training</p></caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td valign="top" align="left"><bold>Input</bold>: original image set <italic>D</italic>, emotion label <italic>Y</italic>, parameters &#x003B3; and &#x003B4;</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Output</bold>: transformation matrix of each view <italic>W</italic><sub><italic>k</italic></sub>, weight of each view &#x003C9;<sub><italic>k</italic></sub></td>
</tr>
<tr>
<td valign="top" align="left">Procedure:</td>
</tr>
<tr>
<td valign="top" align="left">1. Use Sobel operator, Gabor filter, and LC to extract initial multi-view features.</td>
</tr>
<tr>
<td valign="top" align="left">2. Use deep feature extractor to extract deep multi-view features, <italic>X</italic><sub><italic>k</italic></sub>.</td>
</tr>
<tr>
<td valign="top" align="left">3. Randomly assign &#x003C9;<sub><italic>k</italic></sub> under <inline-formula><mml:math id="M9"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>.</td>
</tr>
<tr>
<td valign="top" align="left">4. Set <italic>t</italic> = 0.</td>
</tr>
<tr>
<td valign="top" align="left">5. Repeat.</td>
</tr>
<tr>
<td valign="top" align="left">6. Use Equation (2) to update <italic>W</italic><sub><italic>k</italic></sub> based on current &#x003C9;<sub><italic>k</italic></sub>.</td>
</tr>
<tr>
<td valign="top" align="left">7. Use Equation (3) to update &#x003C9;<sub><italic>k</italic></sub> based on current <italic>W</italic><sub><italic>k</italic></sub>.</td>
</tr>
<tr>
<td valign="top" align="left">8. <italic>t</italic> = <italic>t</italic> &#x0002B; 1.</td>
</tr>
<tr>
<td valign="top" align="left">9. Until (||<italic>W</italic><sub><italic>k</italic></sub>(<italic>t</italic> &#x0002B; 1) &#x02212; <italic>W</italic><sub><italic>k</italic></sub>(<italic>t</italic>)|| &#x0003C; &#x003B5;).</td>
</tr>
</tbody>
</table>
</table-wrap>
 <p>After the training procedure is done, for an unseen textile pattern, we use Algorithm 2 to perform emotion labeling.</p>
<table-wrap position="float">
<label>Algorithm 2</label>
<caption><p>Model testing</p></caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td valign="top" align="left"><bold>Input</bold>: unseen textile patterns, transformation matrix of each view <italic>W</italic><sub><italic>k</italic></sub>, weight of each view &#x003C9;<sub><italic>k</italic></sub></td>
</tr>
<tr>
<td valign="top" align="left"><bold>Output</bold>: emotion labels of unseen textile patterns</td>
</tr>
<tr>
<td valign="top" align="left">Procedure:</td>
</tr>
<tr>
<td valign="top" align="left">1. Use the feature index obtained from training to select features from the unseen textile patterns.</td>
</tr>
<tr>
<td valign="top" align="left">2. Emotion of unseen textile patterns can be determined by
<disp-formula id="E7"><mml:math id="M10"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02032;</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>g</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003C9;</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02032;</mml:mo></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></td>
</tr>
</tbody>
</table>
</table-wrap></sec></sec></sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<p>In this section, we will evaluate our new emotion labeling method from two perspectives, i.e., the effectiveness of the deep feature extractor and the effectiveness of the multi-view classifier.</p>
<sec>
<title>Settings</title>
<p>To evaluate the effectiveness of the deep feature extractor, we first reshape the original shape features, texture features, and the saliency region features from two-dimensional matrices to one-dimensional vectors and then introduce several traditional feature selection methods for discriminant feature selection. The settings of the introduced feature selection methods are shown in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Settings of feature selection methods.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Feature selection methods</bold></th>
<th valign="top" align="left"><bold>Settings</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MRMR (Toga&#x000E7;ar et al., <xref ref-type="bibr" rid="B21">2020</xref>)</td>
<td valign="top" align="left">Use default setting recommended by Toga&#x000E7;ar et al. (<xref ref-type="bibr" rid="B21">2020</xref>)</td>
</tr>
<tr>
<td valign="top" align="left"><italic>l</italic><sub>21</sub>-norm (Nie et al., <xref ref-type="bibr" rid="B15">2010</xref>)</td>
<td valign="top" align="left">The regularized parameter is searched from [0.001, 10]</td>
</tr>
<tr>
<td valign="top" align="left">PCA (Karamizadeh et al., <xref ref-type="bibr" rid="B9">2013</xref>)</td>
<td valign="top" align="left">Use default setting recommended by Karamizadeh et al. (<xref ref-type="bibr" rid="B9">2013</xref>)</td>
</tr>
<tr>
<td valign="top" align="left">Relief (Urbanowicz et al., <xref ref-type="bibr" rid="B22">2018</xref>)</td>
<td valign="top" align="left">Use default setting recommended by Urbanowicz et al. (<xref ref-type="bibr" rid="B22">2018</xref>)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Additionally, to evaluate the effectiveness of the multi-view classifier SERR we proposed, we directly concatenate all features from different views and use classic classifiers SVM (Support Vector Machine), KNN (K-Nearest Neighbor), NB (Naive Bayes), and DT (Decision Tree) for classification. The settings of the introduced feature selection methods are shown in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Settings of classifiers.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Classifiers</bold></th>
<th valign="top" align="left"><bold>Setting</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">SVM (Joachims, <xref ref-type="bibr" rid="B7">1998</xref>)</td>
<td valign="top" align="left">The Gaussian kernel is adopted. The kernel width is searched from [10<sup>&#x02212;5</sup>, 10<sup>5</sup>], the center is searched from [10<sup>&#x02212;5</sup>, 10<sup>5</sup>], and <italic>C</italic> is set to 100</td>
</tr>
<tr>
<td valign="top" align="left">KNN (Zhang et al., <xref ref-type="bibr" rid="B24">2017</xref>)</td>
<td valign="top" align="left"><italic>K</italic> is set to 5</td>
</tr>
<tr>
<td valign="top" align="left">NB (Rish, <xref ref-type="bibr" rid="B18">2001</xref>)</td>
<td valign="top" align="left">Use default setting recommended by Rish (<xref ref-type="bibr" rid="B18">2001</xref>)</td>
</tr>
<tr>
<td valign="top" align="left">DT (Myles et al., <xref ref-type="bibr" rid="B14">2004</xref>)</td>
<td valign="top" align="left">Use default setting recommended by Myles et al. (<xref ref-type="bibr" rid="B14">2004</xref>)</td>
</tr>
<tr>
<td valign="top" align="left">SERR</td>
<td valign="top" align="left">&#x003B3; and &#x003B4; are searched from [0.001, 10]</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>All experiments are conducted on a PC with Intel&#x000AE; Core&#x02122; i7-9700 &#x00040;3.00 GHz Dual, 32G memory, and RTX 2080 Ti. The coding platform is Matlab 2012b.</p></sec>
<sec>
<title>Experimental Results</title>
<p>In this section, we will report our experimental results from two aspects. First of all, our deep feature extractor is used to extract deep features from each initial view. For comparison studies, four commonly used feature selection methods MRMR, <italic>l</italic><sub>21</sub>-norm, PCA, and Relief are also introduced to select discriminant features from each initial view. The Ridge Regression model is taken as the classifier for the classification tasks in each view. <xref ref-type="table" rid="T3">Tables 3</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref> show the classification results in terms of Accuracy, Sensitivity, and Specificity on each view.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Classification performance in terms of accuracy.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Feature selection methods</bold></th>
<th valign="top" align="center"><bold>Shape features</bold></th>
<th valign="top" align="center"><bold>Texture features</bold></th>
<th valign="top" align="center"><bold>Saliency region features</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MRMR</td>
<td valign="top" align="center">0.9458 &#x000B1; 0.0030</td>
<td valign="top" align="center">0.7781 &#x000B1; 0.0028</td>
<td valign="top" align="center">0.7517 &#x000B1; 0.0057</td>
</tr>
<tr>
<td valign="top" align="left"><italic>l</italic><sub>21</sub>-norm</td>
<td valign="top" align="center">0.9784 &#x000B1; 0.0028</td>
<td valign="top" align="center">0.7558 &#x000B1; 0.0047</td>
<td valign="top" align="center">0.7510 &#x000B1; 0.0047</td>
</tr>
<tr>
<td valign="top" align="left">PCA</td>
<td valign="top" align="center">0.9555 &#x000B1; 0.0036</td>
<td valign="top" align="center">0.7697 &#x000B1; 0.0027</td>
<td valign="top" align="center">0.7478 &#x000B1; 0.0092</td>
</tr>
<tr>
<td valign="top" align="left">Relief</td>
<td valign="top" align="center">0.9420 &#x000B1; 0.0014</td>
<td valign="top" align="center">0.7784 &#x000B1; 0.0015</td>
<td valign="top" align="center">0.7149 &#x000B1; 0.0102</td>
</tr>
<tr>
<td valign="top" align="left">Deep feature extractor</td>
<td valign="top" align="center">0.9816 &#x000B1; 0.0021</td>
<td valign="top" align="center">0.7870 &#x000B1; 0.0099</td>
<td valign="top" align="center">0.7579 &#x000B1; 0.0111</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Classification performance in terms of sensitivity.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Feature selection methods</bold></th>
<th valign="top" align="center"><bold>Shape features</bold></th>
<th valign="top" align="center"><bold>Texture features</bold></th>
<th valign="top" align="center"><bold>Saliency region features</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MRMR</td>
<td valign="top" align="center">0.9478 &#x000B1; 0.0023</td>
<td valign="top" align="center">0.5541 &#x000B1; 0.0103</td>
<td valign="top" align="center">0.5147 &#x000B1; 0.0101</td>
</tr>
<tr>
<td valign="top" align="left"><italic>l</italic><sub>21</sub>-norm</td>
<td valign="top" align="center">0.9578 &#x000B1; 0.0030</td>
<td valign="top" align="center">0.5458 &#x000B1; 0.0088</td>
<td valign="top" align="center">0.5368 &#x000B1; 0.0117</td>
</tr>
<tr>
<td valign="top" align="left">PCA</td>
<td valign="top" align="center">0.9412 &#x000B1; 0.0026</td>
<td valign="top" align="center">0.5269 &#x000B1; 0.0201</td>
<td valign="top" align="center">0.5429 &#x000B1; 0.0098</td>
</tr>
<tr>
<td valign="top" align="left">Relief</td>
<td valign="top" align="center">0.9541 &#x000B1; 0.0026</td>
<td valign="top" align="center">0.5578 &#x000B1; 0.0152</td>
<td valign="top" align="center">0.5025 &#x000B1; 0.0100</td>
</tr>
<tr>
<td valign="top" align="left">Deep feature extractor</td>
<td valign="top" align="center">0.9612 &#x000B1; 0.0037</td>
<td valign="top" align="center">0.5670 &#x000B1; 0.0189</td>
<td valign="top" align="center">0.5435 &#x000B1; 0.0125</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Classification performance in terms of specificity.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Feature selection methods</bold></th>
<th valign="top" align="center"><bold>Shape features</bold></th>
<th valign="top" align="center"><bold>Texture features</bold></th>
<th valign="top" align="center"><bold>Saliency region features</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MRMR</td>
<td valign="top" align="center">0.9236 &#x000B1; 0.0036</td>
<td valign="top" align="center">0.8436 &#x000B1; 0.0056</td>
<td valign="top" align="center">0.8178 &#x000B1; 0.0142</td>
</tr>
<tr>
<td valign="top" align="left"><italic>l</italic><sub>21</sub>-norm</td>
<td valign="top" align="center">0.9547 &#x000B1; 0.0021</td>
<td valign="top" align="center">0.8268 &#x000B1; 0.0017</td>
<td valign="top" align="center">0.8362 &#x000B1; 0.0073</td>
</tr>
<tr>
<td valign="top" align="left">PCA</td>
<td valign="top" align="center">0.9632 &#x000B1; 0.0054</td>
<td valign="top" align="center">0.8817 &#x000B1; 0.0026</td>
<td valign="top" align="center">0.8555 &#x000B1; 0.0089</td>
</tr>
<tr>
<td valign="top" align="left">Relief</td>
<td valign="top" align="center">0.9785 &#x000B1; 0.0023</td>
<td valign="top" align="center">0.8557 &#x000B1; 0.0053</td>
<td valign="top" align="center">0.8521 &#x000B1; 0.0107</td>
</tr>
<tr>
<td valign="top" align="left">Deep feature extractor</td>
<td valign="top" align="center">0.9818 &#x000B1; 0.0025</td>
<td valign="top" align="center">0.8929 &#x000B1; 0.0081</td>
<td valign="top" align="center">0.8617 &#x000B1; 0.0133</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Additionally, to highlight our proposed multi-view learning method SERR, with the deep features from different views, we introduce SVM, KNN, DT, and NB as classifiers for comparison studies. <xref ref-type="table" rid="T6">Table 6</xref> shows the classification results in terms of Accuracy, Sensitivity, and Specificity on each view.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Classification performance of SERR, SVM, KNN, NB and DT.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Classifiers</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="center"><bold>Sensitivity</bold></th>
<th valign="top" align="center"><bold>Specificity</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">0.9369 &#x000B1; 0.0024</td>
<td valign="top" align="center">0.9257 &#x000B1; 0.0016</td>
<td valign="top" align="center">0.9478 &#x000B1; 0.0042</td>
</tr>
<tr>
<td valign="top" align="left">KNN</td>
<td valign="top" align="center">0.9478 &#x000B1; 0.0025</td>
<td valign="top" align="center">0.9327 &#x000B1; 0.0019</td>
<td valign="top" align="center">0.9087 &#x000B1; 0.0015</td>
</tr>
<tr>
<td valign="top" align="left">NB</td>
<td valign="top" align="center">0.9368 &#x000B1; 0.0074</td>
<td valign="top" align="center">0.9264 &#x000B1; 0.0015</td>
<td valign="top" align="center">0.9457 &#x000B1; 0.0014</td>
</tr>
<tr>
<td valign="top" align="left">DT</td>
<td valign="top" align="center">0.9698 &#x000B1; 0.0025</td>
<td valign="top" align="center">0.9524 &#x000B1; 0.0036</td>
<td valign="top" align="center">0.9644 &#x000B1; 0.0025</td>
</tr>
<tr>
<td valign="top" align="left">SERR</td>
<td valign="top" align="center">0.9865 &#x000B1; 0.0014</td>
<td valign="top" align="center">0.9782 &#x000B1; 0.0045</td>
<td valign="top" align="center">0.9654 &#x000B1; 0.0024</td>
</tr>
</tbody>
</table>
</table-wrap></sec></sec>
<sec id="s4">
<title>Discussion And Conclusion</title>
<p>With the improvement of automation of home textile production and design and the increasing number of stored home textile pattern images in enterprises, the traditional retrieval methods can no longer meet the needs of home textile manufacturers. It is necessary to conduct aesthetic evaluation and emotional analysis of home textile pattern, so as to provide better services to enterprises and consumers.</p>
<p>Currently, there are two main ways for home textile enterprises to search home textile design patterns. One is to manually classify and number home textile patterns, which is mainly used for enterprise management. However, a large amount of management storage will cause a waste of resources in all aspects. The other is to establish an image retrieval system for the pre-classified home textile patterns, but it needs to manually classify each pattern. However, it requires human classification of each pattern, which is time-consuming and labor-intensive, and not all home textile patterns can be expressed with keywords or symbols, so it is difficult to meet the different retrieval needs of different searchers.</p>
<p>With the development of AI, especially deep learning, in this study, to achieve home textile emotion labeling, we propose a multi-view learning framework that contains three main components. The first component is used to extract initial multi-view features from the shape, texture, and salient region perspectives. The second component is used to extract deep features from the initial multi-view features by CNN. The last component is used to collaboratively learn from multi-view deep features. We demonstrate our method from two perspectives. From the results shown in <xref ref-type="table" rid="T3">Tables 3</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref>, we see that the features extracted by our deep feature extractor drives the best classifier in each kind of features in terms of Accuracy, Sensitivity, and Specificity, respectively. This superiority indicates that deep features are more discriminant than the initial features obtained in the first stage. Additionally, from <xref ref-type="table" rid="T6">Table 6</xref>, we see that our proposed multi-view classifier SERR performs better than the traditional classifiers, SVM, KNN, DT, and NB, which means that collaborative learning in multiple feature space is more reliable than direct feature concatenation.</p>
<p>The experimental results show that the emotional labeling method proposed in this study realizes the emotional labeling of home textile patterns, provides an auxiliary retrieval method for consumers who want to buy home textile with certain emotional semantics, provides convenience for enterprises in the production and design of home textile patterns, and can meet the multiple needs of consumers. In our future work, we will consider more kinds of deep features and develop more deep feature extractors for emotion labeling.</p></sec>
<sec sec-type="data-availability-statement" id="s5">
<title>Data Availability Statement</title>
<p>The datasets presented in this article are not readily available because we will further enrich them. Requests to access the datasets should be directed to Yuanpeng Zhang, <email>maxbirdzhang&#x00040;ntu.edu.cn</email>.</p></sec>
<sec id="s6">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by Ethics Committee of Nantong University. Written informed consent to participate in this study was provided by the participants.</p></sec>
<sec id="s7">
<title>Author Contributions</title>
<p>JY and YZ designed the whole algorithm and experiments. All authors contributed to the article and approved the submitted version.</p></sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
</body>
<back>
<ack><p>We thank the reviewers whose comments and suggestions helped improve this manuscript.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Albawi</surname> <given-names>S.</given-names></name> <name><surname>Mohammed</surname> <given-names>T. A.</given-names></name> <name><surname>Al-Zawi</surname> <given-names>S.</given-names></name></person-group> (<year>2017</year>). <article-title>Understanding of a convolutional neural network</article-title>, in <source>2017 International Conference on Engineering and Technology (ICET)</source> (<publisher-loc>IEEE</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/ICEngTechnol.2017.8308186</pub-id><pub-id pub-id-type="pmid">33748163</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Datta</surname> <given-names>R.</given-names></name> <name><surname>Wang</surname> <given-names>J. Z.</given-names></name></person-group> (<year>2010</year>). <source>Mir 2010 - Proceedings of the 2010 ACM SIGMM International Conference on Multimedia Information</source>. <publisher-loc>Philadelphia, PA</publisher-loc>: <publisher-name>Association for Computing Machinery (ACM)</publisher-name>, <fpage>421</fpage>&#x02013;<lpage>424</lpage>.</citation></ref>
<ref id="B3">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dong</surname> <given-names>Z.</given-names></name> <name><surname>Shen</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>H.</given-names></name></person-group> (<year>2015</year>). <source>Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics).</source> <publisher-loc>Sydney, NSW</publisher-loc>: <publisher-name>Springer Verlag</publisher-name>, <fpage>524</fpage>&#x02013;<lpage>535</lpage>.</citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gan</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Zhai</surname> <given-names>Y.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name></person-group> (<year>2014</year>). <article-title>Deep self-taught learning for facial beauty prediction</article-title>. <source>Neurocomputing</source> <volume>144</volume>, <fpage>295</fpage>&#x02013;<lpage>303</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2014.05.028</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>W.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Yang</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>H.</given-names></name></person-group> (<year>2010</year>). <article-title>An improved Sobel edge detection</article-title>, in <source>2010 3rd International Conference on Computer Science and Information Technology</source>. <publisher-name>IEEE</publisher-name>, <fpage>67</fpage>&#x02013;<lpage>71</lpage>.</citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>Y.</given-names></name> <name><surname>Deng</surname> <given-names>Z.</given-names></name> <name><surname>Chung</surname> <given-names>F. L.</given-names></name> <name><surname>Wang</surname> <given-names>G.</given-names></name> <name><surname>Qian</surname> <given-names>P.</given-names></name> <name><surname>Choi</surname> <given-names>K. S.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Recognition of epileptic EEG signals using a novel multiview TSK fuzzy system</article-title>. <source>IEEE Trans. Fuzzy Syst</source>. <volume>25</volume>, <fpage>3</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1109/TFUZZ.2016.2637405</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Joachims</surname> <given-names>T.</given-names></name></person-group> (<year>1998</year>). <source>Making Large-Scale SVM Learning Practical</source>. Technical report, <publisher-name>Technische Universit&#x000E4;t, Dortmund, Germany</publisher-name>.</citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kalchbrenner</surname> <given-names>N.</given-names></name> <name><surname>Grefenstette</surname> <given-names>E.</given-names></name> <name><surname>Blunsom</surname> <given-names>P.</given-names></name></person-group> (<year>2014</year>). <article-title>A convolutional neural network for modelling sentences</article-title>. <source>arXiv[Preprint].arXiv:1511.08458</source>. <pub-id pub-id-type="doi">10.3115/v1/P14-1062</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karamizadeh</surname> <given-names>S.</given-names></name> <name><surname>Abdullah</surname> <given-names>S. M.</given-names></name> <name><surname>Manaf</surname> <given-names>A. A.</given-names></name> <name><surname>Zamani</surname> <given-names>M.</given-names></name> <name><surname>Hooman</surname> <given-names>A.</given-names></name></person-group> (<year>2013</year>). <article-title>An overview of principal component analysis</article-title>. <source>J. Sign. Inform. Process</source>. <volume>4</volume>:<fpage>173</fpage>. <pub-id pub-id-type="doi">10.4236/jsip.2013.43B031</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Chen</surname> <given-names>T.</given-names></name></person-group> (<year>2009</year>). <article-title>Aesthetic visual quality assessment of paintings</article-title>. <source>IEEE J. Select. Topics Sign. Process</source>. <volume>3</volume>, <fpage>236</fpage>&#x02013;<lpage>252</lpage>. <pub-id pub-id-type="doi">10.1109/JSTSP.2009.2015077</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lo</surname> <given-names>K. Y.</given-names></name> <name><surname>Liu</surname> <given-names>K.</given-names></name> <name><surname>Chen</surname> <given-names>C. S.</given-names></name></person-group> (<year>2012</year>). <source>Proceedings - International Conference on Pattern Recognition</source>. <publisher-loc>Tsukuba</publisher-loc>: <publisher-name>Institute of Electrical and Electronics Engineers, Inc</publisher-name>, <fpage>2186</fpage>&#x02013;<lpage>2189</lpage>.</citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>X.</given-names></name> <name><surname>Lin</surname> <given-names>Z. L.</given-names></name> <name><surname>Jin</surname> <given-names>H.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>J. Z.</given-names></name></person-group> (<year>2015</year>). <article-title>RAPID: rating pictorial aesthetics using deep learning</article-title>. <source>IEEE Trans. Multimedia</source> <volume>17</volume>, <fpage>2021</fpage>&#x02013;<lpage>2034</lpage>. <pub-id pub-id-type="doi">10.1109/TMM.2015.2477040</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mehrotra</surname> <given-names>R.</given-names></name> <name><surname>Namuduri</surname> <given-names>K. R.</given-names></name> <name><surname>Ranganathan</surname> <given-names>N.</given-names></name></person-group> (<year>1992</year>). <article-title>Gabor filter-based edge detection</article-title>. <source>Patt. Recogn.</source> <volume>25</volume>, <fpage>1479</fpage>&#x02013;<lpage>1494</lpage>. <pub-id pub-id-type="doi">10.1016/0031-3203(92)90121-X</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Myles</surname> <given-names>A. J.</given-names></name> <name><surname>Feudale</surname> <given-names>R. N.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Woody</surname> <given-names>N. A.</given-names></name> <name><surname>Brown</surname> <given-names>S. D.</given-names></name></person-group> (<year>2004</year>). <article-title>An introduction to decision tree modeling</article-title>. <source>J. Chemometr.</source> <volume>18</volume>, <fpage>275</fpage>&#x02013;<lpage>285</lpage>. <pub-id pub-id-type="doi">10.1002/cem.873</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nie</surname> <given-names>F.</given-names></name> <name><surname>Huang</surname> <given-names>H.</given-names></name> <name><surname>Cai</surname> <given-names>X.</given-names></name> <name><surname>Ding</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>Efficient and robust feature selection via joint &#x02113;2, 1-norms minimization</article-title>. <source>Adv. Neural Inform. Process. Syst</source>. <volume>23</volume>, <fpage>1813</fpage>&#x02013;<lpage>1821</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2013.146</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>O&#x00027;Shea</surname> <given-names>K.</given-names></name> <name><surname>Nash</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). <article-title>An introduction to convolutional neural networks</article-title>. <source>arXiv[Preprint].arXiv:1404.2188</source>.</citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qian</surname> <given-names>P.</given-names></name> <name><surname>Zhou</surname> <given-names>J.</given-names></name> <name><surname>Jiang</surname> <given-names>Y.</given-names></name> <name><surname>Liang</surname> <given-names>F.</given-names></name> <name><surname>Zhao</surname> <given-names>K.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Multi-view maximum entropy clustering by jointly leveraging inter-view collaborations and intra-view-weighted attributes</article-title>. <source>IEEE Access</source> <volume>6</volume>, <fpage>28594</fpage>&#x02013;<lpage>28610</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2018.2825352</pub-id><pub-id pub-id-type="pmid">31289704</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Rish</surname> <given-names>I.</given-names></name></person-group> (<year>2001</year>). <article-title>An empirical study of the naive Bayes classifier</article-title>, in <source>IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence</source> (<publisher-loc>Santa Barbara, CA</publisher-loc>). Vol. <volume>3</volume>, <fpage>41</fpage>&#x02013;<lpage>46</lpage>.</citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Robinson</surname> <given-names>A. E.</given-names></name> <name><surname>Hammon</surname> <given-names>P. S.</given-names></name> <name><surname>de Sa</surname> <given-names>V. R.</given-names></name></person-group> (<year>2007</year>). <article-title>Explaining brightness illusions using spatial filtering and local response normalization</article-title>. <source>Vis. Res</source>. <volume>47</volume>, <fpage>1631</fpage>&#x02013;<lpage>1644</lpage>. <pub-id pub-id-type="doi">10.1016/j.visres.2007.02.017</pub-id><pub-id pub-id-type="pmid">17459448</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sainath</surname> <given-names>T. N.</given-names></name> <name><surname>Mohamed</surname> <given-names>A.</given-names></name> <name><surname>Kingsbury</surname> <given-names>B.</given-names></name></person-group> (<year>2013</year>). <article-title>Deep convolutional neural networks for LVCSR.</article-title>, in <source>2013 IEEE International Conference on Acoustics, Speech and Signal Processing</source> (<publisher-loc>IEEE</publisher-loc>), <fpage>8614</fpage>&#x02013;<lpage>8618</lpage>. <pub-id pub-id-type="doi">10.1109/ICASSP.2013.6639347</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Toga&#x000E7;ar</surname> <given-names>M.</given-names></name> <name><surname>Ergen</surname> <given-names>B.</given-names></name> <name><surname>C&#x000F6;mert</surname> <given-names>Z.</given-names></name></person-group> (<year>2020</year>). <article-title>A deep feature learning model for pneumonia detection applying a combination of mRMR feature selection and machine learning models</article-title>. <source>Irbm</source> <volume>41</volume>, <fpage>212</fpage>&#x02013;<lpage>222</lpage>. <pub-id pub-id-type="doi">10.1016/j.irbm.2019.10.006</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Urbanowicz</surname> <given-names>R. J.</given-names></name> <name><surname>Meeker</surname> <given-names>M.</given-names></name> <name><surname>La Cava</surname> <given-names>W.</given-names></name> <name><surname>Olson</surname> <given-names>R. S.</given-names></name> <name><surname>Moore</surname> <given-names>J. H.</given-names></name></person-group> (<year>2018</year>). <article-title>Relief-based feature selection: introduction and review</article-title>. <source>J. Biomed. Inform</source>. <volume>85</volume>, <fpage>189</fpage>&#x02013;<lpage>203</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2018.07.014</pub-id><pub-id pub-id-type="pmid">30031057</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhai</surname> <given-names>Y.</given-names></name> <name><surname>Shah</surname> <given-names>M.</given-names></name></person-group> (<year>2006</year>). <article-title>Visual attention detection in video sequences using spatiotemporal cues</article-title>, in <source>Proceedings of the 14th ACM International Conference on Multimedia</source> (<publisher-loc>Santa Barbara, CA</publisher-loc>), <fpage>815</fpage>&#x02013;<lpage>824</lpage>. <pub-id pub-id-type="doi">10.1145/1180639.1180824</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Zong</surname> <given-names>M.</given-names></name> <name><surname>Zhu</surname> <given-names>X.</given-names></name> <name><surname>Cheng</surname> <given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>Learning k for knn classification</article-title>. <source>ACM Trans. Intell. Syst. Technol.</source> <volume>8</volume>, <fpage>1</fpage>&#x02013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1145/2990508</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This work was supported in part by the Philosophy and Social Science Foundation of Jiangsu Province under No. 18YSC009.</p>
</fn>
</fn-group>
</back>
</article>