Next Article in Journal
Frequency Fitting Optimization Using Evolutionary Algorithm in Cochlear Implant Users with Bimodal Binaural Hearing
Next Article in Special Issue
Intrinsic Shapes of Empathy: Functional Brain Network Topology Encodes Intersubjective Experience and Awareness Traits
Previous Article in Journal
Dynamic and Systemic Perspective in Autism Spectrum Disorders: A Change of Gaze in Research Opens to A New Landscape of Needs and Solutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Development of a Multi-Modality Emotion Recognition Test Presented via a Mobile Application

Institute of Behavioral Medicine, College of Medicine, National Cheng Kung University, Tainan 701, Taiwan
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(2), 251; https://doi.org/10.3390/brainsci12020251
Submission received: 25 December 2021 / Revised: 2 February 2022 / Accepted: 7 February 2022 / Published: 11 February 2022

Abstract

:
Emotion recognition ability is the basis of interpersonal communication and detection of brain alterations. Existing tools for assessing emotion recognition ability are mostly single modality, paper-and-pencil test format, and using only Western stimuli. However, various modalities and cultural factors greatly influence emotion recognition ability. We aimed to develop a multi-modality emotion recognition mobile application (MMER app). A total of 169 healthy adults were recruited as participants. The MMER app’s materials were extracted from a published database, and tablets were used as the interface. The Rasch, factor analysis, and related psychometric analyses were performed. The Cronbach alpha was 0.94, and the test–retest reliability was 0.85. Factor analyses identified three factors. In addition, an adjusted score formula was provided for clinical use. The MMER app has good psychometric properties, and its further possible applications and investigations are discussed.

1. Introduction

1.1. Emotion Recognition

Emotion recognition indicates the identification of emotional information in various ways, including through the face and prosody [1]. Different receiving channels affect emotion recognition accuracy. Emotion recognition ability is essential for human social interaction and considerably impacts everyday life [2]. When individuals interact with others, they need to receive and recognize emotional information from various sources, such as verbal (e.g., words and semantics) or non-verbal (e.g., tones, facial emotions, and body movements) stimulation. Studies have shown that emotion recognition ability is an essential factor in developing the ability to “understand others’ intentions” [3]. In the past few years, research on emotion recognition has grown tremendously and attracted the interest of scholars in various fields.
Emotion recognition ability develops during childhood [4] and changes with age. Nevertheless, the relationship between emotion recognition ability and aging is inconclusive. Previous studies have shown that emotion recognition ability declines with age [5]. The elderly find it more challenging to identify anger [6], fear, and sadness [7] than young people. Moreover, emotion recognition is essential for detecting brain alterations or early signs of psychological/neurological disorders. Evidence has shown that patients with neurodegenerative diseases (e.g., Alzheimer’s disease [8] and Parkinson’s disease [9]) or psychiatric disorders (e.g., autism spectrum disorder and schizophrenia [10]) are found to have impairments in facial emotion recognition ability. Furthermore, it has also been shown that alexithymia is a transdiagnostic construct [11], and detecting difficulties in emotion recognition could facilitate the diagnosis of psychiatric disorders. Changes in emotion recognition ability may reflect alterations in the corresponding brain areas. The literature indicates that some brain areas (e.g., the striatum, fusiform gyrus, superior temporal gyrus, amygdala, orbitofrontal cortex, basal ganglia, somatosensory, and insula) might be responsible for this ability [12,13,14].
Based on empirical evidence, it is believed that emotion recognition ability may be a good indicator of detecting subtle changes in the brain and facilitating the diagnosis and management of psychiatry-related diseases. Therefore, an assessment tool that measures emotion recognition ability is essential and has clinical utility.

1.2. Measurement of the Emotion Recognition

1.2.1. Single-Modal Emotion Recognition Tests

Most previous recognition tests have been based on Ekman’s facial emotion database [15,16]; however, most current emotion recognition tests only focus on four basic emotions (i.e., happy, sad, angry, and surprised/frightened) [17]. Unfortunately, only a few Eastern facial stimuli were included in this database. Although facial expression is believed to be a universal language for emotion [18], studies have found that ethnicity and cultural factors must be considered [19,20]. Culture may affect how individuals express their feelings; Westerners use the whole face to express emotions, while Easterners use the upper half of the face [20]. Jack and colleagues (2012) [21] found that Asians have difficulty recognizing European facial emotions. Current studies have revealed a culture-specific decoding strategy applied in various ethnicities and cultures [20].
Few single-modal emotion recognition tests (shown in Table 1), which were applied Eastern materials, have been developed in the past, including the Japanese and Caucasian Brief Affect Recognition Test (JACBART) [22] and the Chinese Facial Emotion Recognition Database (CFERD) [23]. The JACBART rates the degree in the seven types of facial emotions simultaneously (i.e., anger, contempt, disgust, fear, happiness, sadness, and surprise) for one facial stimulus. The CFERD is a three-dimensional color facial emotion recognition task, which involves matching each facial expression with the corresponding emotional word (happiness, disgust, fear, anger, sadness, surprise, or neutral). The tests mentioned above address the culture-specific issues but do not consider the interaction between the facial emotion source and other modalities. Therefore, developing a multi-modal emotion recognition test with Eastern materials is needed.

1.2.2. Multi-Modal Emotion Recognition Tests

A few emotion recognition tests have examined dual or multi-modal emotion recognition (shown in Table 2). The most widely known are the Florida Affect Battery (FAB) [24] and the Diagnostic Analysis Nonverbal Accuracy Scale (DANVA) [17]. These tests are well known and commonly used dual emotional recognition tests; however, they have limitations. For instance, the lack of core emotions (i.e., the FAB lacks disgust and fear, and the DANVA lacks disgust and surprise), the application of only Caucasian faces as stimuli, and the use of only the female face and voice. Fear and disgust are crucial emotions for people’s survival and social interactions. If one cannot recognize another’s fear, it may affect the individual’s understanding of a crisis and messages from others. It is believed that disgust is related to human beings’ specific needs (e.g., hunger and moral) [25], and disgust allows individuals to perform correct behaviors, such as nausea and vomiting after eating expired food; in addition, sexual violence can cause human disgust. Thus, the recognition of these two emotions is crucial for an individual’s life and interpersonal relationships. Although the FAB has been translated and validated in the Chinese population [26], it still uses Caucasian faces for facial stimuli. The FAB altered the voice to Chinese pronunciation; however, the Caucasian faces with Chinese voices may confuse the participants.
The Awareness of Social Inference Test Emotion Evaluation Test [27], the Multimodal Emotion Recognition Test (MERT) [28], the Geneva Emotion Recognition Test (GERT) and its short form (GERT-S) [29,30], and the Emotion Recognition Assessment in Multiple Modalities test [31] were also developed for examining multi-modal emotion recognition; however, they were not similar to the FAB with multiple subtests and were only performed for matching tests, involving different types of stimuli and emotional words. Moreover, most of these tests examine more than ten types of emotions, and too many emotion choices may interfere with recognition. The stimuli for the above multi-modal emotion recognition tests were all Caucasian faces, and they were not validated for the Eastern population.

1.2.3. The Unmet Need for, and Challenges of, the Development of the Emotion Recognition Test

The development of emotion recognition tools has progressed from the single-channel format [22,23,32,33] to the exploration of emotion recognition ability in multiple channels [17,24,26,27,28,29,30,31]. Besides, rather than focusing on basic emotions in the past, a more comprehensive range of emotions and covering various emotions is needed for recent development tools [28,29,30,31]. Another unmet need is that most tools use stimulus material from Western countries or ethnicities, and few tests have been developed for Eastern populations. Furthermore, traditional paper-and-pencil tests have practical limitations. Through using readily available electronic tools as a medium, it will be possible to enhance the accessibility and applicability of the tools. However, to the best of our knowledge, none of the tools for assessing emotion recognition ability in the literature reviewed above were displayed through user-friendly methods (e.g., mobile application).
Table 1. Single-modality emotion recognition tests with eastern facial stimulus.
Table 1. Single-modality emotion recognition tests with eastern facial stimulus.
Test NameCountryNumber of ItemsTest Methods or ProceduresReliability
Validity
Japanese and Caucasian Brief Affect Recognition Test (JACBART) [22]Japan56 aRate the degree of the 7 emotionsInternal reliability Test–retest Reliability
Concurrent validity
Convergent validity
The Chinese Facial Emotion Recognition Database (CFERD) [23]Taiwan100 a7 emotions classificationNA
NA [32]Malaysia56 a7 emotions classificationNA
Chinese Affective Picture System (CAPS) [33]China60 a4 emotions classificationNA
NA—not available. Type of the items: a—images.
Table 2. Multi-modality emotion recognition tests with Western facial stimulus.
Table 2. Multi-modality emotion recognition tests with Western facial stimulus.
Test NameCountryNumber of ItemsTest ContentTest Methods or ProceduresReliability
Validity
Florida Affect Battery [24] US232 a,b,cSimilar to
MMER
5 emotions classification Test–retest 0.89–0.97
The Awareness of Social Inference Test (TASIT)—part 1: Emotion Evaluation Test (EET) [27] Australia28 eSimilar to
subtest 3
7 emotions classification Parallel forms Reliability
Construct validity
Multimodal Emotion Recognition Test (MERT) [28] Switzerland120 a,b,d,eSimilar to
subtest 3
10 emotions classificationInterrater 0.38
Test–retest 0.78
EFA
Florida Affect Battery (Chinese version) [26] Taiwan225 a,b,c
(Western facial stimulus; Chinese prosody stimulus)
Similar to
MMER
5 emotions classification Content validity
Criterion validity
Norm comparison
Geneva Emotion Recognition Test (GERT) [29] Switzerland108 eSimilar to
subtest 3
14 emotions classification NA
Geneva Emotion Recognition Test—short form (GERT-S) [30] Switzerland42 eSimilar to
subtest 3
14 emotions classification Cronbach alpha = 0.80
CFA good fit
Emotion Recognition Assessment in Multiple Modalities Test (ERAM) [31] Sweden72 a,b,cSimilar to
subtest 3
12 emotions classification Cronbach alpha = 0.74
CFA good fit
NA—not available. The types of the items: a—images; b—audio; c—audio–image; d—video; e—audio–video.

1.3. Aim

We aimed to develop the Multi-Modality Emotion Recognition mobile application (MMER app), a multi-modal emotion recognition test. Chinese faces and prosody were used as stimuli. Moreover, comprehensive emotions (i.e., neutral, happy, sad, angry, disgust, fear, and surprise) were included. Furthermore, our test was designed as a device application and performed on a tablet due to the popularity and handiness of mobile devices. We used the tablet as an interface to calculate and output the scores, which improves the convenience and applicability of the test.

2. Materials and Methods

2.1. Participants

We recruited 169 healthy adults (demographic characteristics are presented in Table 3) from community activity centers and our college. Participants suspected to have dementia, based on a mini mental state examination (MMSE) score below 24, were excluded from the study [34]. In addition, patients with a history of psychiatric illness, substance abuse, severe systemic diseases, and traumatic brain injury were excluded. Before participation, the participants provided informed consent; ethical standards were drawn up based on the 1964 Declaration of Helsinki. The ethical research committee of National Cheng Kung University Hospital IRB (approval number: A-ER-107-425) confirmed the study protocols.

2.2. Procedure of the Development of the MMER App

2.2.1. Item Generation of the MMER App

We generated the items and format of the MMER app through a literature review and reference to other tests (e.g., the FAB). Facial and prosodic emotions were chosen as the materials for the MMER app. The materials were obtained from the Emotional Speech Database in Taiwan and from the Taiwan Corpora of Chinese Emotion and Relevant Psychophysiological Database [35,36,37]. Those datasets included many Eastern faces and tones. We randomly selected 2272 Chinese face pictures and 188 sound segment stimuli in 7 emotions to establish the MMER app. The pretest version of the MMER app included 9 subtests (5 facial tests, 2 prosodic tests, and 2 cross-modal tests), with a total of 325 items.

2.2.2. The Subtests of the MMER App

We referred to the format of the FAB to generate the subtests. Subtest 1 was a facial feature discrimination test. Subtests 2–5 were facial-related emotion recognition tests, subtests 6 and 7 were prosodic-related emotion recognition tests, and subtests 8 and 9 were facial–prosodic, cross-modal emotion recognition tests. We wrote the items into the app and used the tablet to collect the participants’ responses.
Subtest 1 was a “facial feature discrimination test”, with 24 items, including the front, side, and two-thirds of the face. Two faces were presented in the tablet at a time, and the participants were asked to determine whether the two faces were the same person. Subtest 2 was a “facial emotion discrimination test” using faces. The subtest had 35 items. Two faces were presented in the tablet at a time, and the participants were asked to determine whether the emotions of the two faces were the same. Subtest 3 was a “face-word matched test” and contained 42 items, including front, side, and two-thirds of the face. One face and seven emotion terms (i.e., neutral, happy, sad, angry, disgust, fear, and surprise) were displayed on the tablet, and the participants were asked to choose one emotion term that best fit the target face. Subtest 4 was a “word-face matched test” and contained 14 items. One emotion term and seven faces were shown on the tablet, and the participants were asked to choose one face that best fit the target emotion term. Subtest 5 was a “face-face emotion mated test” containing 28 items. Six faces were shown in the tablet (along with the target face), and the participants were asked to select one face whose emotion best resembled that of the target face.
Subtest 6 was a “prosodic emotion discrimination test”. Two emotional sentences were played via the tablet, and the participants were asked to determine whether the two sentences displayed the same emotions. Subtest 7 was a “prosodic-word matched test” with 28 items. One sentence and seven emotion terms were presented, and participants were asked to choose one emotion term that most suitably represented the target sentence’s emotion.
Subtest 8 was a “prosodic-face matched test” and contained 35 items. One sentence and four faces were presented on the tablet, and participants were asked to select one face whose emotion was the most suitable for the target sentence’s emotion. Subtest 9 was a “face-prosodic matched test” and contained 42 items. One face and four sentences were presented, and participants were asked to select one sentence whose emotion was the most suitable for the target face’s emotion.

2.2.3. The Pretest Stage

Seven participants were recruited to join the pretest procedure to examine the reaction time, test item accuracy, and modification of some items. In addition, we collected the participants’ comments after completing the test to improve the quality of the MMER app. After the pretest, we kept the items with an accuracy above 50%, increased the number of practice items (subtests 3–5 and subtest 7), and added feedback in the practice section. We modified the items in which accuracy was below 50% by excluding stimuli that presented conflicting emotions (e.g., actors acting out sad emotions, but most participants rated it as scared). The option of subtest 9 was originally a four-segment prosodic emotion; however, considering that prosodic emotion has only two opportunities to play and is more likely to be affected by the participant’s cognitive function (e.g., poor memory may make them forget the sound of the previous option), the number of options were reduced to three. In the analysis of the previous version of the MMER app, subtest 2 had unacceptable internal consistency (Cronbach’s alpha = 0.46) and test–retest reliability (0.28). Subtest 6 had poor internal consistency (Cronbach’s alpha = 0.56) and questionable test–retest reliability (0.67). Thus, we deleted these two subtests. The Rasch analysis results were used to modify the MMER app. Based on the Rasch analysis, the comparison between participant ability and item difficulty was conducted, and the items below participant ability were deleted.
Item difficulty, which was below the minimum value of participant performance, was eliminated. Nevertheless, it retained half of the items in each emotion if over half of the items in each emotion were canceled. As the purpose of the MMER app was to determine emotion recognition ability, seven types of emotion items were needed to fulfill its aims, even though it was easy for most of the population. The retained items were selected based on difficulty—the higher difficulty items were retained until each emotion item reached half.

2.2.4. The MMER App

The MMER app included 7 subtests and 198 items (100 items of facial tests, 25 items of prosodic tests, and 73 items of cross-modal tests), with a scoring method of 1 point per question (total score of 198 points).

2.3. Measurement

The MMER app was displayed on a 10.1-inch tablet to evaluate the participants’ emotion recognition ability (Figure 1). We used the MMSE [34] to exclude participants with dementia. MMSE has a total score of 30 and is often used to assess an individual’s general cognitive function. Those with scores below 24 were considered likely to have dementia. The Reading Mind in the Eyes Test (RMET) [38] was applied to establish criterion-related validity. The RMET is a test that is often used to assess individuals’ judgment of others’ feelings or emotions through the expressions around their eyes. This test presents a series of black-and-white photos expressing emotions and asks participants to choose the adjective that best fits the emotion expressed by the people in the photos.

2.4. Data Analysis

Descriptive statistics of performance were calculated for total scores, accuracy in each subtest, and demographic characteristics of all participants. Cronbach’s alpha was used to calculate internal consistency to examine the internal reliability of the MMER app. Twenty-eight participants were invited to complete the test again after three–four months from the first visit, and Pearson’s correlation was employed to confirm the test–retest reliability. The RMET [37] was applied to establish criterion-related validity. Confirmatory factor analysis was conducted to examine the factorial structure of the MMER app. A one-factor model (model 1) was first used, as the purpose of the MMER app was related to one factor—“emotion recognition”. Multi-dimensional models were conducted to examine the relationship between factors; models 2 and 4 were oblique, and models 3 and 5 were orthogonal. Moreover, two-factor models (models 2 and 3) and three-factor models were also tested to study the relationship between the three theoretical concepts (facial recognition, facial emotion recognition, and prosody emotion recognition). Oblique and orthogonal is a factor rotation for transform gained factors from factor analysis. It would maximize the large factor loadings and minimize the small factor loadings to enhance the interpretability for the factors. The major difference between oblique and orthogonal rotation is that the factors in the oblique rotation model could be correlated, and the correlation between factors in the orthogonal rotation model is equal to zero.

3. Results

3.1. Performance

The mean and standard deviation of the demographic characteristics and performance of the participants are presented in Table 3. The correct score and accuracy of the MMER app for the seven types of emotion are shown in Table 4.

3.2. Reliability, Confirmatory Factor Analyses, and Criterion-Related Validity

3.2.1. Reliability

The internal consistency of the MMER app was excellent (Cronbach’s alpha = 0.94) with good test–retest reliability (0.87).

3.2.2. Confirmatory Factor Analysis

The results for the different models are presented in Table 5, Figure 2, Figure 3 and Figure 4. The fit index showed that the orthogonal models fit more with the data. The results of Models 2 and 4 are similar. Moreover, the structure of Model 4 is equal to the theoretical structure of the MMER app. The factor loadings of Model 4 are shown in Table 6.

3.2.3. Criterion-Related Validity

The result of the Pearson’s correlation indicated that there was a strong positive association between RMET and the total score of the MMER app (r = 0.53, p < 0.001). Moreover, all subtests were also significantly correlated with RMET. For subtest 3 (“face-word matched test”), the Pearson correlation showed a strong positive association with RMET (r = 0.54, p < 0.001). In addition, the Pearson correlation showed that the RMET had a moderate positive association with subtest 4 (“word-face matched test”) (r = 0.43, p < 0.001), subtest 5 (“face-face emotion mated test”) (r = 0.47, p < 0.001), subtest 7 (“prosodic-word matched test”) (r = 0.40, p < 0.001), subtest 8 (“prosodic-face matched test”) (r = 0.43, p < 0.001), and subtest 9 (“face-prosodic matched test”) (r = 0.41, p < 0.001). For subtest 1 (“facial feature discrimination test”), the Pearson correlation showed a weak positive association with RMET (r = 0.23, p < 0.001).

3.3. Multiple Stepwise Regression for Application

According to the correlation analysis, the demographic variables of the sample were correlated with the MMER app; for instance, age (r = −0.77, p < 0.001) and education (r = 0.41, p < 0.001) were correlated to the total score. Multiple stepwise backward regression was conducted to modify the score of the MMER app using a formula. The demographic variables, including age, gender, and education, explained 65.67% of the MMER app performance based on multiple stepwise backward regression. The adjustment formula was as follows:
Adjusted MMER app score =
MMER app total score-4.007(gender-0.33)-0.750(age-48.28) + 1.645(education-13.73)
where gender was defined as male = 1 and female = 0.

4. Discussion

Emotion recognition is vital for social interaction and intimacy, and the evaluation of emotion recognition ability is necessary for detecting brain dysfunction and further developing rehabilitation programs. In the current study, we have overcome the limitations of previous studies and use the tablet’s advantages to develop a reliable and effective emotion recognition test—the MMER app. The MMER app contains 7 subtests with 198 items. The stimuli used in the MMER app are all Eastern stimuli (both sexes), and the MMER app can measure multi-modal emotion recognition ability (i.e., face and prosody). The MMER app has good psychometric properties and takes only 20 minutes to complete. To the best of our knowledge, the MMER app is the first suitable test in the Chinese population to measure emotion recognition ability via various modalities (i.e., visual and auditory).
The total accuracy of the MMER app was 69%, and subtest 1 had the highest accuracy (93%). Compared with other subtests, healthy people generally have the highest accuracy in subtest 1. The face-related subtest (76%) had the second-highest accuracy, followed by the face–prosody subtest (61%). The prosody-related subtest (44%) had the lowest accuracy. The accuracy was similar to previous study findings, which reported facial emotion recognition accuracy between 65 and 78% and prosodic emotion recognition between 52 and 65% [39,40]. The highest accuracy of each emotion was anger, followed by happiness; the lowest accuracy was fear, followed by disgust. These findings are consistent with previous research [35,37] and indicate that people have the highest consistency in evaluating happiness and anger and the lowest consistency in assessing fear and disgust. The highest accuracies of the face-related and prosodic-related subtests were for happiness and anger, respectively. In addition, the lowest accuracy of the face-related subtest was for fear. For the prosody-related subtest, the lowest accuracy was for disgust, and these findings are consistent with Scherer et al.’s results. [40].
The MMER app has high internal consistency and adequate test–retest reliability. Moreover, the RMET, as criterion-related validity, is closely related to the MMER app. Among the subtests, the face-related subtests and the RMET scores, which also use facial pictures as stimulus materials, are significantly correlated. Previous studies have shown that these two abilities are highly related [41], and lesion studies have also found that emotion recognition and mind-reading abilities partially share the exact neural mechanisms [13]. Thus, we believe that the MMER app has well-established, criterion-related validity. In addition, the factor analyses confirm that the MMER app has three factors and concepts: facial recognition and facial and prosody emotion recognition. The findings of our factor analyses show that the MMER app is a multi-modality tool for measuring emotion recognition ability. Moreover, the seven subtests may represent a different psychological mechanism of emotion recognition owing to the different presentation methods. Further investigations that apply this test in clinical practice—focusing on the relationship between different types of emotion recognition defects and related brain pathologies, which may assist in disease detection and rehabilitation—are encouraged.
The MMER app has a certain degree of discrimination on healthy participants, which is different from previous tests (e.g., FAB and DANVA) that have a ceiling effect. Studies suggest that demographic variables, such as age, gender [42], and education level, are crucial in emotion recognition ability [43]. Multiple stepwise regression can be adjusted due to demographic variables, potentially being used as the norm exploration in the future. An individual’s emotion recognition ability can be determined after using the adjusted score to query the percentage level comparison table. If one’s adjusted MMER app score is below the common standard (5th percentile), his/her emotion recognition ability is considered defective. This is the first multi-modality emotional recognition test for an app to the best of our knowledge. Our MMER app can improve the accessibility and clinical efficiency of assessment.
The limitation of this study lies in the lack of developed Eastern emotion recognition models to compare with and refer to, which is also the reason for this study. Although the MMER app is significantly related to RMET, and criterion-related validity was established, the RMET cannot perfectly play the role of the criterion, especially the part of voice–emotion subtests. Second, the cultural divergence among Eastern countries is considerable. Third, the intensity of emotional stimuli was not considered in this study. Emotional intensity may be one reason that affects the ability to recognize emotions and needs further investigation. Fourth, the accuracy of the prosody-related subtests was low. Previous studies have found that other factors easily affect prosody-related emotions (e.g., semantic meaning) [31]. Our findings also confirmed that prosody-related emotion recognition accuracy is lower than face-related emotion recognition. Although the accuracy of these subtests was poor, other psychometric properties of these subtests were acceptable–good. Finally, only healthy participants were recruited because there is evidence that psychiatric disorders may impair emotion recognition. The MMER app could potentially be created as a screening test or a standard emotion recognition test within Eastern cultures. Thus, other populations should be recruited in future studies to validate our findings. Modification of the MMER app was also required for a prospective study on sample variation, multicultural comparison in Eastern culture, and item selection.

5. Conclusions

The MMER app has well-established psychometric properties and provides an integrated Eastern version of an emotion recognition test, with multiple modalities involved in comprehensive emotions, and no sex bias in the stimulus. Moreover, we offered a formula to generate the adjusted score, which can be used to determine whether an individual’s emotion recognition ability is impaired through the percentile scale correspondence. Further research is needed to recruit other populations (e.g., clinical cases) to cross-validate the MMER app. This test also has the potential to be used in future clinical practice.

Author Contributions

All authors have read and agreed to the published version of the manuscript. Supervision, project administration, funding acquisition, and resources, R.-L.Y.; conceptualization, R.-L.Y. and P.-H.H.; methodology, R.-L.Y. and S.-F.P.; formal analysis, S.-F.P.; investigation and data curation, S.-F.P., H.-J.Y., C.-Y.C. and P.-H.H.; writing—original draft preparation: R.-L.Y. and S.-F.P.; writing—review and editing: R.-L.Y., S.-F.P., H.-J.Y. and C.-Y.C.; visualization: R.-L.Y., S.-F.P., H.-J.Y. and C.-Y.C.

Funding

This research was funded by the Higher Education Sprout Project, Ministry of Education to the Headquarters of University Advancement at National Cheng Kung University (NCKU) and the Ministry of Science and Technology (MOST), Taipei, Taiwan, grant number (MOST 110-2628-B-006-020 and 108-2410-H-006-046-).

Institutional Review Board Statement

The study was conducted following the Declaration of Helsinki, and approved by the Ethics Committee of hospitals (National Cheng Kung University Hospital IRB) on 3 February 2020. (Protocol code A-ER-107-425).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The dataset can be obtained with the corresponding author’s permission.

Acknowledgments

We thank all the participants, and this research was supported in part by Higher Education Sprout Project, Ministry of Education to the Headquarters of University Advancement at National Cheng Kung University (NCKU) and the Ministry of Science and Technology (MOST), Taipei, Taiwan (MOST 110-2628-B-006-020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Schirmer, A. Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing. Soc. Cogn. Affect. Neurosci. 2018, 13, 1–13. [Google Scholar] [CrossRef] [PubMed]
  2. Frith, C. Role of facial expressions in social interactions. Philos. Trans. R. Soc. B Biol. Sci. 2009, 364, 3453–3458. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Mier, D.; Lis, S.; Neuthe, K.; Sauer, C.; Esslinger, C.; Gallhofer, B.; Kirsch, P. The involvement of emotion recognition in affective theory of mind. Psychophysiology 2010, 47, 1028–1039. [Google Scholar] [CrossRef] [PubMed]
  4. Leppanen, J.M.; Hietanen, J.K. Emotion recognition and social adjustment in school-aged girls and boys. Scand. J. Psychol. 2001, 42, 429–435. [Google Scholar] [CrossRef] [PubMed]
  5. Ruffman, T.; Henry, J.; Livingstone, V.; Phillips, L.H. A meta-analytic review of emotion recognition and aging: Implications for neuropsychological models of aging. Neurosci. Biobehav. Rev. 2008, 32, 863–881. [Google Scholar] [CrossRef]
  6. Sullivan, S.; Ruffman, T.; Hutton, S.B. Age Differences in Emotion Recognition Skills and the Visual Scanning of Emotion Faces. J. Gerontol. Ser. B Psychol. Sci. Soc. Sci. 2007, 62, P53–P60. [Google Scholar] [CrossRef] [Green Version]
  7. Calder, A.J.; Keane, J.; Manly, T.; Sprengelmeyer, R.H.; Scott, S.; Nimmo-Smith, I.; Young, A.W. Facial expression recognition across the adult life span. Neuropsychologia 2003, 41, 195–202. [Google Scholar] [CrossRef]
  8. Hargrave, R.; Maddock, R.J.; Stone, V. Impaired recognition of facial expressions of emotion in Alzheimer’s disease. J. Neuropsychiatry Clin. Neurosci. 2002, 14, 64–71. [Google Scholar] [CrossRef]
  9. Yu, R.-L.; Chen, P.S.; Tu, S.-C.; Tsao, W.-C.; Tan, C.-H. Emotion-Specific Affective Theory of Mind Impairment in Parkinson’s Disease. Sci. Rep. 2018, 8, 16043. [Google Scholar] [CrossRef]
  10. Sachse, M.; Schlitt, S.; Hainz, D.; Ciaramidaro, A.; Walter, H.; Poustka, F.; Bölte, S.; Freitag, C.M. Facial emotion recognition in paranoid schizophrenia and autism spectrum disorder. Schizophr. Res. 2014, 159, 509–514. [Google Scholar] [CrossRef]
  11. Valdespino, A.; Antezana, L.; Ghane, M.; Richey, J.A. Alexithymia as a Transdiagnostic Precursor to Empathy Abnormalities: The Functional Role of the Insula. Front. Psychol. 2017, 8, 2234. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Haxby, J.V.; Hoffman, E.A.; Gobbini, M. Human neural systems for face recognition and social communication. Biol. Psychiatry 2002, 51, 59–67. [Google Scholar] [CrossRef]
  13. Mitchell, R.L.; Phillips, L. The overlapping relationship between emotion perception and theory of mind. Neuropsychologia 2015, 70, 1–10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Adolphs, R. Recognizing Emotion from Facial Expressions: Psychological and Neurological Mechanisms. Behav. Cogn. Neurosci. Rev. 2002, 1, 21–62. [Google Scholar] [CrossRef] [Green Version]
  15. Calder, A.J.; Keane, J.; Manes, F.; Antoun, N.; Young, A.W. Impaired recognition and experience of disgust following brain injury. Nat. Neurosci. 2000, 3, 1077–1078. [Google Scholar] [CrossRef]
  16. Dodich, A.; Cerami, C.; Canessa, N.; Crespi, C.; Marcone, A.; Arpone, M.; Realmuto, S.; Cappa, S. Emotion recognition from facial expressions: A normative study of the Ekman 60-Faces Test in the Italian population. Neurol. Sci. 2014, 35, 1015–1021. [Google Scholar] [CrossRef]
  17. Nowicki, S.; Duke, M.P. Nonverbal Receptivity: The Diagnostic Analysis of Nonverbal Accuracy (DANVA); Erlbaum Associates: Mahwah, NJ, USA, 2001. [Google Scholar]
  18. Ekman, P.; Friesen, W.V.; O’sullivan, M.; Chan, A.; Diacoyanni-Tarlatzis, I.; Heider, K.; Krause, R.; LeCompte, W.A.; Pitcairn, T.; Ricci-Bitti, P.E.; et al. Universals and cultural differences in the judgments of facial expressions of emotion. J. Personal. Soc. Psychol. 1987, 53, 712. [Google Scholar] [CrossRef]
  19. Jack, R.E.; Garrod, O.G.B.; Yu, H.; Caldara, R.; Schyns, P.G. Facial expressions of emotion are not culturally universal. Proc. Natl. Acad. Sci. USA 2012, 109, 7241–7244. [Google Scholar] [CrossRef] [Green Version]
  20. Jack, R.E.; Blais, C.; Scheepers, C.; Schyns, P.G.; Caldara, R. Cultural Confusions Show that Facial Expressions Are Not Universal. Curr. Biol. 2009, 19, 1543–1548. [Google Scholar] [CrossRef] [Green Version]
  21. Jack, R.E.; Caldara, R.; Schyns, P.G. Internal representations reveal cultural diversity in expectations of facial expressions of emotion. J. Exp. Psychol. Gen. 2012, 141, 19–25. [Google Scholar] [CrossRef] [Green Version]
  22. Matsumoto, D.; LeRoux, J.; Wilson-Cohn, C.; Raroque, J.; Kooken, K.; Ekman, P.; Yrizarry, N.; Loewinger, S.; Uchida, H.; Yee, A.; et al. A New Test to Measure Emotion Recognition Ability: Matsumoto and Ekman’s Japanese and Caucasian Brief Affect Recognition Test (JACBART). J. Nonverbal Behav. 2000, 24, 179–209. [Google Scholar] [CrossRef]
  23. Huang, C.L.-C.; Hsiao, S.; Hwu, H.-G.; Howng, S.-L. The Chinese Facial Emotion Recognition Database (CFERD): A computer-generated 3-D paradigm to measure the recognition of facial emotional expressions at different intensities. Psychiatry Res. 2012, 200, 928–932. [Google Scholar] [CrossRef] [PubMed]
  24. Bowers, D.; Blonder, L.; Heilman, K. Florida Affect Battery; Center for Neuropsychological Studies, Department of Neurology: Gainesville, FL, USA, 1998. [Google Scholar]
  25. Sedda, A.; Petito, S.; Guarino, M.; Stracciari, A. Identification and intensity of disgust: Distinguishing visual, linguistic and facial expressions processing in Parkinson disease. Behav. Brain Res. 2017, 330, 30–36. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, Y.C.; Yeh, Z.T.; Tsai, M.C. The Validity of the Florida Affect Battery Test and Performance in Patients with Prefrontal Cortex Damage. J. Ment. Health 2012, 25, 299–334. [Google Scholar]
  27. McDonald, S.; Flanagan, S.; Rollins, J.; Kinch, J. TASIT: A New Clinical Tool for Assessing Social Perception After Traumatic Brain Injury. J. Head Trauma Rehabil. 2003, 18, 219–238. [Google Scholar] [CrossRef]
  28. Bänziger, T.; Grandjean, D.; Scherer, K.R. Emotion recognition from expressions in face, voice, and body: The Multimodal Emotion Recognition Test (MERT). Emotion 2009, 9, 691–704. [Google Scholar] [CrossRef] [Green Version]
  29. Schlegel, K.; Grandjean, D.; Scherer, K.R. Introducing the Geneva Emotion Recognition Test: An example of Rasch-based test development. Psychol. Assess. 2014, 26, 666–672. [Google Scholar] [CrossRef] [Green Version]
  30. Schlegel, K.; Scherer, K.R. Introducing a short version of the Geneva Emotion Recognition Test (GERT-S): Psychometric properties and construct validation. Behav. Res. Methods 2016, 48, 1383–1392. [Google Scholar] [CrossRef]
  31. Laukka, P.; Bänziger, T.; Israelsson, A.; Cortes, D.S.; Tornberg, C.; Scherer, K.R.; Fischer, H. Investigating individual differences in emotion recognition ability using the ERAM test. Acta Psychol. 2021, 220, 103422. [Google Scholar] [CrossRef]
  32. Tan, C.B.; Sheppard, E.; Stephen, I.D. A change in strategy: Static emotion recognition in Malaysian Chinese. Cogent Psychol. 2015, 2, 1085941. [Google Scholar] [CrossRef]
  33. Xia, M.; Li, X.; Zhong, H.; Li, H. Fixation Patterns of Chinese Participants while Identifying Facial Expressions on Chinese Faces. Front. Psychol. 2017, 8, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  35. Cheng, C.M.; Chen, H.C.; Chan, Y.C.; Su, Y.C.; Tseng, C.C. Taiwan corpora of Chinese emotions and relevant psychophysiological data—Normative Data for Chinese Jokes. Chin. J. Psychol. 2013, 55, 555–569. [Google Scholar]
  36. Chiou, B.-C.; Chen, C.-P. Speech emotion recognition with cross-lingual databases. In Fifteenth Annual Conference of the International Speech Communication Association; ISCA: Baixas, France, 2014. [Google Scholar]
  37. Huang, S.T.; Lee, M.C.; Lee, L.; Chan, Y.T.; Tsai, H.T. Taiwan corpora of Chinese emotional stimuli database—The study of emotional prosody. Chin. J. Psychol. 2014, 56, 383–398. [Google Scholar]
  38. Baron-Cohen, S.; Wheelwright, S.; Hill, J.; Raste, Y.; Plumb, I. The “Reading the Mind in the Eyes” test revised version: A study with normal adults, and adults with Asperger syndrome or high-functioning autism. J. Child Psychol. Psychiatry 2001, 42, 241–251. [Google Scholar] [CrossRef] [PubMed]
  39. Waaramaa, T. Gender differences in identifying emotions from auditory and visual stimuli. Logop. Phoniatr. Vocology 2016, 42, 160–166. [Google Scholar] [CrossRef]
  40. Scherer, K.R.; Johnstone, T.; Klasmeyer, G. Vocal Expression of Emotion; Oxford University Press: Oxford, UK, 2003. [Google Scholar]
  41. Blair, R.J.R. Fine Cuts of Empathy and the Amygdala: Dissociable Deficits in Psychopathy and Autism. Q. J. Exp. Psychol. 2008, 61, 157–170. [Google Scholar] [CrossRef]
  42. Abbruzzese, L.; Magnani, N.; Robertson, I.H.; Mancuso, M. Age and Gender Differences in Emotion Recognition. Front. Psychol. 2019, 10, 2371. [Google Scholar] [CrossRef] [Green Version]
  43. Trauffer, N.M.; Widen, S.C.; Russell, J.A. Education and the attribution of emotion to facial expressions. Psihol. Teme 2013, 22, 237–247. [Google Scholar]
Figure 1. The diagrammatic sketch of the MMER app, taking subtest four as an example. (A) The step in which the participant’s information was entered; (B) the step to show the participants the instructions; (C) the step where the participant was asked to choose the correct answer.
Figure 1. The diagrammatic sketch of the MMER app, taking subtest four as an example. (A) The step in which the participant’s information was entered; (B) the step to show the participants the instructions; (C) the step where the participant was asked to choose the correct answer.
Brainsci 12 00251 g001
Figure 2. Model 1 is a one-factor (emotion recognition) model.
Figure 2. Model 1 is a one-factor (emotion recognition) model.
Brainsci 12 00251 g002
Figure 3. Model 2 and 3 are two-factor (prosody emotion recognition and facial emotion recognition) models with oblique and orthogonal rotation, respectively.
Figure 3. Model 2 and 3 are two-factor (prosody emotion recognition and facial emotion recognition) models with oblique and orthogonal rotation, respectively.
Brainsci 12 00251 g003
Figure 4. Model 4 and 5 are three-factor (facial recognition, prosody emotion recognition, and facial emotion recognition) models with oblique and orthogonal rotation, respectively.
Figure 4. Model 4 and 5 are three-factor (facial recognition, prosody emotion recognition, and facial emotion recognition) models with oblique and orthogonal rotation, respectively.
Brainsci 12 00251 g004
Table 3. The demographic characteristics and performance of participants in the MMER app.
Table 3. The demographic characteristics and performance of participants in the MMER app.
Demographic VariablesMeanRange
Gender (male %)32.54 (46.99%)-
Age, years48.28 (20.27 +)18–80
Education, years13.73 (2.97 +)6–20
Subtests of the MMER AppFull MarkCorrect Score
Mean (SD)
Correct Score RangeAccuracy
Mean (SD)
Total score198135.88 (22.14)71.00–174.000.68 (0.11)
Subtest 12422.42 (2.09)13.00–24.000.93 (0.09)
Subtest 33022.14 (4.08)10.00–29.000.74 (0.14)
Subtest 42015.74 (2.82)6.00–20.000.79 (0.14)
Subtest 52620.07 (3.96)3.00–26.000.77 (0.15)
Face-related subtests a10057.95 (9.85)24.00–74.000.76 (0.13)
Subtest 7 (prosody-related subtest)2511.12 (3.90)3.00–21.000.44 (0.16)
Subtest 83520.62 (5.19)5.00–30.000.59 (0.15)
Subtest 93823.76 (5.25)8.00–35.000.63 (0.14)
Face–prosody subtest b7344.38 (9.63)22.00–65.000.61 (0.13)
Abbreviations: MMER app—Multi-Modalities Emotion Recognition Mobile Application; SD—standard deviation; a—sum of subtests 1, 3, 4, and 5; b—sum of subtests 8 and 9; +—standard deviation.
Table 4. Correct score and accuracy of the MMER app in 7 types of emotion.
Table 4. Correct score and accuracy of the MMER app in 7 types of emotion.
EmotionFull MarkCorrect Score
Mean (SD)
Correct Score RangeAccuracy
Mean (SD)
Total score174113.46 (21.58)49.00–151.000.66 (0.12)
 Neutral2517.20 (4.12)4.00–25.000.69 (0.16)
 Happiness 2115.89 (2.61)9.00–21.000.76 (0.12)
 Sadness 2315.91 (4.02)4.00–22.000.69 (0.17)
 Angry 2520.17 (3.42)9.00–25.000.84 (0.14)
 Disgust2915.78 (4.43)4.00–24.000.54 (0.15)
 Fear2913.51 (4.60)4.00–23.000.47 (0.16)
Surprise2215.01 (3.43)3.00–21.000.68 (0.16)
Version and subtests of the MMER app
Face-related subtests a76
Neutral107.88 (1.69)2.00–10.000.79 (0.17)
Happiness 87.67 (0.72)3.00–8.000.96 (0.09)
Sadness 107.33 (1.75)2.00–10.000.73 (0.17)
Angry 1210.47 (1.62)4.00–12.000.87 (0.13)
Disgust1410.76 (2.98)1.00–14.000.77 (0.21)
Fear146.90 (3.05)1.00–14.000.49 (0.22)
Surprise86.95 (1.21)1.00–8.000.87 (0.15)
Prosody-related subtest (subtest 7)25
Neutral42.27 (1.17)0.00–4.000.57 (0.29)
Happiness 31.33 (0.92)0.00–3.000.44 (0.31)
Sadness 42.21 (1.23)0.00–4.000.55 (0.31)
Angry 21.59 (0.59)0.00–2.000.80 (0.30)
Disgust40.93 (0.92)0.00–3.000.23 (0.23)
Fear41.33 (1.03)0.00–4.000.33 (0.26)
Surprise41.46 (0.97)0.00–4.000.37 (0.24)
Face–prosody subtest b73
Neutral117.05 (2.43)1.00–11.000.64 (0.22)
Happiness 106.88 (1.83)2.00–10.000.69 (0.18)
Sadness 96.37 (1.88)1.00–9.000.71 (0.21)
Angry 118.11 (2.12)1.00–11.000.74 (0.19)
Disgust114.08 (1.87)0.00–9.000.37 (0.17)
Fear115.29 (1.79)1.00–9.000.48 (0.16)
Surprise106.60 (2.09)1.00–10.000.66 (0.21)
Abbreviations: please see Table 3. a—sum of subtests 3, 4, and 5; b—sum of subtests 8 and 9.
Table 5. Confirmatory factor analysis of MMER app with different models.
Table 5. Confirmatory factor analysis of MMER app with different models.
Modelχ2dfRMSEA (90% CI)CFITLIAIC
Model 1: Unidimensional45.282140.115 (0.079–0.153)0.9580.9367425.075
Model 2: Oblique (2 factors)12.44413<0.001 (0.000–0.074)1.0001.0017394.237
Model 3: Orthogonal (2 factors)174.631140.261 (0.227–0.296)0.7820.6737554.424
Model 4: Oblique (3 factors)10.78712<0.001 (0.000–0.072)1.0001.0037394.580
Model 5: Orthogonal (3 factors)186.673150.260 (0.228–0.294)0.7670.6747564.466
χ2—chi-square; df—degree of freedom; RMSEA—root mean square effort of approximation; CFI—comparative fit index; TLI—Tucker–Lewis Index; AIC—Akaike.
Table 6. Factor loadings for confirmatory factor analysis with models 4.
Table 6. Factor loadings for confirmatory factor analysis with models 4.
SubtestsFactors
Facial RecognitionFacial Emotion RecognitionProsody Emotion Recognition
11.000
3 0.861
4 0.858
5 0.840
7 0.846
8 0.891
9 0.807
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yu, R.-L.; Poon, S.-F.; Yi, H.-J.; Chien, C.-Y.; Hsu, P.-H. The Development of a Multi-Modality Emotion Recognition Test Presented via a Mobile Application. Brain Sci. 2022, 12, 251. https://doi.org/10.3390/brainsci12020251

AMA Style

Yu R-L, Poon S-F, Yi H-J, Chien C-Y, Hsu P-H. The Development of a Multi-Modality Emotion Recognition Test Presented via a Mobile Application. Brain Sciences. 2022; 12(2):251. https://doi.org/10.3390/brainsci12020251

Chicago/Turabian Style

Yu, Rwei-Ling, Shu-Fai Poon, Hsin-Jou Yi, Chia-Yi Chien, and Pei-Hsuan Hsu. 2022. "The Development of a Multi-Modality Emotion Recognition Test Presented via a Mobile Application" Brain Sciences 12, no. 2: 251. https://doi.org/10.3390/brainsci12020251

APA Style

Yu, R. -L., Poon, S. -F., Yi, H. -J., Chien, C. -Y., & Hsu, P. -H. (2022). The Development of a Multi-Modality Emotion Recognition Test Presented via a Mobile Application. Brain Sciences, 12(2), 251. https://doi.org/10.3390/brainsci12020251

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop