Machine Learning Algorithms for Detection and Classifications of Emotions in Contact Center Applications
Abstract
:1. Introduction
- ▪
- ▪
- ▪
- ▪
- ▪
- ▪
- Context—assessing the fit between the emotion and the context of expression [31].
- ▪
- ▪
- ▪
- ▪
- In the industry, by creating personalized offerings for specific audiences [45].
- ▪
- Development of the Emotion Classification for Machine Detection of Affect-Tinged Conversational Contents directly dedicated to the CC industry;
- ▪
- Verification of the developed Classification based on the developed algorithms for machine detection of customers’ affective states in CC voice and text channels.
2. Method
2.1. Samples
2.2. Testing Procedure
2.2.1. Emotion Classification—Voice and Text Channel
2.2.2. Machine Emotion Detection
2.2.3. Assessing the Usefulness of Classification in Machine Detection of Emotions
3. Results
3.1. Emotion Classification for Machine Detection of the Affective Coloring of Conversational Content in Voice and Text Channels of CC Systems
3.2. Findings on Machine Detection of Emotions in CC Voice and Text Channels
4. Discussion
5. Limitations
- ▪
- For voice conversations conducted in natural context, noise interference from the environment can affect the correctness of emotion detection by the judges, as well as the performance of the developed system. Due to the typically poor quality of recordings and lack of standardization, detection based on listening to natural utterances (including those from CC systems) is far more difficult than detection based on voice samples prepared by trained actors for experiments [16,70]. The recorded voice strength in natural conversations, as well as a number of other parameters, is not necessarily indicative of, for example, a person’s arousal, but of the quality of the telephone call, the quality of the equipment used for the call, the distance of the handset from the mouth, and many other important interfering variables. Moreover, actors’ utterances are usually exaggerated, and experimental samples are abstracted from the natural context of the utterances, i.e., their place, purpose, etc., i.e., deprived of information critical for the direction of interpretation of the expressed emotions [70,85];
- ▪
- Similar problems can also be encountered in the case of written text samples, e.g., when a person, due to time pressure, simplifies the message and makes syntactic and grammatical errors, which can then significantly affect the detection of emotions by the judges as well as the information technology system;
- ▪
- The accuracy of emotion detection by judges and then the system may be higher for intentionally prepared materials, where emotions are induced under standardized conditions using pre-prepared affective stimuli, e.g., excerpts from horror films, comedy, etc., compared to natural conversations recorded in CCs as an example [82]. It follows that the results of studies that use intentionally prepared material have higher relevance but lower ecological accuracy [86]. In contrast, results from studies based on conversations conducted in natural settings have relatively lower reliability but higher ecological accuracy and, therefore, generalizability to other natural settings and contexts [87];
- ▪
- The cultural context of the study may be an important limitation. It cannot be ruled out that certain types of emotions may be more or less easily expressed in speech and writing, and consequently recognized, within specific groups of languages, e.g., Slavic vs. Germanic, Romance, Semitic, etc. Furthermore, speakers of different languages may have different sensitivities to specific groups of emotions expressed in speech or writing [88,89]. The problem discussed therefore concerns the representativeness of the results obtained. Perhaps, based on conversations conducted in a language other than Polish, different results would be obtained, contrasting with those presented. This issue is worth addressing in future research.
- ▪
- It is worth noting that the field of psychology has also developed a concept that makes the strength of emotional reactions dependent on the language used at a given moment. For example, Zajonc [90] noted that humans are often unaware of the role of emotions in decision making—especially under conditions of uncertainty. This process does not occur as strongly in a foreign language as in the native language. A phenomenon called the foreign-language effect has been defined as a result of research that has shown weaker emotional responses when content is processed in a language the user is less fluent in rather than the native language [91]. Despite understanding its meaning, a language user always perceives more emotionally a message formulated in the native language than in a foreign language, which is supposed to promote the rationalization of thinking in a foreign language. The reason for the foreign language effect is most likely due to the lower level of emotional reactions felt when thinking in a foreign language [92].
- ▪
- Interview samples were obtained within the three interview topics mentioned earlier (invoices and payments; technical information; contracts and annexes). As such, it is difficult to assess how representative the results derived from them are of the entire CC context and, more broadly, how typical the collected sample of CC conversations is of contexts other than the CC system. This issue can be considered in the context of further research on the topic discussed in the paper. To be able to conduct this research, it may be necessary to optimize selected components of the authors’ emotion recognition method used. Among other things, for conversations in the text channel, the integrated dictionary of emotional expressions may need to be supplemented to some extent. This dictionary was implemented as a dedicated component of an emotion recognition method designed for both text channels and audio channels. In audio channels, the ability to use the dictionary is provided by automatically generated conversation transcripts. The ability to additionally include transcripts of conducted audio conversations increases the effectiveness of classification. Nevertheless, the transcription method used dedicated to the CC industry [60] was optimized for the above three topics. Therefore, another component that may require potential optimization is the integrated transcription method. The method would primarily need to optimize the algorithms operating on the post-processing side. In addition, appropriate learning processes can be implemented to make the solution even more versatile. The task of these solutions should be to continuously and dynamically adjust the reference models used. Intensive work is currently being carried out in this area, which is focused on the following possibilities: (a) the addition of learning processes to the existing model and (b) the preparation of a new model based on additional learning data, for which the classification results are considered taking into account the results obtained for the original model. Consideration of new learning data, particularly the amount of data, may further impact the need to optimize the data balancing techniques used. The potential activities listed above illustrate the additional work that may need to be conducted when implementing a solution for a completely different industry.
- ▪
- Not all emotion types known in psychology are identifiable to a satisfactory level of efficiency using intelligent computer systems. This limitation significantly narrows the possibilities in the development of the Emotion Classification for Machine Detection of Affect-Tinged Conversational Contents serving the needs of CC systems.
6. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Picard, R.W. Affective Computing; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
- Picard, R.W.; Vyzas, E.; Healey, J. Toward machine emotional intelligence: Analysis of affective physiological state. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 1175–1191. [Google Scholar] [CrossRef] [Green Version]
- Engelmann, J.; Pogosyan, M. Emotion perception across cultures: The role of cognitive mechanisms. Front. Psychol. 2013, 4, 118. [Google Scholar] [CrossRef] [Green Version]
- Dzedzickis, A.; Kaklauskas, A.; Bucinskas, V. Human Emotion Recognition: Review of Sensors and Methods. Sensors 2020, 20, 592. [Google Scholar] [CrossRef] [Green Version]
- Pal, S.; Mukhopadhyay, S.; Suryadevara, N. Development and Progress in Sensors and Technologies for Human Emotion Recognition. Sensors 2021, 21, 5554. [Google Scholar] [CrossRef]
- Saganowski, S. Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning. Electronics 2022, 11, 496. [Google Scholar] [CrossRef]
- Arana, J.M.; Gordillo, F.; Darias, J.; Mestas, L. Analysis of the efficacy and reliability of the Moodies app for detecting emotions through speech: Does it actually work? Comput. Hum. Behav. 2020, 104, 106156. [Google Scholar] [CrossRef]
- Er, M.B. A Novel Approach for Classification of Speech Emotions Based on Deep and Acoustic Features. IEEE Access 2020, 8, 221640–221653. [Google Scholar] [CrossRef]
- Mocanu, B.; Tapu, R.; Zaharia, T. Utterance Level Feature Aggregation with Deep Metric Learning for Speech Emotion Recognition. Sensors 2021, 21, 4233. [Google Scholar] [CrossRef]
- Koromilas, P.; Giannakopoulos, T. Deep Multimodal Emotion Recognition on Human Speech: A Review. Appl. Sci. 2021, 11, 7962. [Google Scholar] [CrossRef]
- Kedar, S.V.; Bormane, D.S.; Dhadwal, A.; Alone, S.; Agarwal, R. Automatic Emotion Recognition through Handwriting Analysis: A Review. In Proceedings of the International Conference on Computing Communication Control and Automation, Pune, India, 26–27 February 2015; pp. 811–816. [Google Scholar]
- Bhade, V.; Baraskar, T. A Model for Determining Personality by Analyzing Off-line Handwriting. Adv. Intell. Syst. Comput. 2018, 705, 345–354. [Google Scholar]
- Likforman-Sulem, L.; Esposito, A.; Faundez-Zanuy, M.; Clémençon, S. Extracting Style and Emotion from Handwriting. In Advances in Neural Networks: Computational and Theoretical Issues. Smart Innovation, Systems and Technologies; Bassis, S., Esposito, A., Morabito, F., Eds.; Springer: Cham, Switzerland, 2015; Volume 37, pp. 347–355. [Google Scholar]
- Erenel, Z.; Adegboye, O.R.; Kusetogullari, H. A New Feature Selection Scheme for Emotion Recognition from Text. Appl. Sci. 2020, 10, 5351. [Google Scholar] [CrossRef]
- Krommyda, M.; Rigos, A.; Bouklas, K.; Amditis, A. An Experimental Analysis of Data Annotation Methodologies for Emotion Detection in Short Text Posted on Social Media. Informatics 2021, 8, 19. [Google Scholar] [CrossRef]
- Johnstone, T.; Scherer, K.R. Vocal communication of emotion. In Handbook of Emotions, 2nd ed.; Lewis, M., Haviland-Jones, J.M., Eds.; The Guilford Press: New York, NY, USA, 2000; pp. 220–235. [Google Scholar]
- Sun, X.; Xia, P.; Ren, F. Multi-attention based Deep Neural Network with hybrid features for Dynamic Sequential Facial Expression Recognition. Neurocomputing 2021, 444, 378–389. [Google Scholar] [CrossRef]
- Dharanya, V.; Raj, A.N.J.; Gopi, V.P. Facial Expression Recognition through person-wise regeneration of expressions using Auxiliary Classifier Generative Adversarial Network (AC-GAN) based model. J. Vis. Commun. Image Represent. 2021, 77, 103110. [Google Scholar]
- Minaee, S.; Minaei, M.; Abdolrashidi, A. Deep-Emotion: Facial Expression Recognition Using Attentional Convolutional Network. Sensors 2021, 21, 3046. [Google Scholar] [CrossRef]
- Zhi, J.; Song, T.; Yu, K.; Yuan, F.; Wang, H.; Hu, G.; Yang, H. Multi-Attention Module for Dynamic Facial Emotion Recognition. Information 2022, 13, 207. [Google Scholar] [CrossRef]
- Lee, S.; Bae, M.; Lee, W.; Kim, H. CEPP: Perceiving the Emotional State of the User Based on Body Posture. Appl. Sci. 2017, 7, 978. [Google Scholar] [CrossRef] [Green Version]
- Sapiński, T.; Kamińska, D.; Pelikant, A.; Anbarjafari, G. Emotion Recognition from Skeletal Movements. Entropy 2019, 21, 646. [Google Scholar] [CrossRef] [Green Version]
- Caro, K.; Tentori, M.; Martinez-Garcia, A.I.; Zavala-Ibarra, I. FroggyBobby: An exergame to support children with motor problems practicing motor coordination exercises during therapeutic interventions. Comput. Hum. Behav. 2017, 71, 479–498. [Google Scholar] [CrossRef]
- Shi, J.; Liu, C.; Ishi, C.T.; Ishiguro, H. Skeleton-Based Emotion Recognition Based on Two-Stream Self-Attention Enhanced Spatial-Temporal Graph Convolutional Network. Sensors 2021, 21, 205. [Google Scholar] [CrossRef]
- Dissanayake, T.; Rajapaksha, Y.; Ragel, R.; Nawinne, I. An Ensemble Learning Approach for Electrocardiogram Sensor Based Human Emotion Recognition. Sensors 2019, 19, 4495. [Google Scholar] [CrossRef] [Green Version]
- Mehmood, R.; Lee, H. Towards Building a Computer Aided Education System for Special Students Using Wearable Sensor Technologies. Sensors 2017, 17, 317. [Google Scholar] [CrossRef]
- Chang, C.Y.; Chang, C.W.; Zheng, J.Y.; Chung, P.C. Physiological emotion analysis using support vector regression. Neurocomputing 2013, 122, 79–87. [Google Scholar] [CrossRef]
- Sharma, R.; Pachori, R.B.; Sircar, P. Automated emotion recognition based on higher order statistics and deep learning algorithm. Biomed. Signal Processing Contr. 2020, 58, 101867. [Google Scholar] [CrossRef]
- Shu, L.; Yu, Y.; Chen, W.; Hua, H.; Li, Q.; Jin, J.; Xu, X. Wearable Emotion Recognition Using Heart Rate Data from a Smart Bracelet. Sensors 2020, 20, 718. [Google Scholar] [CrossRef] [Green Version]
- Algarni, M.; Saeed, F.; Al-Hadhrami, T.; Ghabban, F.; Al-Sarem, M. Deep Learning-Based Approach for Emotion Recognition Using Electroencephalography (EEG) Signals Using Bi-Directional Long Short-Term Memory (Bi-LSTM). Sensors 2022, 22, 2976. [Google Scholar] [CrossRef]
- Deborah, A.S.; Rajalakshmi, S.; Rajendram, M.S.; Mirnalinee, T.T. Contextual Emotion Detection. Lect. Notes Data Eng. Commun. Technol. 2019, 35, 1179–1186. [Google Scholar]
- Lin, C.-J.; Lin, C.-H.; Wang, S.-H.; Wu, C.-H. Multiple Convolutional Neural Networks Fusion Using Improved Fuzzy Integral for Facial Emotion Recognition. Appl. Sci. 2019, 9, 2593. [Google Scholar] [CrossRef] [Green Version]
- Jianhua, T. Multimodal information processing for Affective Computing. In Speech Technology: Theory and Applications; Chen, F., Jokinen, K., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 151–166. [Google Scholar]
- Bertacchini, F.; Bilotta, E.; Pantano, P. Shopping with a robotic companion. Comput. Hum. Behav. 2017, 77, 382–395. [Google Scholar] [CrossRef]
- Park, S.; Cho, J.; Park, K.; Shin, H. Customer sentiment analysis with more sensibility. Eng. Appl. Artif. Intell. 2021, 104, 104356. [Google Scholar] [CrossRef]
- Aldayel, M.; Ykhlef, M.; Al-Nafjan, A. Deep learning for EEG-based preference classification in neuromarketing. Appl. Sci. 2020, 10, 1525. [Google Scholar] [CrossRef] [Green Version]
- Bojanić, M.; Delić, V.; Karpov, A. Call Redistribution for a Call Center Based on Speech Emotion Recognition. Appl. Sci. 2020, 10, 4653. [Google Scholar] [CrossRef]
- Spekman, M.L.C.; Konijn, E.A.; Hoorn, J.F. Perceptions of healthcare robots as a function of emotion-based coping: The importance of coping appraisals and coping strategies. Comput. Hum. Behav. 2018, 85, 308–318. [Google Scholar] [CrossRef] [Green Version]
- Picarra, N.; Giger, J.C. Predicting intention to work with social robots at anticipation stage: Assessing the role of behavioral desire and anticipated emotions. Comput. Hum. Behav. 2018, 86, 129–146. [Google Scholar] [CrossRef]
- Hui, T.K.L.; Sherratt, R.S. Coverage of Emotion Recognition for Common Wearable Biosensors. Biosensors 2018, 8, 30. [Google Scholar] [CrossRef] [Green Version]
- Yan, F.; Iliyasu, A.M.; Hirota, K. Emotion space modelling for social robots. Eng. Appl. Artif. Intell. 2021, 100, 104178. [Google Scholar] [CrossRef]
- Ashwin, T.S.; Ram, M.; Reddy, G. Affective database for e-learning and classroom environments using Indian students’ faces, hand gestures and body postures. Future Gener. Comput. Syst. 2020, 108, 334–348. [Google Scholar]
- Sarrafzadeh, A.; Alexander, S.; Dadgostar, F.; Fan, C.; Bigdeli, A. How do you know that I don’t understand? A look at the future of intelligent tutoring systems. Comput. Hum. Behav. 2008, 24, 1342–1363. [Google Scholar] [CrossRef]
- Crescenzi-Lanna, L. Emotions, private speech, involvement and other aspects of young children’s interactions with educational apps. Comput. Hum. Behav. 2020, 111, 106430. [Google Scholar] [CrossRef]
- Yan, Q.; Zhou, S.; Wu, S. The influences of tourists’ emotions on the selection of electronic word of mouth platforms. Tour. Manag. 2018, 66, 348–363. [Google Scholar] [CrossRef]
- Harley, J.M.; Bouchet, F.; Sazzad Hussain, S.; Azevedo, R.; Calvo, R. A multi-componential analysis of emotions during complex learning with an intelligent multi-agent system. Comput. Hum. Behav. 2015, 48, 615–625. [Google Scholar] [CrossRef]
- Chen, L.; Xie, B.; Zhang, C.; Zheng, J.; Limeng Xiao, L. An overview of the application of deep learning in face age image recognition. In Proceedings of the 2021 the 5th International Conference on Video and Image Processing (ICVIP 2021), Hayward CA, USA, 22–25 December 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 26–31. [Google Scholar]
- Wierzbicka, A. Emotions across Languages and Cultures: Diversity and Universals (Studies in Emotion and Social Interaction); Cambridge University Press: Cambridge, UK, 1999. [Google Scholar]
- Bryant, G.A.; Barrett, H.C. Vocal emotion recognition across disparate cultures. J. Cogn. Cult. 2008, 8, 135–148. [Google Scholar] [CrossRef] [Green Version]
- Laukka, P.; Elfenbein, H.A. Cross-Cultural Emotion Recognition and In-Group Advantage in Vocal Expression: A Meta-Analysis. Emot. Rev. 2021, 30, 3–11. [Google Scholar] [CrossRef]
- Rothbart, M.K.; Sheese, B.E.; Posner, M.I. Temperament and Emotion Regulation. In Handbook of Emotion Regulation, 2nd ed.; Gross, J.J., Ed.; Guilford Press: New York, NY, USA, 2014; pp. 331–350. [Google Scholar]
- Mauss, I.B.; Tamir, M. Emotion goals: How their content, structure, and operation shape emotion regulation. In Handbook of Emotion Regulation, 2nd ed.; Gross, J.J., Ed.; Guilford Press: New York, NY, USA, 2014; pp. 361–375. [Google Scholar]
- Keltner, D.; Tracy, J.; Sauter, D.A.; Cordaro, D.C.; McNeil, G. Expression of Emotion. In Handbook of Emotion Regulation, 4th ed.; Guilford Press: New York, NY, USA, 2018; pp. 467–482. [Google Scholar]
- Vogt, T.; André, E.; Wagner, J. Automatic Recognition of Emotions from Speech: A Review of the Literature and Recommendations for Practical Realisation. In Affect and Emotion in Human-Computer Interaction. Lecture Notes in Computer Science; Peter, C., Beale, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; p. 4868. [Google Scholar]
- Chatterjee, A.; Gupta, U.; Chinnakotla, M.K.; Srikanth, R.; Galley, M.; Agrawal, P. Understanding Emotions in Text Using Deep Learning and Big Data. Comput. Hum. Behav. 2018, 93, 309–317. [Google Scholar] [CrossRef]
- Lane, H.; Howard, C.; Hapke, H. Natural Language Processing in Action: Understanding, Analyzing, and Generating Text with Python; Manning Publications: Shelter Island, NY, USA, 2019. [Google Scholar]
- Anagnostopoulos, C.N.; Iliou, T.; Giannoukos, I. Features and classifiers for emotion recognition from speech: A survey from 2000 to 2011. Artif. Intell. Rev. 2015, 43, 155–177. [Google Scholar] [CrossRef]
- Jiang, W.; Wang, Z.; Jin, J.S.; Han, X.; Li, C. Speech Emotion Recognition with Heterogeneous Feature Unification of Deep Neural Network. Sensors 2019, 19, 2730. [Google Scholar] [CrossRef] [Green Version]
- Papakostas, M.; Siantikos, G.; Giannakopoulos, T.; Spyrou, E.; Sgouropoulos, D. Recognizing Emotional States Using Speech Information. Adv. Exp. Med. Biol. 2017, 989, 155–164. [Google Scholar]
- Płaza, M.; Pawlik, Ł.; Deniziak, S. Call Transcription Methodology for Contact Center Systems. IEEE Access 2021, 9, 110975–110988. [Google Scholar] [CrossRef]
- Pawlik, Ł.; Płaza, M.; Deniziak, S.; Boksa, E. A Method for Improving Bot Effectiveness by Recognising Implicit Customer Intent in Contact Centre Conversations. Speech Communication 2022. [Google Scholar]
- Sabharwal, N.; Agrawal, A. Introduction to Google Dialogflow. In Cognitive Virtual Assistants Using Google Dialogflow; Apress: Berkeley, CA, USA, 2020. [Google Scholar]
- Płaza, M.; Pawlik, Ł. Influence of the Contact Center Systems Development on Key Performance Indicators. IEEE Access 2021, 9, 44580–44591. [Google Scholar] [CrossRef]
- Scherer, K.R. Expression of emotion in voice and music. J. Voice 1995, 9, 235–248. [Google Scholar] [CrossRef] [Green Version]
- Juslin, P.N.; Laukka, P. Communication of Emotions in Vocal Expression and Music Performance: Different Channels, Same Code? Psychol. Bull. 2003, 129, 770–814. [Google Scholar] [CrossRef] [Green Version]
- Banse, R.; Scherer, K.R. Acoustic profiles in vocal emotion expression. J. Personal. Soc. Psychol. 1996, 70, 614–636. [Google Scholar] [CrossRef]
- Bachorowski, J.A.; Owren, M.J. Vocal expressions of emotion. In Handbook of Emotions, 3rd ed.; Lewis, M., Haviland-Jones, J.M., Feldman Barrett, I.L., Eds.; Gulford Press: New York, NY, USA; London, UK, 2008; pp. 196–210. [Google Scholar]
- Bachorowski, J.A. Vocal expression and perception of emotion. Curr. Dir. Psychol. Sci. 1999, 8, 53–57. [Google Scholar] [CrossRef]
- Kamiloğlu, R.G.; Fischer, A.H.; Sauter, D.A. Good vibrations: A Review of Vocal Expressions of Positive Emotions. Psychon. Bull. Rev. 2020, 27, 237–265. [Google Scholar] [CrossRef] [Green Version]
- Scherer, K.R. Vocal communication of emotion: A review of research paradigms. Speech Commun. 2003, 40, 227–256. [Google Scholar] [CrossRef]
- Strukowska, M. Anger in action: Socio-pragmatic analysis of verbal exchanges in UK parliamentary debates. Stud. Contemp. Linguist. 2019, 55, 671–703. [Google Scholar] [CrossRef]
- Aldunate, N.; Gonzalez-Ibanez, R. An integrated review of emoticons in computer-mediated communication. Front. Psychol. 2017, 7, 1–6. [Google Scholar] [CrossRef] [Green Version]
- LeDoux, J. Anxious: Using the Brain to Understand and Treat Fear and Anxiety; Viking: New York, NY, USA, 2015. [Google Scholar]
- McCulloch, G. Because Internet: Understanding the New Rules of Language; Riverhead Books: New York, NY, USA, 2020. [Google Scholar]
- Feldman, L.B.; Aragon, C.R.; Chen, N.C.; Kroll, J.F. Emoticons in text may function like gestures in spoken or signed communication. Behav. Brain Sci. 2017, 40, 26–27. [Google Scholar] [CrossRef]
- Danesi, M. The Semiotics of Emoji; Bloomsbury Publishing: London, UK, 2016. [Google Scholar]
- Chandrasekar, P.; Chapaneri, S.; Jayaswal, D. Emotion Recognition from Speech using Discriminative Features. Int. J. Comput. Appl. 2014, 101, 31–36. [Google Scholar] [CrossRef]
- Kou, J.; Zhang, W. Multi-kernel neural networks for nonlinear unsteady aerodynamic reduced-order modeling. Aerosp. Sci. Technol. 2017, 67, 309–326. [Google Scholar] [CrossRef]
- Chai, R.; Tsourdos, A.; Savvaris, A.; Chai, S.; Xia, Y.; Chen, C.L.P. Six-dof spacecraft optimal trajectory planning and real-time attitude con1195 trol: A deep neural network-based approach. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5005–5013. [Google Scholar] [CrossRef]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Scherer, K.R.; Banse, R.; Wallbott, H.G. Emotion Inferences from Vocal Expression Correlate Across Languages and Cultres. J. Cross-Cult. Psychol. 2001, 32, 76–92. [Google Scholar] [CrossRef] [Green Version]
- Juslin, P.N.; Laukka, P. Impact of intended emotion intensity on cue utilization and decoding accuracy in vocal expression of emotion. Emotion 2001, 1, 381–412. [Google Scholar] [CrossRef]
- Vasilateanu, A.; Ene, R. Call-Center Virtual Assistant Using Natural Language Processing and Speech Recognition. J. ICT Des. Eng. Technol. Sci. Juhriyansyah Dalle 2018, 2, 40–46. [Google Scholar] [CrossRef] [Green Version]
- Roche, J.M.; Peters, B.; Dale, R. Your Tone Says It All: The processing and interpretation of affective language. Speech Commun. 2015, 66, 47–64. [Google Scholar] [CrossRef]
- Russell, J.A.; Bachorowski, J.A.; Fernández-Dols, J.M. Facial and Vocal Expressions of Emotion. Annu. Rev. Psychol. 2003, 54, 329–349. [Google Scholar] [CrossRef] [Green Version]
- Drolet, M.; Schubotz, R.I.; Fischer, J. Authenticity affects the recognition of emotions in speech: Behavioral and fMRI evidence. Cogn. Affect. Behav. Neurosci. 2012, 12, 140–150. [Google Scholar] [CrossRef] [Green Version]
- Jürgens, R.; Grass, A.; Drolet, M.; Fischer, J. Effect of Acting Experience on Emotion Expression and Recognition in Voice: Non-Actors Provide Better Stimuli than Expected. J. Nonverbal. Behav. 2015, 39, 195–214. [Google Scholar] [CrossRef] [Green Version]
- Lausen, A.; Schacht, A. Gender Differences in the Recognition of Vocal Emotions. Front. Psychol. 2018, 9, 882. [Google Scholar] [CrossRef]
- Tanaka, A.; Koizumi, A.; Imai, H.; Hiramatsu, S.; Hiramoto, E.; de Gelder, B. I feel your voice. Cultural differences in the multisensory perception of emotion. Psychol. Sci. 2010, 21, 1259–1262. [Google Scholar] [CrossRef]
- Zajonc, R.B. Feeling and Thinking: Preferences Need no Inferences. Am. Psychol. 1980, 35, 151–175. [Google Scholar] [CrossRef]
- Opitz, B.; Degner, J. Emotionality in a Second Language: It’s a Matter of Time. Neuropsychologia 2012, 50, 1961–1967. [Google Scholar] [CrossRef] [Green Version]
- Costa, A.; Foucart, A.; Arnon, I.; Aparici, M.; Apesteguia, J. “Piensa” Twice: On the Foreign Language Effect in Decision Making. Cognition 2014, 130, 236–254. [Google Scholar] [CrossRef]
- Królikowski, M.; Płaza, M.; Szcześniak, Z. Chosen sources of signal interference in HD-TVI technology. Proc. Spie 2017, 10445, 104455M. [Google Scholar]
- Wilk-Jakubowski, J. Information systems engineering using VSAT networks. Yugosl. J. Oper. Res. 2021, 31, 409–428. [Google Scholar] [CrossRef]
System | Affective State | Parameters | Technical Solutions | References |
---|---|---|---|---|
An analysis of the emotions of people using intelligent tutoring systems (ITS) | Anger, neutrality, disgust, sadness, fear, happiness, surprise | Facial expressions, skin conductivity | Sensor electrodermal activity (EDA), face reader, recording camera, neural network | [46] |
Dynamic facial expression recognition sequence | Neutrality, disgust, sadness, fear, happiness, surprise, anger | Facial expressions | Convolutional Neural Network (CNN) | [47] |
Detection of emotions based on computer analysis of body posture | Lividness, boredom, disgust, happiness, interest | Body posture | Algorithm on C++, a sensor that analyzes body posture | [21] |
Detection of emotions based on gestures and body movements | Happiness, sadness, surprise, fear, anger, disgust and neutral state | Gestures and body movements | A sensor that analyzes posture, Convolutional Neural Network (CNN) | [22] |
Moodies for voice-based emotion detection | Disgust, happiness, anger, fear, tenderness and sadness | Sound | An app that detects the emotion in people’s voice | [7] |
Techniques for recognizing emotions from voice | Anger, happiness, sadness and neutral state | Sound | Deep neural networks, hybrid CNN and SVM model | [8] |
A prototype system for detecting emotions in a text based on social media posts | Anger, anticipation, disgust, fear, joy, sadness, surprise, trust | Text | Long Short Term Memory (LSTM) networks | [15] |
A model for emotion recognition based on ECG signal analysis | Happiness, sadness, pleasure, anger | ECG | Spiker-Shield Heart and Brain sensor, Extra Tree Classification, ADA Boost Classification with SVM, Python Scikit API | [25] |
Development of an emotion recognition system based on physiological reactions of the organism | Sadness, fear and pleasure | ECG 1, GSR 2, BVP 3, pulse, respiration | A system with five physiological signal sensors (ML870 PowerLab 8/30 sensor), Support vector regression (SVR) | [27] |
Emotion Recognition Using Heart Rate Data from a Smart Bracelet | Happiness, sadness and neutral | Pulse | A smart bracelet (Algoband F8), k-Nearest Neighbor (kNN), Random Forests (RF), Decision Tree (DT), Gradient Boosting Decision Tree (GBDT), Adaptive Boosting | [29] |
Affective State | Descriptive Statistics | Family of Related Emotions | |
---|---|---|---|
N(%) 1 | L(%) 2 | ||
Anger | 937 (32) | 1 h 19 min 7 s (5.74) | irritation, impatience, negative surprise, disappointment, bitterness, anger, irony, sarcasm, rage |
Examples: “This agreement is ridiculous, it’s not insurance at all;” “Wait, wait, are you kidding me?! This is complete nonsense;” “Why do you cheat people like this?! I’ve legitimately been duped.” | |||
Fear | 166 (5) | 36 min 58 s (2.68) | uncertainty, fear, worry, confusion, anxiety, panic |
Examples: “It’s taking an awfully long time, could you please hurry up?;” “This sort of thing shouldn’t happen because the GDPR is in force now;” “I’ve tried to explain but there seem to be problems, I’m having a problem.” | |||
Happiness | 196 (7) | 3 min 16 s (0.24) | interest, satisfaction, positive surprise, excitement, gratitude, hope, happiness, amusement |
Examples: “I am very happy, this is great, thank you very much;” “Thank you kindly, you helped me a lot, have a nice day;” “Okay, let’s do it this way, I am very happy, everything has been explained.” | |||
Neutral | 1491 (51) | 14 h 36 min 46 s (63.57) | not applicable |
Examples: “Okay, thank you for the information;” “Yes, that’s right, I understand everything;” “I would like to know something about my complaint. Can we please check it out? Thank you.” | |||
Sadness | 145 (5) | 7 min 18 s (0.53) | resignation, bitterness, helplessness, regret, melancholy |
Examples: “Oh well, that’s too bad, if it can’t be handled otherwise;” “It’s a bit unfair (…) This doesn’t sit square with me;” “And now what? I don’t want that phone, I don’t understand, this is impossible” |
Affective State | Descriptive Statistics | Family of Related Emotions | |
---|---|---|---|
N (%) 1 | L (%) 2 | ||
Anger | 312 (4.15) | 9910 (5.67) | irritation, impatience, negative surprise, disappointment, bitterness, anger, irony, sarcasm, rage |
Examples: “The bubbles are still inactive—As they’ve been for five days;” “2. I’m a serious guy and i don’t have time for these divagations: (“Are you kidding me?; “I don’t believe it, you guys are in such a mess it’s unbelievable.” | |||
Fear | 102 (1.35) | 3659 (2.09) | uncertainty, fear, worry, confusion, anxiety, panic |
Examples: “Pesel necessary:-O;” “GDPR”; “whoops… something’s gone wrong” | |||
Happiness | 761 (10.12) | 21,521 (12.32) | interest, satisfaction, positive surprise, excitement, gratitude, hope, happiness, amusement |
Examples: “Well that’s great. Thank you sincerely);” “Concrete answer! Thank you! You helped me a lot!;” “Have a nice day XD” | |||
Neutral | 2269 (30.19) | 90,817 (52.00) | not applicable |
Examples: “How long will the migration take?;” “I’m having trouble logging into the a;” “What are the internet access packages?” | |||
Sadness | 305 (4.05) | 8741 (5.00) | resignation, bitterness, helplessness, regret, melancholy |
Examples: “I’ve been misled: (“I’ve written to you, but unfortunately no one is writing back;” “But it doesn’t work on my phone yet: (Why?)” |
No | Classifier Type | Voice Channel | |||||
---|---|---|---|---|---|---|---|
Accuracy | SD 1 | Precision | SD 1 | F1-Score | SD 1 | ||
[%] | [%] | [%] | [%] | [%] | [%] | ||
1. | CNN | 72.9 | 2.58 | 80.7 | 0.89 | 75.0 | 1.58 |
2. | kNN | 70.0 | 1.00 | 67.8 | 1.64 | 67.2 | 1.64 |
3. | SVM | 69.2 | 3.11 | 69.4 | 5.81 | 63.4 | 3.84 |
No | Classifier Type | Text Channel | |||||
---|---|---|---|---|---|---|---|
Accuracy | SD 1 | Precision | SD 1 | F1-Score | SD 1 | ||
[%] | [%] | [%] | [%] | [%] | [%] | ||
1. | ANN | 63.9 | 2.45 | 64.0 | 2.49 | 64.4 | 2.57 |
2. | DT | 54.4 | 2.23 | 58.1 | 4.34 | 55.6 | 3.14 |
3. | kNN | 55.5 | 2.14 | 62.1 | 4.98 | 56.1 | 3.45 |
4. | RFC | 53.4 | 1.58 | 66.0 | 2.81 | 55.8 | 2.78 |
5. | SVM | 49.2 | 1.54 | 57.7 | 4.57 | 49.1 | 4.18 |
No | Classifier Type | Accuracy [%] | Precision [%] | F1-Score [%] |
---|---|---|---|---|
VOICE CHANNEL | ||||
1. | CNN | 67.5 | 80.3 | 74.5 |
2. | kNN | 52.7 | 67.6 | 57.5 |
3. | SVM | 62.4 | 63.9 | 62.2 |
TEXT CHANNEL | ||||
1. | ANN | 55.8 | 62.4 | 58.4 |
2. | DT | 49.6 | 58.9 | 53.4 |
3. | kNN | 55.3 | 57.2 | 55.7 |
4. | RFC | 56.5 | 59.2 | 57.0 |
5. | SVM | 65.9 | 58.5 | 61.7 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Płaza, M.; Trusz, S.; Kęczkowska, J.; Boksa, E.; Sadowski, S.; Koruba, Z. Machine Learning Algorithms for Detection and Classifications of Emotions in Contact Center Applications. Sensors 2022, 22, 5311. https://doi.org/10.3390/s22145311
Płaza M, Trusz S, Kęczkowska J, Boksa E, Sadowski S, Koruba Z. Machine Learning Algorithms for Detection and Classifications of Emotions in Contact Center Applications. Sensors. 2022; 22(14):5311. https://doi.org/10.3390/s22145311
Chicago/Turabian StylePłaza, Mirosław, Sławomir Trusz, Justyna Kęczkowska, Ewa Boksa, Sebastian Sadowski, and Zbigniew Koruba. 2022. "Machine Learning Algorithms for Detection and Classifications of Emotions in Contact Center Applications" Sensors 22, no. 14: 5311. https://doi.org/10.3390/s22145311
APA StylePłaza, M., Trusz, S., Kęczkowska, J., Boksa, E., Sadowski, S., & Koruba, Z. (2022). Machine Learning Algorithms for Detection and Classifications of Emotions in Contact Center Applications. Sensors, 22(14), 5311. https://doi.org/10.3390/s22145311