Next Article in Journal
Energy Efficient Routing in Wireless Sensor Networks: A Comprehensive Survey
Next Article in Special Issue
Ensemble Deep Learning for Multilabel Binary Classification of User-Generated Content
Previous Article in Journal
Oil Spill Monitoring of Shipborne Radar Image Features Using SVM and Local Adaptive Threshold
Previous Article in Special Issue
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspectral Image Analysis and Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition

by
Kudakwashe Zvarevashe
and
Oludayo Olugbara
*
ICT and Society Research Group, South Africa Luban Workshop, Durban University of Technology, Durban 4001, South Africa
*
Author to whom correspondence should be addressed.
Algorithms 2020, 13(3), 70; https://doi.org/10.3390/a13030070
Submission received: 15 January 2020 / Revised: 14 March 2020 / Accepted: 17 March 2020 / Published: 22 March 2020
(This article belongs to the Special Issue Ensemble Algorithms and Their Applications)

Abstract

:
Automatic recognition of emotion is important for facilitating seamless interactivity between a human being and intelligent robot towards the full realization of a smart society. The methods of signal processing and machine learning are widely applied to recognize human emotions based on features extracted from facial images, video files or speech signals. However, these features were not able to recognize the fear emotion with the same level of precision as other emotions. The authors propose the agglutination of prosodic and spectral features from a group of carefully selected features to realize hybrid acoustic features for improving the task of emotion recognition. Experiments were performed to test the effectiveness of the proposed features extracted from speech files of two public databases and used to train five popular ensemble learning algorithms. Results show that random decision forest ensemble learning of the proposed hybrid acoustic features is highly effective for speech emotion recognition.

1. Introduction

Emotion plays an important role in the daily interpersonal interactions and is considered an essential skill for human communication [1]. It helps humans to understand the opinions of others by conveying feelings and giving feedback to people. Emotion has many useful benefits of affective computing and cognitive activities such as rational decision making, perception and learning [2]. It has opened up an exhilarating research agenda because constructing an intelligent robotic dialogue system that can recognize emotions and precisely respond in the manner of human conversation is presently arduous. The requirement of emotion recognition is steadily increasing with the pervasiveness of intelligent systems [3]. Huawei intelligent video surveillance systems, for instance, can support real-time tracking of a person in a distressed phase through emotion recognition. The capability to recognize human emotions is considered an essential future requirement of intelligent systems that are inherently supposed to interact with people to a certain degree of emotional intelligence [4]. The necessity to develop emotionally intelligent systems is exceptionally important for the modern society of the internet of things (IoT) because such systems have great impact on decision making, social communication and smart connectivity [5].
Practical applications of emotion recognition systems can be found in many domains such as audio/video surveillance [6], web-based learning, commercial applications [7], clinical studies, entertainment [8], banking [9], call centers [10], computer games [11] and psychiatric diagnosis [12]. In addition, other real applications include remote tracking of persons in a distressed phase, communication between human and robots, mining sentiments of sport fans and customer care services [13], where emotion is perpetually expressed. These numerous applications have led to the development of emotion recognition systems that use facial images, video files or speech signals [14]. In particular, speech signals carry emotional messages during their production [15] and have led to the development of intelligent systems habitually called speech emotion recognition systems [16]. There is an avalanche of intrinsic socioeconomic advantages that make speech signals a good source for affective computing. They are economically easier to acquire than other biological signals like electroencephalogram, electrooculography and electrocardiograms [17], which makes speech emotion recognition research attractive [18]. Machine learning algorithms extract a set of speech features with a variety of transformations to appositely classify emotions into different classes. However, the set of features that one chooses to train the selected learning algorithm is one of the most important tools for developing effective speech emotion recognition systems [3,19]. Research has suggested that features extracted from the speech signal have a great effect on the reliability of speech emotion recognition systems [3,20], but selecting an optimal set of features is challenging [21].
Speech emotion recognition is a difficult task because of several reasons such as an ambiguous definition of emotion [22] and the blurring of separation between different emotions [23]. Researchers are investigating different heterogeneous sources of features to improve the performance of speech emotion recognition systems. In [24], performances of Mel-frequency cepstral coefficient (MFCC), linear predictive cepstral coefficient (LPCC) and perceptual linear prediction (PLP) features were examined for recognizing speech emotion that achieved a maximum accuracy of 91.75% on an acted corpus with PLP features. This accuracy is relatively low when compared to the recognition accuracy of 95.20% obtained for a fusion of audio features based on a combination of MFCC and pitch for recognizing speech emotion [25]. Other researchers have tried to agglutinate different acoustic features with the optimism of boosting accuracy and precision rates of speech emotion recognition [25,26]. This technique has shown some improvement, nevertheless it has yielded low accuracy rates for the fear emotion in comparison with other emotions [25,26]. Semwal et al. [26] fused some acoustic features such as MFCC, energy, zero crossing rate (ZCR) and fundamental frequency that gave an accuracy of 77.00% for the fear emotion. Similarly, Sun et al. [27] used a deep neural network (DNN) to extract bottleneck features that achieved an accuracy of 62.50% for recognizing the fear emotion. The overarching objective of this study was to construct a set of hybrid acoustic features (HAFs) to improve the recognition of the fear emotion and other emotions from speech signal with high precision. This study has contributed to the understanding of the topical theme of speech emotion recognition in the following unique ways.
  • The application of a specialized software to extract highly discriminating speech emotion feature representations from multiple sources such as prosodic and spectral to achieve an improved precision in emotion recognition.
  • The agglutination of the extracted features using the specialized software to form a set of hybrid acoustic features that can recognize the fear emotion and other emotions from speech signal better than the state-of-the-art unified features.
  • The comparison of the proposed set of hybrid acoustic features with other prominent features of the literature using popular machine learning algorithms to demonstrate through experiments, the effectiveness of our proposal over the others.
The content of this paper is succinctly organized as follows. Section 1 provides the introductory message, including the objective and contributions of the study. Section 2 discusses the related studies in chronological order. Section 3 designates the experimental databases and the study methods. The details of the results and concluding statements are given in Section 4 and Section 5 respectively.

2. Related Studies

Speech emotion recognition research has been exhilarating for a long time and several papers have presented different ways of developing systems for recognizing human emotions. The authors in [3] presented a majority voting technique (MVT) for detecting speech emotion using fast correlation based feature (FCBF) and Fisher score algorithms for feature selection. They extracted 16 low-level features and tested their methods over several machine learning algorithms, including artificial neural network (ANN), classification and regression tree (CART), support vector machine (SVM) and K-nearest neighbor (KNN) on berlin emotion speech database (Emo-DB). Kerkeni et al. [11] proposed a speech emotion recognition method that extracted MFCC with modulation spectral (MS) features and used recurrent neural network (RNN) learning algorithm to classify seven emotions from Emo-DB and Spanish database. The authors in [15] proposed a speech emotion recognition model where glottis was used for compensation of glottal features. They extracted features of glottal compensation to zero crossings with maximal teager (GCZCMT) energy operator using Taiyuan University of technology (TYUT) speech database and Emo-DB.
Evaluation of feature selection algorithms on a combination of linear predictive coefficient (LPC), MFCC and prosodic features with three different multiclass learning algorithms to detect speech emotion was discussed in [19]. Luengo et al. [20] used 324 spectral and 54 prosody features combined with five voice quality features to test their proposed speech emotion recognition method on the Surrey audio-visual expressed emotion (SAVEE) database after applying the minimal redundancy maximal relevance (mRMR) to reduce less discriminating features. In [28], a method for recognizing emotions in an audio conversation based on speech and text was proposed and tested on the SemEval-2007 database using SVM. Liu et al. [29] used the extreme learning machine (ELM) method for feature selection that was applied to 938 features based on a combination of spectral and prosodic features from Emo-DB for speech emotion recognition. The authors in [30] extracted 988 spectral and prosodic features from three different databases using the OpenSmile toolkit with SVM for speech emotion recognition. Stuhlsatz et al. [31] introduced the generalized discriminant analysis (GerDA) based on DNN to recognize emotions from speech using the OpenEar specialized software to extract 6552 acoustic features based on 39 functional of 56 acoustic low-level descriptor (LLD) from Emo-DB and speech under simulated and actual stress (SUSAS) database. Zhang et al. [32] applied the cooperative learning method to recognize speech emotion from FAU Aibo and SUSAS databases using GCZCMT based acoustic features.
In [33], Hu moments based weighted spectral features (HuWSF) were extracted from Emo-DB, SAVEE and Chinese academy of sciences - institute of automation (CASIA) databases to classify emotions from speech using SVM. The authors used HuWSF and multicluster feature selection (MCFS) algorithm to reduce feature dimensionality. In [34], extended Geneva minimalistic acoustic parameter set (eGeMAPS) features were used to classify speech emotion, gender and age from the Ryerson audio-visual database of emotional speech and song (RAVDESS) using the multiple layer perceptron (MLP) neural network. Pérez-Espinosa et al. [35] analyzed 6920 acoustic features from the interactive emotion dyadic motion capture (IEMOCAP) database to discover that features based on groups of MFCC, LPC and cochleagrams are important for estimating valence, activation and dominance emotions in speech respectively. The authors in [36] proposed a DNN architecture for extracting informative feature representatives from heterogeneous acoustic feature groups that may contain redundant and unrelated information. The architecture was tested by training the fusion network to jointly learn highly discriminating acoustic feature representations from the IEMOCAP database for speech emotion recognition using SVM to obtain an overall accuracy of 64.0%. In [37], speech features based on MFCC and facial on maximally stable extremal region (MSER) were combined to recognize human emotions through a systematic study on Indian face database (IFD) and Emo-DB.
Narendra and Alku [38] proposed a new dysarthric speech classification method from a coded telephone speech using glottal features with a DNN based glottal inverse filtering method. They considered two sets of glottal features based on time and frequency domain parameters plus parameters based on principal component analysis (PCA). Their results showed that a combination of glottal and acoustic features resulted in an improved classification after applying PCA. In [39], four pitch and spectral energy features were combined with two prosodic features to distinguish two high activation states of angry and happy plus low activation states of sadness and boredom for speech emotion recognition using SVM with Emo-DB. Alshamsi et al. [40] proposed a smart phone method for automated facial expression and speech emotion recognition using SVM with MFCC features extracted from SAVEE database.
Li and Akagi [41] presented a method for recognizing emotions expressed in a multilingual speech using the Fujitsu, Emo-DB, CASIA and SAVEE databases. The highest weighted average precision was obtained after performing speaker normalization and feature selection. The authors in [42] used MFCC related features for recognizing speech emotion based on an improved brain emotional learning (BEL) model inspired by the emotional processing mechanism of the limbic system in human brain. They tested their method on CASIA, SAVEE and FAU Aibo databases using linear discriminant analysis (LDA) and PCA for dimensionality reduction to obtain the highest average accuracy of 90.28% on CASIA database. Mao et al. [43] proposed the emotion-discriminative and domain-invariant feature learning method (EDFLM) for recognizing speech emotion. In their method, both domain divergence and emotion discrimination were considered to learn emotion-discriminative and domain-invariant features using emotion label and domain label constraints. Wang et al. [44] extracted MFCC, Fourier parameters, fundamental frequency, energy and ZCR from three different databases of Emo-DB, Chinese elderly emotion database (EESDB) and CASIA for recognizing speech emotion. Muthusamy et al. [45] extracted a total of 120 wavelet packet energy and entropy features from speech signals and glottal waveforms from Emo-DB, SAVEE and Sahand emotional speech (SES) databases for speech emotion recognition. The extracted features were enhanced using the Gaussian mixture model (GMM) with ELM as the learning algorithm.
Zhu et al. [46] used a combination of acoustic features based on MFCC, pitch, formant, short-term ZCR and short-term energy to recognize speech emotion. They extracted the most discriminating features and performed classification using the deep belief network (DBN) with SVM. Their results showed an average accuracy of 95.8% on the CASIA database across six emotions, which is an improvement when compared to other related studies. Álvarez et al. [47] proposed a classifier subset selection (CSS) for the stacked generalization to recognize speech emotion. They used the estimation of distribution algorithm (EDA) to select optimal features from a collection of features that included eGeMAPS and SVM for classification that achieved an average accuracy of 82.45% on Emo-Db. Bhavan et al. [48] used a combination of MFCCs, spectral centroids and MFCC derivatives of spectral features with a bagged ensemble algorithm based on Gaussian kernel SVM for recognizing speech emotion that achieved an accuracy of 92.45% on Emo-DB. Shegokar et al. [49] proposed the use of continuous wavelet transform (CWT) with prosodic features to recognize speech emotion. They used PCA for feature transformation with quadratic kernel SVM as a classification algorithm that achieved an average accuracy of 60.1% on the RAVDESS database. Kerkeni et al. [50] proposed a model for recognizing speech emotion using empirical mode decomposition (EMD) based on optimal features that included the reconstructed signal based on Mel frequency cepstral coefficient (SMFCC), energy cepstral coefficient (ECC), modulation frequency feature (MFF), modulation spectral (MS) and frequency weighted energy cepstral coefficient (FECC). They achieved an average accuracy of 91.16% on the Spanish database using the RNN algorithm for classification.
Emotion recognition is still a great challenge because of several reasons as previously alluded in the introductory message. Further reasons include the existence of a gap between acoustic features and human emotions [31,36] and the non-existence of a solid theoretical foundation relating the characteristics of voice to the emotions of a speaker [20]. These intrinsic challenges have led to the disagreement in the literature on which features are best for speech emotion recognition [20,36]. The trend in the literature shows that a combination of heterogeneous acoustic features is promising for speech emotion recognition [20,29,30,36,39], but how to effectively unify the different features is highly challenging [21,36]. The importance of selecting relevant features to improve the reliability of speech emotion recognition systems is strongly emphasized in the literature [3,20,29]. Researchers frequently apply specialized software such as OpenEar [31,32,35], OpenSmile [30,32,34,36] and Praat [35] for extraction, selection and unification of speech features from heterogeneous sources to ease the intricacy inherent in the processes. Moreover, the review of numerous literature has revealed that different learning algorithms are trained and validated with the specific features extracted from public databases for speech emotion recognition. Table 1 shows the result of the comparative analysis of our method with the related methods in terms of the experimental database, type of recording, number of emotions, feature set, classification method and maximum percentage accuracy results obtained. It is conspicuous from the comparative analysis that our emotion recognition method is highly promising because it has achieved the highest average accuracy of 99.55% across two dissimilar public experimental databases.

3. Materials and Methods

Materials used for this study included speech emotion multimodal databases of RAVDESS [51] and SAVEE [52]. The databases, because of their popularity were chosen to test for the effectiveness of the proposed HAF against two other sets of features. The study method followed three phases of feature extraction, feature selection and classification, which were subsequently described. The method was heralded by the feature extraction process that involves the abstraction of prosodic and spectral features from the raw audio files of RAVDESS and SAVEE databases. This phase was subsequently followed by the feature selection process that involved the filtration, collection and agglutination of features that have high discriminating power of recognizing emotions in human speech. Classification was the last phase that involved the application of the selected learning algorithm to recognize human emotions and a comparative analysis of experimental results of classification. In the classification phase, several experiments were carried out on a computer with an i7 2.3GHz processor and 8 GB of random access memory (RAM). The purpose of the experiments was to apply standard metrics of accuracy, precision, recall and F1-score to evaluate the effectiveness of the proposed HAF with respect to a given database and ensemble learning algorithm.

3.1. Databases

The speech emotion database is an important precursor for analyzing speech and recognizing emotions. The database provides speech data for training and testing the effectiveness of emotion recognition algorithms [15]. RAVDESS and SAVEE are two public speech emotion databases used in this study for experiments to verify the effectiveness of the proposed HAF for emotion recognition. The two selected experimental databases were fleetingly discussed in this subsection of the paper.

3.1.1. RAVDESS

RAVDESS is a gender balanced set of validated speeches and songs that consists of eight emotions of 24 professional actors speaking similar statements in a North American accent. It is a multiclass database of angry, calm, disgust, fear, happy, neutral, sad and surprise emotions with 1432 American English utterances. Each of the 24 recorded vocal utterances comprises of three formats, which are audio-only (16bit, 48kHz .wav), audio-video (720p H.264, AAC 48kHz, .mp4) and video-only (no sound). The audio-only files were used across all the eight emotions because this study concerns speech emotion recognition. Figure 1 shows that angry, calm, disgust, fear, happy and sad emotion classes constituted 192 audio files each. The surprise emotion had 184 files and the neutral emotion had the lowest number of audio files of 96.

3.1.2. SAVEE

SAVEE is a speech emotion database that consists of recordings from four male actors in seven different emotion classes. The database comprises of a total of 480 British English utterances, which is quite different from the North American accent used in the RAVDESS database. These vocal utterances were processed and labeled in a standard media laboratory with high quality audio-visual equipment. The recorded vocal utterances were classified into seven classes of angry, disgust, fear, happy, neutral, sad and surprise emotional expressions. The neutral emotion constituted 120 audio files, while all the other remaining emotions comprised of 60 audio files each as illustrated in Figure 2.

3.2. Feature Extraction

The speech signal carries a large number of useful information that reflects emotion characteristics such as gender, age, stuttering and identity of the speaker. Feature extraction is an important mechanism in audio processing to capture the information in a speech signal and most of the related studies have emphasized on the extraction of low-level acoustic features for speech recognition. Feature representation plays a prominent role in distinguishing the speech of speakers from each other. Since no universal consensus on the best features for speech recognition, certain authors have mentioned that prosody carries most of the useful information about emotions and can create a suitable feature representation [20]. Moreover, spectral features describe the properties of a speech signal in the frequency domain to identify important aspects of speech and manifest the correlation between vocal movements and changes of channel shape [35,42]. Prosodic and spectral acoustic features were suggested as the most important characteristics of speech because of the improved recognition results of their integration [25,42,53,54]. The blend of prosodic and spectral features is capable of enhancing the performance of emotion recognition, giving better recognition accuracy and a lesser number of emotions in a gender dependent model than many existing systems with the same number of emotions [25]. Banse et al. [54] examined vocal cues for 14 emotions using a combination of prosodic speech features and spectral information in a voiced segment and unvoiced segment to achieve impressive results.
The present authors have applied the selective approach [35] to extract prosodic and spectral features, taking cognizance of features that could be useful for improving emotion recognition. These are features that have produced positive results in the related studies and those from the other comparable tasks. We carefully applied the specialized jAudio software [55] to construct MFCC1, MFCC2 and HAF as three sets of features representing several voice aspects. MFCC1 is the set of MFCC features inspired by the applications of MFCC features [24,40,42]. MFCC2 is the set of features based on MFCC, energy, ZCR and fundamental frequency as inspired by the fusion of MFCC with other acoustic features [11,25,26,44,46]. The HAF is the proposed set of hybrid acoustic features of prosodic and spectral carefully selected based on the interesting results in the literature [11,20,24,25,26,29,30,39,40,42,44,46,48].
The feature extraction process was aided by the application of a specialized jAudio to effectively manage the inherent complexity in the selection process. The jAudio software is a digital signal processing library of advanced audio feature extraction algorithms designed to eliminate effort duplication in manually calculating emotion features from an audio signal. It provides a unique method of handling multidimensional features, dependency and allows for iterative development of new features from the existing ones through the meta-feature template. Moreover, it simplifies the intrinsic stiffness of the existing feature extraction methods by allowing low-level features to be fused to build increasingly high-level musically meaningful features. In addition, it provides audio samples as simple vectors of features and offers aggregation functions that aggregate a sequence of separate feature sets into a single feature vector.

3.2.1. Prosodic Features

Prosodic features are acoustic features prominently used in emotion recognition and speech signal processing because they carry essential paralinguistic information. This information complements a message with an intention that can paint a flawless picture about attitude or emotion [35,56]. In addition, prosodic features are considered as suprasegmental information because they help in defining and structuring the flow of speech [35]. Prosodic continuous speech features such as pitch and energy convey much content of emotions in speech [57] and are important for delivering the emotional cues of the speakers. These features include formant, timing and articulation features and they characterize the perceptual properties of speech typically used by human beings to perform different speech tasks [57].
The present authors have included three important prosodic features of energy, fundamental frequency and ZCR. Signal energy models the voice intensity, volume or loudness and reflects the pause and ascent of the voice signal. It is often associated with the human respiratory system and is one of the most important characteristics of human aural perception. The logarithm function is often used to reflect minor changes of energy because the energy of an audio signal is influenced by the recording conditions. Fundamental frequency provides tonal plus rhythmic characteristics of a speech and carries useful information about the speaker. ZCR determines the information about the number of times a signal waveform crosses the zero amplitude line because of a transition from a positive/negative value to a negative/positive value in a given time. It is suitable for detecting voice activity, end point, voiced sound segment, unvoiced sound segment, silent segment and approximating the measure of noisiness in speech [58]. ZCR is an acoustic feature that has been classified as a prosodic feature [49,59,60]. In particular, energy and pitch were declared prosody features with a low frequency domain while ZCR and formants are high frequency features [60].

3.2.2. Spectral Features

Spectral features of this study include timbral features that have been successful in music recognition [61]. Timbral features define the quality of a sound [62] and they are a complete opposite of most general features like pitch and intensity. It has been revealed that a strong relationship exists between voice quality and emotional content in a speech [58]. Siedenburg et al. [63] considered spectral features as significant in distinguishing between classes of speech and music. They claimed that the temporal evolution of the spectrum of audio signals mainly accounts for the timbral perception. Timbral features of a sound help in differentiating between sounds that have the same pitch and loudness and they assist in the classification of audio samples with similar timbral features into unique classes [61]. The application of spectral timbral features like spectral centroid, spectral roll-off point, spectral flux, time domain zero crossings and root mean squared energy amongst others has been demonstrated for speech analysis [64].
The authors have included eight spectral features of MFCC, spectral roll-off point, spectral flux, spectral centroid, spectral compactness, fast Fourier transforms, spectral variability and LPCC. MFCC is one of the most widely used feature extraction methods for speech analysis because of its computational simplicity, superior ability of distinction and high robustness to noise [42]. It has been successfully applied to discriminate phonemes and can determine what is said and how it is said [35]. The method is based on the human hearing system that provides a natural and real reference for speech recognition [46]. In addition, it is based on the sense that human ears perceive sound waves of different frequencies in a non-linear mode [42]. Spectral roll-off point defines a frequency below which at least 85% of the total spectral energy are present and provides a measure of spectral shape [65]. Spectral flux characterizes the dynamic variation of spectral information, is related to perception of music rhythm and it captures spectrum difference between two adjacent frames [66].
Spectral centroid describes the center of gravity of the magnitude spectrum of short-time Fourier transform (STFT). It is a spectral moment that is helpful in modeling sharpness or brightness of sound [67]. Spectral compactness is obtained by comparing the components in the magnitude spectrum of a frame and magnitude spectrum of neighboring frames [68]. Fast Fourier transform (FFT) is a useful method of analyzing the frequency spectrum of a speech signal and features based on the FTT algorithm have the strongest frequency component in Hertz [69,70]. Spectral variability is a measure of variance of the magnitude spectrum of a signal and is attributed to different sources, including phonetic content, speaker, channel, coarticulation and context [71]. Spectral compactness is an audio feature that measures the noisiness of a speech signal and is closely related to the spectral smoothness of speech signal [72]. LPCCs are spectral features obtained from the envelope of LPC and are computed from sample points of a speech waveform. They have low vulnerability to noise, yielded a lower error rate in comparison to LPC features and can discriminate between different emotions [73]. LPCC features are robust, but they are not based on an auditory perceptual frequency scale like MFCC [74].

3.3. Feature Selection

Feature selection is a process of filtering insignificant features to create a subset of the most discriminating features in the input data. It reduces the cause of dimensionality and improves the success of a recognition algorithm. Based on the literature review, we observed that most methods used for feature selection are computationally expensive because of the enormous processing time involved. Moreover, accuracy and precision values of some of these methods are reduced because of redundancy, less discriminating features and poor accuracy values in certain emotions such as the fear emotion. The authors have selected 133 MFCC spectral features for the MFCC1 and 90 features based on MFCC, ZCR, energy and fundamental frequency for the MFCC2. In a nutshell, the effectiveness of the HAF was tested against one homogenous set of acoustic features (MFCC1) and another hybrid set of acoustic features (MFCC2).
Table 2 presents the list of prosodic and spectral features of this study in terms of group, type and number of features with their statistical characteristics obtained with the aid of jAudio that generated a csv file with numerical values for each feature together with the respective emotion class. The literature inspired brute force approach was then explored to select non-redundant features that are likely to help improve emotion recognition performance. This implies inspecting through a series of experiments for possible combinations of features to discover the right features that define HAF. This approach is although laborious to require the help of a stochastic optimization technique, it has yielded excellent performance results in this study.

3.4. Classification

Emotion classification is aimed at obtaining an emotional state for the input feature vector using a machine learning algorithm. In this study, classification experiments were conducted using ensemble learning algorithms to verify the effectiveness of the proposed HAF. Ensemble learning is considered a state-of-the-art approach for solving different machine learning problems by combining the predictions of multiple base learners [75]. It improves the overall predictive performance, decreases the risk of obtaining a local minimum and provides a better fit to the data space by combining the predictions of several weak learners into a strong learning algorithm. In this study, we applied bagging and boosting machine learning because they are widely used effective approaches for constructing ensemble learning algorithms [76,77,78,79]. Bagging is a technique that utilizes bootstrap sampling to reduce the variance of a decision tree and improve the accuracy of a learning algorithm by creating a collection of learning algorithms that are learned in parallel [79,80,81]. It performs random sampling with replacement over a simple averaging of all the predictions from different decision trees to give a more robust result than a single decision tree. Boosting creates a collection of learning algorithms that are learned sequentially with early learners to fit decision trees to the data and analyze data for errors. It performs random sampling with replacement over a weighted average and reduces classification bias and variance of a decision tree [79]. One of the bagging ensemble algorithms investigated in this study was random decision forest (RDF), which is an extension over bagging that is popularly called random forest [75,82]. RDF can be constituted by making use of bagging based on the CART approach to raise trees [83]. The other bagging ensemble learning algorithms investigated in this study were Bagging with SVM as the base learner (BSVM) and bagging with the multilayer perceptron neural network as the base learner (BMLP). The boosting ensemble algorithms investigated in this study were the gradient boosting machine (GBM), which extends boosting by combining the gradient descent optimization algorithm with boosting technique [75,82,84], and AdaBoost with CART as the base learner (ABC), which is one of the most widely used boosting algorithm to reduce sensitivity to class label noise [79,85]. AdaBoost is an iterative learning algorithm for constructing a strong classifier by enhancing weak classification algorithms and it can improve data classification ability by reducing both bias and variance through continuous learning [81].
The standard metrics of accuracy, precision, recall and F1-score have been engaged to measure the performances of the learning algorithms with respect to a particular set of features. The accuracy of a classification algorithm is often judged as one of the most intuitive performance measures. It is the ratio of the number of instances correctly recognized to the total number of instances. Precision is the ratio of the number of positive instances correctly recognized to the total number of positive instances. Recall is the ratio of the number of positive instances correctly recognized to the number of all the instances of the actual class [86]. F1-score constitutes a harmonic mean of precision and recall [87]. Moreover, the training times of the learning algorithms are also compared to analyze the computational complexities of training different ensemble learning classifiers. The flow chart for different configurations of the methods of experimentation is illustrated in Figure 3.

4. Results and Discussion

The present authors have extracted the set of 404 HAFs from RAVDESS and SAVEE experimental databases. The features were used to train five famous ensemble learning algorithms in order to determine their effectiveness using four standard performance metrics. The performance metrics were selected because the experimental databases were not equally distributed. Data were subjected to 10 fold cross-validation and divided into training and testing groups as the norm permits. Gaussian noise was added to the training data to provide a regularizing effect, reduce overfitting, increase resilience and improve generalization performance of the ensemble learning algorithms.
Table 3 shows the result of the CPU computational time of training an individual ensemble learning algorithm with three different sets of features across two experimental databases. The training time analysis shows that RDF spent the least amount of training time when compared to other ensemble learning algorithms. The least training time was recorded with respect to MFCC1 and MFCC2 irrespective of databases. However, the training time of RDF (0.43) was slightly higher than that of BSVM (0.39) with respect to SAVEE HAF, but it claimed superiority with respect to the RAVDESS HAF. The reason for this inconsistent result might be because of the smaller data size (480) of SAVEE when compared to the size (1432) of RAVDESS. The experimental results of a study that was aimed at fully assessing the predictive performances of SVM and SVM ensembles over small and large scale datasets of breast cancer show that SVM ensemble based on BSVM method can be the better choice for a small scale dataset [78]. Moreover, it can be observed that GBM and BMLP were generally time consuming when compared to the other ensemble learning algorithms because they took 1172.3 milliseconds to fit MFCC2 RAVDESS and 1319.39 milliseconds to fit MFCC2 SAVEE features respectively.
The experimental results of the overall recognition performance obtained are given in Table 4. The percentage average recognition performance rates differ depending on critical factors such as the type of database, the accent of the speakers, the way data were collected, the type of features and the learning algorithm used. In particular, RDF consistently recorded the highest average accuracy, while BMLP consistently recorded the lowest average accuracy across the sets of features, irrespective of databases. These results indicate that RDF gave the highest performance because the computed values of the precision, recall, F1-score and accuracy of the ensemble learning algorithm are generally impressive when compared to other ensemble learning algorithms. However, average recognition accuracies of 61.5% and 66.7% obtained by RDF for the SAVEE MFCC1 and MFCC2 were respectively low, but the values were better than that of BMLP (55.4% and 60.7%) for SAVEE MFCC1 and MFCC2 respectively. In general, an improvement can be observed when a combination of features (MFCC2) was used in an experiment drawing inspiration from the work done in [46], where MFCC features were combined with pitch and ZCR features. In addition, we borrowed the same line of thought from Sarker and Alam [3] and Bhaskar et al. [28] where some of the above mentioned features were combined to form hybrid features.
The results in Table 4 can be used to demonstrate the effectiveness of the proposed HAF. The values of the precision, recall, F1-score and accuracy performance measures can be seen to be generally high for the five ensemble learning algorithms using the HAF irrespective of databases. Specifically, RDF and GBM concomitantly recorded the highest accuracy of 99.3% for the SAVEE HAF, but RDF claims superiority with 99.8% accuracy when compared to the 92.6% accuracy for the HAF RAVDESS. Consequently, the results of RDF learning of the proposed HAF were exceptionally inspiring with an average accuracy of 99.55% across the two databases. The ranking of the investigated ensemble learning algorithms in terms of percentage average accuracy computed for HAF across the databases is RDF (99.55%), GBM (95.95%), ABC (95.00%), BSVM (93.90%) and BMLP (92.95%). It can be inferred that application of ensemble learning of HAF is highly promising for recognizing emotions in human speech. In general, high accuracy values computed by all learning algorithms across databases indicate the effectiveness of the proposed HAF for speech emotion recognition. Moreover, the confidence factors show that the true classification accuracies of all the ensemble learning algorithms investigated lie in the range between 0.002% and 0.045% across all the corpora. The results in Table 4 show that HAF significantly improved the performance of emotion recognition when compared to the methods of the SAVEE MFCC related features [42] and RAVDESS wavelet transform features [49] that recorded the highest recognition accuracies of 76.40% and 60.10% respectively.
In particular, the results of precision, recall, F1-score and accuracy per each emotion, learning algorithm and database were subsequently discussed. Table 5 shows that angry, calm and disgust emotions were easier to recognize using RDF and GBM ensemble learning of RAVDESS MFCC1 features. The results were particularly impressive with RDF ensemble learning with perfect precision, recall and F1-score computed by the algorithm, which also recorded highest accuracy values across all emotions. However, BMLP generally gave the least performing result across all emotions. RAVDESS MFCC1 features achieved a precision rate of 92.0% for the fear emotion using RDF ensemble learning, which is relatively high when compared to the results in [27].
Table 6 shows the results of percentage precision, recall, F1-score and accuracy performance analysis of RAVDESS MFCC2 features. The performance values were generally higher for recognizing angry, calm and disgust emotions using RDF. However, the performance values were relatively low for happy and neutral emotions showing that the hybridization of MFCC, ZCR, energy and fundamental frequency features yielded a relatively poor result in recognizing happy and neutral emotions using ensemble learning. The MFCC2 combination of acoustic features achieved an improved recognition performance of the fear (90.0%) emotion using RDF in comparison with the report in [44], where 81.0% recognition was achieved for the fear emotion using the CASIA database.
Emotion recognition results shown in Table 7 with respect to the RAVDESS HAF were generally impressive across emotions, irrespective of the ensemble learning algorithms utilized. RDF was highly effective in recognizing all the eight emotions because it had achieved perfect accuracy value by 100.0% on almost all the emotions, except fear (98.0%). Moreover, perfect precision and accuracy values were achieved in recognizing the neutral emotion, which was really difficult to identify using other sets of features. In addition, 98.0% accuracy and perfect values obtained for precision, recall and F1-score for the fear emotion is impressive because of the difficulty of recognizing the fear emotion as reported in the literature [25,26,27]. All the other learning algorithms performed relatively well, considering the fact that noise was added to the training data. These results attest to the effectiveness of the proposed HAF for recognizing emotions in human speech. In addition, the results point to the significance of sourcing highly discriminating features for recognizing human emotions. They further show that the fear emotion could be recognized efficiently when compared to other techniques used in [44]. It can be seen that the recognition of fear emotion was successfully improved using the proposed HAF. In addition, HAF gave better performance than the bottleneck features used for six emotions from CASIA database [27] that achieved a recognition rate of 60.5% for the fear emotion.
Table 8 shows the results of percentage precision, recall, F1-score and accuracy analysis of SAVEE MFCC1 features. In the results, the maximum percentage precision F1-score of the neutral emotion had achieved 94.0% using the RDF ensemble algorithm. This F1-score was quite high when compared to the F1-score obtained from RAVDESS database using the same features. The percentage increase in overall precision of neutral emotion was observed to be 133.33%, 77.36%, 18.57%, 12.66% and 10.39% using GBM, RDF, BMLP, ABC and BSVM respectively. Moreover, RDF and GBM algorithms achieved perfect prediction (100.0%) for recognizing the surprise emotion. The results show that MFCC features still achieved low recognition rates for the fear emotion as reported by other authors [25]. However, RDF (94.0%) and GBM (91.0%) show higher precision rates for the neutral emotion, which is higher than what was reported in [11].
Table 9 shows the analysis results of the percentage precision, recall, F1-score and accuracy obtained after performing the classification with SAVEE MFCC2 features. It can be observed from the table that all the learning algorithms performed well for recognizing the surprise emotion, but performed poorly in recognizing other emotions. In general, all learning algorithms achieved a low prediction of emotions with the exception of surprise emotion in which the perfect F1-score was obtained using the RDF algorithm. The lowest precision value of 93.0% was achieved for the surprise emotion using MFCC2 features with BMLP. These results indicate the unsuitability of MFCC, ZCR, energy and fundamental frequency features for recognizing human emotions. However, RDF and GBM gave higher recognition results for happy and sad emotions, RDF gave a higher recognition result for fear emotion while ABC and BSVM gave higher recognition results for the sad emotion that the traditional and bottleneck features [27].
Table 10 shows the analysis results of the percentage precision, recall, F1-score and accuracy obtained after performing classification with SAVEE HAF. Accordingly, all the emotions were much easier to recognize using the HAF that had tremendously improved the recognition of the fear emotion reported in the literature to be difficult with lower performance results recorded with traditional and bottleneck features for fear emotion [27]. Moreover, Kerkeni et al. [11] obtained a recognition accuracy rate of 76.16% for the fear emotion using MFCC and MS features based on the Spanish database of seven emotions. These results show that using HAF with ensemble learning is highly promising for recognizing emotions in human speech. The study findings were consistent with the literature that ensemble learning gives a better predictive performance through the fusion of information knowledge of predictions of multiple inducers [75,85]. The ability of ensemble learning algorithms to mimic the nature of human by seeking opinions from several inducers for informed decision distinguish them from inducers [73]. Moreover, it can be seen across Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10 that the set of HAF presented the most effective acoustic features for speech emotion recognition, while the set MFCC1 presented the worst features. In addition, the study results supported the hypothesis that a combination of acoustic features based on prosodic and spectral reflects the important characteristics of speech [42,53,54].
Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10 show that HAF features had a higher discriminating power because higher performance results were achieved across all emotion classes in the SAVEE and RAVDESS corpora. The performance of each ensemble learning algorithm was generally low when MFCC1 and MFCC2 were applied because the sets of the features had a much lower discriminating power compared to HAF. The recognition results across the experimental databases were different because RAVDESS had more instances than SAVEE. Moreover, SAVEE constitutes male speakers only while RAVDESS is comprised of both male and female speakers. According to the literature, this diversity has an impact on the performance of emotion recognition [35]. The literature review revealed that most emotion recognition models have not been able to recognize the fear emotion with a higher classification accuracy of 98%, but the results in Table 4 show that HAF significantly improved the recognition performance of emotion beyond expectation.

5. Conclusions

The automatic recognition of emotion is still an open research because human emotions are highly influenced by quite a number of external factors. The primary contribution of this study is the construction and validation of a set of hybrid acoustic features based on prosodic and spectral features for improving speech emotion recognition. The constructed acoustic features have made it easy to effectively recognize eight classes of human emotions as seen in this paper. The agglutination of prosodic and spectral features has been demonstrated in this study to yield excellent classification results. The proposed set of acoustic features is highly effective in recognizing all the eight emotions investigated in this study, including the fear emotion that has been reported in the literature to be difficult to classify. The proposed methodology seems to work well because combining certain prosodic and spectral features increases the overall discriminating power of the features hence increasing classification accuracy. This superiority is further underlined by the fact that the same ensemble learning algorithms were used on different feature sets, exposing the performance of each group of features regarding their discriminating power. In addition, we saw that ensemble learning algorithms generally performed well. They are widely judged to improve recognition performance better than a single inducer by decreasing variability and reducing bias. Experimental results show that it was quite difficult to achieve high precisions and accuracies when recognizing the neutral emotion with either pure MFCC features or a combination of MFCC, ZCR, energy and fundamental frequency features that were seen to be effective in recognizing the surprise emotion. However, we provided evidence through intensive experimentation that random decision forest ensemble learning of the proposed hybrid acoustic features was highly effective for speech emotion recognition.
The results of this study were generally fantastic, even though the Gaussian noise was added to the training data. The gradient boosting machine algorithm yielded good results, but the main challenge with the algorithm is that its training time is long. In terms of effectiveness, the random decision forest algorithm is superior to the gradient boosting machine algorithm for speech emotion recognition with the proposed acoustic features. The results obtained in this study were quite stimulating, nevertheless, the main limitation of this study was that the experiments were done on acted speech databases in consonance with the research culture. There can be major differences between working with acted and real data. Moreover, there are new features such as breath and spectrogram not investigated in this study for recognizing speech emotion. In the future, we would like to pursue this direction to determine how the new features will compare with the proposed hybrid acoustic features. In addition, future research will harvest real emotion data for emotion recognition using the Huawei OceanConnect cloud based IoT platform and Narrow Band IoT resources in South Africa Luban workshop at the institution of the authors, which is a partnership project with the Tianjin Vocational Institute in China.

Author Contributions

Conceptualization by K.Z. and O.O., Data curation by K.Z. Methodology by K.Z. Writing original draft by K.Z. and O.O. Writing review and editing by O.O., Supervision by O.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, Y.; Hao, Y.; Chen, M.; Chen, J.; Lu, P.; Košir, A. Audio-visual emotion fusion (AVEF): A deep efficient weighted approach. Inf. Fusion 2019, 46, 184–192. [Google Scholar] [CrossRef]
  2. Picard, R.W. Affective computing: Challenges. Int. J. Hum. Comput. Stud. 2003, 59, 55–64. [Google Scholar] [CrossRef]
  3. Sarker, K.; Alam, K.R. Emotion recognition from human speech: Emphasizing on relevant feature selection and majority voting technique. In Proceedings of the 3rd International Conference on Informatics, Electronics & Vision (ICIEV), Dhaka, Bangladesh, 23–24 May 2014; pp. 89–95. [Google Scholar]
  4. Li, S.; Xu, L.; Yang, Z. Multidimensional speaker information recognition based on proposed baseline system. In Proceedings of the 2017 IEEE 2nd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 25–26 March 2017; pp. 1776–1780. [Google Scholar]
  5. Jiang, H.; Hu, B.; Liu, Z.; Yan, L.; Wang, T.; Liu, F.; Kang, H.; Li, X. Investigation of different speech types and emotions for detecting depression using different classifiers. Speech Commun. 2017, 90, 39–46. [Google Scholar] [CrossRef]
  6. Subhashini, R.; Niveditha, P.R. Analyzing and detecting employee’s emotion for amelioration of organizations. Procedia Comput. Sci. 2015, 48, 530–536. [Google Scholar] [CrossRef] [Green Version]
  7. Feinberg, R.A.; Hokama, L.; Kadam, R.; Kim, I. Operational determinants of caller satisfaction in the banking/financial services call center. Int. J. Bank Mark. 2002, 20, 174–180. [Google Scholar] [CrossRef]
  8. Gomes, J.; El-Sharkawy, M. Implementation of i-vector algorithm in speech emotion recognition by using two different classifiers: Gaussian mixture model and support vector machine. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2016, 6, 8–16. [Google Scholar]
  9. Yu, T.-K.; Fang, K. Measuring the post-adoption customer perception of mobile banking services. Cyberpsychol. Behav. 2009, 12, 33–35. [Google Scholar] [CrossRef]
  10. Chakraborty, R.; Pandharipande, M.; Kopparapu, S.K. Knowledge-based framework for intelligent emotion recognition in spontaneous speech. Procedia Comput. Sci. 2016, 96, 587–596. [Google Scholar] [CrossRef] [Green Version]
  11. Kerkeni, L.; Serrestou, Y.; Mbarki, M.; Raoof, K.; Mahjoub, M.A. Speech emotion recognition: Methods and cases study. In Proceedings of the 10th International Conference on Agents and Artificial Intelligence (ICAART 2018), Funchal, Madeira, Portugal, 16–18 January 2018; Volume 2, pp. 175–182. [Google Scholar]
  12. Arias, J.P.; Busso, C.; Yoma, N.B. Shape-based modeling of the fundamental frequency contour for emotion detection in speech. Comput. Speech Lang. 2014, 28, 278–294. [Google Scholar] [CrossRef]
  13. Atayero, A.A.; Olugbara, O.O.; Ayo, C.K.; Ikhu-Omoregbe, N.A. Design, development and deployment of an automated speech-controlled customer care service system. In Proceedings of the GSPx, The International Embedded Solutions Event, Santa Clara, CA, USA, 27–30 September 2004; pp. 27–30. [Google Scholar]
  14. Hess, U.; Thibault, P. Darwin and emotion expression. Am. Psychol. 2009, 64, 120–128. [Google Scholar] [CrossRef]
  15. Ying, S.; Xue-Ying, Z. Characteristics of human auditory model based on compensation of glottal features in speech emotion recognition. Future Gener. Comput. Syst. 2018, 81, 291–296. [Google Scholar] [CrossRef]
  16. Xiaoqing, J.; Kewen, X.; Yongliang, L.; Jianchuan, B. Noisy speech emotion recognition using sample reconstruction and multiple-kernel learning. J. China Univ. Posts Telecommun. 2017, 24, 17–19. [Google Scholar] [CrossRef]
  17. Papakostas, M.; Spyrou, E.; Giannakopoulos, T.; Siantikos, G.; Sgouropoulos, D.; Mylonas, P.; Makedon, F. Deep visual attributes vs. hand-crafted audio features on multidomain speech emotion recognition. Computation 2017, 5, 26. [Google Scholar] [CrossRef] [Green Version]
  18. Arruti, A.; Cearreta, I.; Álvarez, A.; Lazkano, E.; Sierra, B. Feature selection for speech emotion recognition in Spanish and Basque: On the use of machine learning to improve human-computer interaction. PLoS ONE 2014, 9, e108975. [Google Scholar] [CrossRef] [Green Version]
  19. Altun, H.; Polat, G. Boosting selection of speech related features to improve performance of multi-class SVMs in emotion detection. Expert Syst. Appl. 2009, 36, 8197–8203. [Google Scholar] [CrossRef]
  20. Luengo, I.; Navas, E.; Hernaez, I. Feature analysis and evaluation for automatic emotion identification in speech. IEEE Trans. Multimed. 2010, 12, 490–501. [Google Scholar] [CrossRef]
  21. Basu, S.; Chakraborty, J.; Bag, A.; Aftabuddin, M. A review on emotion recognition using speech. In Proceedings of the 2017 International Conference on Inventive Communication and Computational Technologies (ICICCT), Coimbatore, India, 10–11 March 2017; pp. 109–114. [Google Scholar]
  22. Cong, P.; Wang, C.; Ren, Z.; Wang, H.; Wang, Y.; Feng, J. Unsatisfied customer call detection with deep learning. In Proceedings of the 2016 10th International Symposium on Chinese Spoken Language Processing (ISCSLP), Tianjin, China, 17–20 October 2016; pp. 1–5. [Google Scholar]
  23. Getahun, F.; Kebede, M. Emotion identification from spontaneous communication. In Proceedings of the 2016 12th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Naples, Italy, 28 November–1 December 2016; pp. 151–158. [Google Scholar]
  24. Palo, H.K.; Chandra, M.; Mohanty, M.N. Emotion recognition using MLP and GMM for Oriya language. Int. J. Comput. Vis. Robot. 2017, 7, 426–442. [Google Scholar] [CrossRef]
  25. Khan, A.; Roy, U.K. Emotion recognition using prosodie and spectral features of speech and Naïve Bayes classifier. In Proceedings of the 2017 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), Chennai, India, 22–24 March 2017; pp. 1017–1021. [Google Scholar]
  26. Semwal, N.; Kumar, A.; Narayanan, S. Automatic speech emotion detection system using multi-domain acoustic feature selection and classification models. In Proceedings of the 2017 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA), New Delhi, India, 22–24 February 2017; pp. 1–6. [Google Scholar]
  27. Sun, L.; Zou, B.; Fu, S.; Chen, J.; Wang, F. Speech emotion recognition based on DNN-decision tree SVM model. Speech Commun. 2019, 115, 29–37. [Google Scholar] [CrossRef]
  28. Bhaskar, J.; Sruthi, K.; Nedungadi, P. Hybrid approach for emotion classification of audio conversation based on text and speech mining. Procedia Comput. Sci. 2015, 46, 635–643. [Google Scholar] [CrossRef] [Green Version]
  29. Liu, Z.T.; Wu, M.; Cao, W.H.; Mao, J.W.; Xu, J.P.; Tan, G.Z. Speech emotion recognition based on feature selection and extreme learning machine decision tree. Neurocomputing 2018, 273, 271–280. [Google Scholar] [CrossRef]
  30. Cao, H.; Verma, R.; Nenkova, A. Speaker-sensitive emotion recognition via ranking: Studies on acted and spontaneous speech. Comput. Speech Lang. 2015, 29, 186–202. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Stuhlsatz, A.; Meyer, C.; Eyben, F.; Zielke, T.; Meier, G.; Schuller, B. Deep neural networks for acoustic emotion recognition: Raising the benchmarks. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5688–5691. [Google Scholar]
  32. Zhang, Z.; Coutinho, E.; Deng, J.; Schuller, B. Cooperative learning and its application to emotion recognition from speech. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 115–126. [Google Scholar] [CrossRef] [Green Version]
  33. Sun, Y.; Wen, G.; Wang, J. Weighted spectral features based on local Hu moments for speech emotion recognition. Biomed. Signal Process. Control 2015, 18, 80–90. [Google Scholar] [CrossRef]
  34. Shaqra, F.A.; Duwairi, R.; Al-Ayyoub, M. Recognizing emotion from speech based on age and gender using hierarchical models. Procedia Comput. Sci. 2019, 151, 37–44. [Google Scholar] [CrossRef]
  35. Pérez-Espinosa, H.; Reyes-García, C.A.; Villaseñor-Pineda, L. Acoustic feature selection and classification of emotions in speech using a 3D continuous emotion model. Biomed. Signal Process. Control 2012, 7, 79–87. [Google Scholar] [CrossRef]
  36. Jiang, W.; Wang, Z.; Jin, J.S.; Han, X.; Li, C. Speech emotion recognition with heterogeneous feature unification of deep neural network. Sensors 2019, 19, 2730. [Google Scholar] [CrossRef] [Green Version]
  37. Prasada Rao, K.; Chandra Sekhara Rao, M.V.P.; Hemanth Chowdary, N. An integrated approach to emotion recognition and gender classification. J. Vis. Commun. Image Represent. 2019, 60, 339–345. [Google Scholar] [CrossRef]
  38. Narendra, N.P.; Alku, P. Dysarthric speech classification from coded telephone speech using glottal features. Speech Commun. 2019, 110, 47–55. [Google Scholar] [CrossRef]
  39. Alonso, J.B.; Cabrera, J.; Medina, M.; Travieso, C.M. New approach in quantification of emotional intensity from the speech signal: Emotional temperature. Expert Syst. Appl. 2015, 42, 9554–9564. [Google Scholar] [CrossRef]
  40. Alshamsi, H.; Kepuska, V.; Alshamsi, H.; Meng, H. Automated facial expression and speech emotion recognition app development on smart phones using cloud computing. In Proceedings of the 2018 IEEE 9th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON), Vancouver, BC, Canada, 1–3 November 2018; pp. 730–738. [Google Scholar]
  41. Li, X.; Akagi, M. Improving multilingual speech emotion recognition by combining acoustic features in a three-layer model. Speech Commun. 2019, 110, 1–12. [Google Scholar] [CrossRef]
  42. Liu, Z.T.; Xie, Q.; Wu, M.; Cao, W.H.; Mei, Y.; Mao, J.W. Speech emotion recognition based on an improved brain emotion learning model. Neurocomputing 2018, 309, 145–156. [Google Scholar] [CrossRef]
  43. Mao, Q.; Xu, G.; Xue, W.; Gou, J.; Zhan, Y. Learning emotion-discriminative and domain-invariant features for domain adaptation in speech emotion recognition. Speech Commun. 2017, 93, 1–10. [Google Scholar] [CrossRef]
  44. Wang, K.; An, N.; Li, B.N.; Zhang, Y.; Li, L. Speech emotion recognition using Fourier parameters. IEEE Trans. Affect. Comput. 2015, 6, 69–75. [Google Scholar] [CrossRef]
  45. Muthusamy, H.; Polat, K.; Yaacob, S. Improved emotion recognition using gaussian mixture model and extreme learning machine in speech and glottal signals. Math. Probl. Eng. 2015, 2015, 13. [Google Scholar] [CrossRef]
  46. Zhu, L.; Chen, L.; Zhao, D.; Zhou, J.; Zhang, W. Emotion recognition from chinese speech for smart affective services using a combination of SVM and DBN. Sensors 2017, 17, 1694. [Google Scholar] [CrossRef] [Green Version]
  47. Álvarez, A.; Sierra, B.; Arruti, A.; Lópezgil, J.M.; Garay-Vitoria, N. Classifier subset selection for the stacked generalization method applied to emotion recognition in speech. Sensors 2016, 16, 21. [Google Scholar] [CrossRef] [Green Version]
  48. Bhavan, A.; Chauhan, P.; Shah, R.R. Bagged support vector machines for emotion recognition from speech. Knowl. Based Syst. 2019, 184, 104886. [Google Scholar] [CrossRef]
  49. Shegokar, P.; Sircar, P. Continuous wavelet transform based speech emotion recognition. In Proceedings of the 2016 10th International Conference on Signal Processing and Communication Systems (ICSPCS), Gold Coast, QLD, Australia, 19–21 December 2016; pp. 1–8. [Google Scholar]
  50. Kerkeni, L.; Serrestou, Y.; Raoof, K.; Mbarki, M.; Mahjoub, M.A.; Cleder, C. Automatic speech emotion recognition using an optimal combination of features based on EMD-TKEO. Speech Commun. 2019, 114, 22–35. [Google Scholar] [CrossRef]
  51. Livingstone, S.R.; Russo, F.A. The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north American english. PLoS ONE 2018, 13, 1–38. [Google Scholar] [CrossRef] [Green Version]
  52. Wang, W.; Klinger, K.; Conapitski, C.; Gundrum, T.; Snavely, J. Machine Audition: Principles, Algorithms, and Systems; IGI Global Press: Hershey, PA, USA, 2010. [Google Scholar]
  53. Ibrahim, N.J.; Idris, M.Y.I.; Yakub, M.; Yusoff, Z.M.; Rahman, N.N.A.; Dien, M.I. Robust feature extraction based on spectral and prosodic features for classical Arabic accents recognition. Malaysian J.Comput. Sci. 2019, 46–72. [Google Scholar]
  54. Banse, R.; Scherer, K.R. Acoustic profiles in vocal emotion expression. J. Pers. Soc. Psychol. 1996, 70, 614–636. [Google Scholar] [CrossRef] [PubMed]
  55. McEnnis, D.; McKay, C.; Fujinaga, I.; Depalle, P. jAudio: A feature extraction library. In Proceedings of the International Conference on Music Information Retrieval, London, UK, 11–15 September 2005; pp. 600–603. [Google Scholar]
  56. Hellbernd, N.; Sammler, D. Prosody conveys speaker’s intentions: Acoustic cues for speech act perception. J. Mem. Lang. 2016, 88, 70–86. [Google Scholar] [CrossRef] [Green Version]
  57. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  58. Guidi, A.; Gentili, C.; Scilingo, E.P.; Vanello, N. Analysis of speech features and personality traits. Biomed. Signal Process. Control. 2019, 51, 1–7. [Google Scholar] [CrossRef]
  59. Pervaiz, M.; Ahmed, T. Emotion recognition from speech using prosodic and linguistic features. Int. J. Adv. Comput. Sci. Appl. 2016, 7, 84–90. [Google Scholar] [CrossRef] [Green Version]
  60. Chen, L.; Mao, X.; Xue, Y.; Cheng, L.L. Speech emotion recognition: Features and classification models. Digit. Signal Process. 2012, 22, 1154–1160. [Google Scholar] [CrossRef]
  61. Agostini, G.; Longari, M.; Pollastri, E. Musical instrument timbres classification with spectral features. EURASIP J. Appl. Signal Process. 2003, 2003, 5–14. [Google Scholar] [CrossRef] [Green Version]
  62. Avisado, H.G.; Cocjin, J.V.; Gaverza, J.A.; Cabredo, R.; Cu, J.; Suarez, M. Analysis of music timbre features for the construction of user-specific affect model. Theory Pract. Comput. 2012, 5, 28–35. [Google Scholar]
  63. Siedenburg, K.; Fujinaga, I.; McAdams, S. A Comparison of approaches to timbre descriptors in music information retrieval and music psychology. J. New Music Res. 2016, 45, 27–41. [Google Scholar] [CrossRef] [Green Version]
  64. Istening, E.M.L.; Imbre, T. Embodied listening and timbre: Perceptual, acoustical and neural correlates. Music Percept. 2018, 35, 332–363. [Google Scholar]
  65. Kos, M.; Kačič, Z.; Vlaj, D. Acoustic classification and segmentation using modified spectral roll-off and variance-based features. Digit. Signal Process. Rev. J. 2013, 23, 659–674. [Google Scholar] [CrossRef]
  66. Burger, B.; Ahokas, R.; Keipi, A.; Toiviainen, P. Relationships between spectral flux, perceived rhythmic strength, and the propensity to move. In Proceedings of the Sound and Music Computing Conference 2013, SMC 2013, Stockholm, Sweden, 30 July–3 August 2013; pp. 179–184. [Google Scholar]
  67. Rouillard, J.M.; Simler, M. Signal estimation from modified short-time Fourier transform. Trans. Acoust. Speech Signal Process. 1952, 3, 772–776. [Google Scholar]
  68. Razuri, J.D.; Sundgren, G.; Rahmani, R.; Larsson, A.; Moran, A.; Bonet, I. Speech emotion recognition in emotional feedback for Human-Robot Interaction. Int. J. Adv. Res. Artif. Intell. 2015, 4, 20–27. [Google Scholar]
  69. Ernawan, F.; Abu, N.A.; Suryana, N. Spectrum analysis of speech recognition via discrete Tchebichef transform. In Proceedings of the International Conference on Graphic and Image Processing (ICGIP 2011), Cairo, Egypt, 1–3 October 2011; Volume 8285, p. 82856L. [Google Scholar]
  70. James, A.P. Heart rate monitoring using human speech spectral features. Hum. Cent. Comput. Inf. Sci. 2015, 5, 1–12. [Google Scholar] [CrossRef] [Green Version]
  71. Kajarekar, S.; Malayath, N.; Hermansky, H. Analysis of sources of variability in speech. In Proceedings of the Sixth European Conference on Speech Communication and Technology, Budapest, Hungary, 5–9 September 1999. [Google Scholar]
  72. Pachet, F.; Roy, P. Analytical features: A knowledge-based approach to audio feature generation. EURASIP J. Audio Speech Music. Process. 2009, 2009, 153017. [Google Scholar] [CrossRef] [Green Version]
  73. Turgut, Ö. The acoustic cues of fear: Investigation of acoustic parameters of speech containing fear. Arch. Acoust. 2018, 43, 245–251. [Google Scholar]
  74. Thakur, S.; Adetiba, E.; Olusgbara, O.O.; Millham, R. Experimentation using short-term spectral features for secure mobile internet voting authentication. Math. Probl. Eng. 2015, 2015, 564904. [Google Scholar] [CrossRef]
  75. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  76. Kotsiantis, S.B.; Pintelas, P.E. Combining bagging and boosting. Int. J. Comput. Intell. 2004, 1, 324–333. [Google Scholar]
  77. de Almeida, R.; Goh, Y.M.; Monfared, R.; Steiner, M.T.A.; West, A. An ensemble based on neural networks with random weights for online data stream regression. Soft Comput. 2019, 1–21. [Google Scholar] [CrossRef] [Green Version]
  78. Huang, M.W.; Chen, C.W.; Lin, W.C.; Ke, S.W.; Tsai, C.F. SVM and SVM ensembles in breast cancer prediction. PLoS ONE 2017, 12, e0161501. [Google Scholar] [CrossRef] [PubMed]
  79. Xing, H.J.; Liu, W.T. Robust AdaBoost based ensemble of one-class support vector machines. Inf. Fusion 2020, 55, 45–58. [Google Scholar] [CrossRef]
  80. Navarro, C.F.; Perez, C. A Color–texture pattern classification using global–local feature extraction, an SVM classifier with bagging ensemble post-processing. Appl. Sci. 2019, 9, 3130. [Google Scholar] [CrossRef] [Green Version]
  81. Wu, Y.; Ke, Y.; Chen, Z.; Liang, S.; Zhao, H.; Hong, H. Application of alternating decision tree with AdaBoost and bagging ensembles for landslide susceptibility mapping. Catena 2020, 187, 104396. [Google Scholar] [CrossRef]
  82. Zvarevashe, K.; Olugbara, O.O. Gender voice recognition using random forest recursive feature elimination with gradient boosting machines. In Proceedings of the 2018 International Conference on Advances in Big Data, Computing and Data Communication Systems (icABCD), Durban, South Africa, 6–7 August 2018; pp. 1–6. [Google Scholar]
  83. Yaman, E.; Subasi, A. Comparison of bagging and boosting ensemble machine learning methods for automated EMG signal classification. BioMed Res. Int. 2019, 2019, 9152506. [Google Scholar] [CrossRef]
  84. Friedman, J. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189. [Google Scholar] [CrossRef]
  85. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  86. Olugbara, O.O.; Taiwo, T.B.; Heukelman, D. Segmentation of melanoma skin lesion using perceptual color difference saliency with morphological analysis. Math. Probl. Eng. 2018, 2018, 1524286. [Google Scholar] [CrossRef]
  87. Livieris, I.E.; Kotsilieris, T.; Tampakas, V.; Pintelas, P. Improving the evaluation process of students’ performance utilizing a decision support software. Neural Comput. Appl. 2019, 31, 1683–1694. [Google Scholar] [CrossRef]
Figure 1. The Ryerson audio-visual database of emotional speech and song (RAVDESS).
Figure 1. The Ryerson audio-visual database of emotional speech and song (RAVDESS).
Algorithms 13 00070 g001
Figure 2. The Surrey audio-visual expressed emotion (SAVEE).
Figure 2. The Surrey audio-visual expressed emotion (SAVEE).
Algorithms 13 00070 g002
Figure 3. Flowchart for different configurations of emotion recognition methods.
Figure 3. Flowchart for different configurations of emotion recognition methods.
Algorithms 13 00070 g003
Table 1. A comparison of our method with related methods.
Table 1. A comparison of our method with related methods.
ReferenceDatabaseType of RecordingNumber of EmotionsFeature SetClassification MethodResult (%)
[3]Emo-DBActedAngry, Happy, Neutral, Sad Energy + MFCC + ZCR + voicing probability + fundamental frequency FCBF + MVT84.19
[11]Spanish ActedAnger, Disgust, Fear, Neutral, Surprise, Sadness, JoyMFCC + MS RNN classifier90.05
[15]Emo-DBActedAngry, Happy, NeutralGCZCMT SVM 84.45
[19]Emo-DBActedAnger, Happiness, Neutral, SadnessProsodic + sub-band + MFCC + LPC SFS algorithm + SVM83.00
[28]eNTERFACE’05 ElicitedDisgust, Surprise, Happy, Anger, Sad, Fear,Pitch, energy, formants, intensity and ZCR + text SVM90.00
[29]CASIA ActedNeutral, Happy, Sadness, Fear, Angry, SurpriseProsodic + quality characteristics + MFCC correlation analysis + Fisher + ELM decision tree89.60
[36]IEMOCAP ActedAngry, Happy, Neutral, SadIS10 + MFCCs + eGemaps + SoundNet + VGGishSVM64.00
[39]Emo-DBActedAnger, Boredom, Happy, Neutral, SadnessProsodic features + paralinguistic features SVM94.90
[42]CASIA ActedSurprise, Happy, Sad, Angry, Fear, NeutralMFCC GA-BEL + PCA + LDA90.28
[46]CASIA ActedAngry, Fear, Happy, Neutral, Surprise, SadMFCC, pitch, formant, ZCR and short-term energy SVM + DBN95.80
[47]Emo-DBActedSadness, Fear, Joy, Anger, Surprise, Disgust, NeutraleGeMAPS CSS stacking system +SVM82.45
[48]Emo-DBActedAnger, Happiness, Sadness, Boredom, Neutral, Disgust, Fear,MFCCs bagged ensemble of SVMs92.45
[49]RAVDESSActedNeutral, Surprise, Happy, Angry, Calm, Sad, Fearful, DisgustCWT, prosodic coefficients SVM60.10
[50]Spanish ActedAnger, Joy, Disgust, Neutral, Surprise, Fear, SadnessSMFCC, ECC, MFF, MS and EFCC RNN 91.16
Proposed modelRAVDESS/SAVEEActedAngry, Sad, Happy, Disgust, Calm, Fear, Neutral, SurpriseProsodic + spectral RDF ensemble99.55
Table 2. List of the selected prosodic and spectral features.
Table 2. List of the selected prosodic and spectral features.
GroupTypeNumber
Prosodic
EnergyLogarithm of Energy10
PitchFundamental Frequency70
TimesZero Crossing Rate24
Spectral
CepstralMFCC133
ShapeSpectral Roll-off Point12
AmplitudeSpectral Flux12
MomentSpectral Centroid22
AudioSpectral Compactness10
FrequencyFast Fourier Transform9
SignatureSpectral Variability21
EnvelopeLPCC81
Total404
Table 3. Processing time of learning algorithms trained with Mel-frequency cepstral coefficient 1 (MFCC1), Mel-frequency cepstral coefficient 2 (MFCC2) and a hybrid acoustic feature (HAF).
Table 3. Processing time of learning algorithms trained with Mel-frequency cepstral coefficient 1 (MFCC1), Mel-frequency cepstral coefficient 2 (MFCC2) and a hybrid acoustic feature (HAF).
SAVEE FeatureRAVDESS Feature
ClassifierMFCC1MFCC2HAFMFCC1MFCC2HAF
RDF0.030.380.430.0521.620.06
GBM2.30579.40343.105.901172.306.90
ABC0.4722.161.382.592.791.89
BSVM0.2049.790.390.42211.540.48
BMLP3.131319.39952.4676.00231.167.18
Table 4. Percentage average precision, recall, F1-score and accuracy with confidence intervals of learning algorithms trained with MFCC1, MFCC2 and HAF.
Table 4. Percentage average precision, recall, F1-score and accuracy with confidence intervals of learning algorithms trained with MFCC1, MFCC2 and HAF.
SAVEE FeatureRAVDESS Feature
Classifier/MeasureMFCC1MFCC2HAFMFCC1MFCC2HAF
RDF
Precision63.4 (±0.044)78.1 (±0.037)99.1(±0.009)90.1 (±0.016)93.1 (±0.013)99.6 (±0.003)
Recall72.1 (±0.041)77.4 (±0.038)99.1(±0.009)88.9 (±0.016)94.0 (±0.012)99.5 (±0.004)
F1-score63.6 (±0.043)76.6 (±0.039)99.1(±0.009)88.5 (±0.017)92.5 (±0.014)99.8 (±0.002)
Accuracy61.5 (±0.044)66.7 (±0.043)99.3 (±0.008)90.7 (±0.015)93.7 (±0.013)99.8 (±0.002)
GBM
Precision64.0 (±0.043)75.6 (±0.039)99.4 (±0.007)85.0 (±0.019)86.5 (±0.018)95.3(±0.011)
Recall66.6 (±0.043)74.6 (±0.039)99.1 (±0.009)82.9 (±0.020)88.1 (±0.017)94.8(±0.012)
F1-score62.6 (±0.044)72.7 (±0.040)99.3 (±0.008)82.6 (±0.020)85.9 (±0.018)96.8(±0.009)
Accuracy61.5 (±0.044)65.5 (±0.043)99.3 (±0.008)85.4(±0.018)86.7(±0.018)92.6(±0.014)
ABC
Precision62.9 (±0.044)73.4 (±0.040)99.1 (±0.009)81.3 (±0.020)85.4 (±0.018)94.3(±0.012)
Recall64.1 (±0.044)73.7 (±0.040)99.4 (±0.007)84.0 (±0.019)87.5 (±0.017)94.5(±0.012)
F1-score63.1 (±0.044)74.1 (±0.040)97.9 (±0.013)82.6 (±0.020)84.5 (±0.019)95.9(±0.010)
Accuracy58.0 (±0.044)62.8 (±0.043)98.0 (±0.008)83.0 (±0.020)85.2 (±0.018)92.0 (±0.014)
BSVM
Precision61.4 (±0.044)71.4 (±0.041)98.7 (±0.010)81.0 (±0.020)84.1 (±0.019)91.1(±0.015)
Recall61.0 (±0.044)72.6 (±0.041)99.0 (±0.009)82.5 (±0.020)82.3(±0.020)92.0(±0.014)
F1-score61.9 (±0.044)72.0 (±0.041)99.3 (±0.008)80.5 (±0.021)83.5 (±0.019)90.3(±0.015)
Accuracy56.0 (±0.045)61.5 (±0.044)96.0 (±0.013)82.7 (±0.020)84.5 (±0.019)91.8 (±0.014)
BMLP
Precision59.4 (±0.044)69.0 (±0.042)97.9 (±0.013)78.8 (±0.021)77.5 (±0.022)93.0 (±0.013)
Recall60.1 (±0.044)72.0 (±0.041)98.4 (±0.011)75.8 (±0.022)80.3 (±0.021)88.0(±0.017)
F1-score59.6 (±0.044)70.3 (±0.041)98.1 (±0.012)74.4(±0.023)78.3 (±0.021)89.0 (±0.016)
Accuracy55.4 (±0.045)60.7 (±0.044)94.6 (±0.002)75.6 (±0.022)79.3 (±0.021)91.3 (±0.015)
Table 5. Percentage precision, recall and F1-score and accuracy on RAVDESS MFCC1.
Table 5. Percentage precision, recall and F1-score and accuracy on RAVDESS MFCC1.
Emotion
Classifier/MeasureAngryCalmDisgustFearHappyNeutralSadSurprise
RDF
Precision100.0100.0100.092.080.053.087.096.0
Recall100.0100.0100.083.082.086.079.091.0
F1-score100.0100.0100.087.081.066.083.094.0
Accuracy91.090.091.089.091.089.091.093.0
GBM
Precision94.0100.0100.087.069.039.087.085.0
Recall100.0100.0100.074.069.078.072.087.0
F1-score97.0100.0100.080.069.052.079.086.0
Accuracy86.085.287.081.083.085.086.090.0
ABC
Precision91.082.078.079.081.079.081.081.0
Recall100.0100.0100.081.056.085.071.082.0
F1-score95.0100.091.065.069.070.081.079.0
Accuracy85.084.085.075.081.079.087.088.0
BSVM
Precision83.084.082.075.075.077.079.089.0
Recall82.083.091.073.079.081.084.087.0
F1-score81.082.079.079.081.079.082.084.0
Accuracy83.085.081.079.080.681.086.086.0
BMLP
Precision72.081.078.068.070.070.075.081.0
Recall78.084.082.070.076.079.078.083.0
F1-score74.082.080.067.071.075.077.080.0
Accuracy80.077.076.071.074.075.075.077.0
Table 6. Percentage precision, recall and F1-score and accuracy on RAVDESS MFCC2.
Table 6. Percentage precision, recall and F1-score and accuracy on RAVDESS MFCC2.
Emotion
Classifier/MeasureAngryCalmDisgustFearHappyNeutralSadSurprise
RDF
Precision100.0100.0100.091.095.064.090.0100.0
Recall100.0100.0100.089.078.0100.090.095.0
F1-score100.0100.0100.093.086.078.090.098.0
Accuracy96.093.092.090.094.091.096.097.6
GBM
Precision100.0100.0100.087.079.055.080.086.0
Recall100.0100.0100.087.075.086.067.090.0
F1-score100.0100.0100.087.077.067.073.088.0
Accuracy88.084.083.082.085.085.088.098.6
ABC
Precision94.093.091.078.071.079.081.089.0
Recall96.0100.0100.086.076.073.083.086.0
F1-score92.095.0100.079.074.076.081.086.0
Accuracy86.082.083.081.087.082.088.092.6
BSVM
Precision81.094.091.077.074.079.078.094.0
Recall92.089.090.076.074.077.074.090.0
F1-score94.091.091.077.074.078.076.092.0
Accuracy84.081.080.078.085.085.087.096.0
BMLP
Precision76.089.086.072.070.076.074.083.0
Recall75.086.086.074.067.075.072.085.0
F1-score77.091.087.074.073.076.074.090.0
Accuracy85.081.078.073.078.077.078.084.0
Table 7. Percentage precision, recall and F1-score and accuracy on RAVDESS HAF.
Table 7. Percentage precision, recall and F1-score and accuracy on RAVDESS HAF.
Emotion
Classifier/MeasureAngryCalmDisgustFearHappyNeutralSadSurprise
RDF
Precision100.0100.0100.0100.0100.0100.0100.098.0
Recall100.0100.0100.0100.0100.096.0100.0100.0
F1-score100.0100.0100.0100.0100.098.0100.099.0
Accuracy100.0100.0100.098.0100.0100.0100.0100.0
GBM
Precision93.0100.0100.094.0100.095.099.093.0
Recall92.0100.0100.093.097.0100.076.0100.0
F1-score92.0100.0100.093.098.097.086.096.0
Accuracy95.091.092.088.091.090.096.098.0
ABC
Precision92.0100.096.094.0100.095.093.097.0
Recall91.0100.0100.095.093.076.0100.0100.0
F1-score92.0100.098.093.095.091.087.098.0
Accuracy93.090.092.086.093.091.095.096.0
BSVM
Precision91.0100.099.092.089.074.083.094.0
Recall91.0100.0100.090.086.085.086.098.0
F1-score93.0100.0100.091.086.077.087.095.0
Accuracy93.093.092.086.090.091.093.096.0
BMLP
Precision92.0100.093.090.088.073.081.091.0
Recall91.0100.097.093.088.086.088.097.0
F1-score90.0100.091.088.083.075.084.091.0
Accuracy92.091.093.089.091.088.092.094.0
Table 8. Percentage precision, recall and F1-score and accuracy on SAVEE MFCC1.
Table 8. Percentage precision, recall and F1-score and accuracy on SAVEE MFCC1.
Emotion
Classifier/MeasureAngryDisgustFearHappyNeutralSadSurprise
RDF
Precision65.024.065.048.094.048.0100.0
Recall65.083.057.065.052.083.0100.0
F1-score65.037.060.055.067.061.0100.0
Accuracy69.066.060.075.064.077.081.0
GBM
Precision65.029.075.026.091.062.0100.0
Recall62.067.048.055.062.072.0100.0
F1-score63.040.059.035.074.067.0100.0
Accuracy70.066.057.074.062.079.084.0
ABC
Precision66.028.064.036.089.059.098.0
Recall68.029.062.040.091.059.0100.0
F1-score65.028.063.038.091.058.099.0
Accuracy60.057.053.066.062.079.083.0
BSVM
Precision65.023.061.041.085.057.098.0
Recall64.026.057.038.081.061.0100.0
F1-score66.029.062.036.083.058.099.0
Accuracy56.054.051.069.056.077.085.0
BMLP
Precision64.023.059.038.083.054.095.0
Recall63.025.055.039.085.057.097.0
F1-score64.025.059.037.081.055.096.0
Accuracy52.050.049.051.053.064.069.0
Table 9. Percentage precision, recall and F1-score and accuracy on SAVEE MFCC2.
Table 9. Percentage precision, recall and F1-score and accuracy on SAVEE MFCC2.
Emotion
Classifier/MeasureAngryDisgustFearHappyNeutralSadSurprise
RDF
Precision71.085.069.068.095.059.0100.0
Recall70.084.065.067.068.088.0100.0
F1-score71.084.068.068.074.071.0100.0
Accuracy77.073.063.070.069.082.0100.0
GBM
Precision72.065.083.047.093.069.0100.0
Recall69.074.069.062.071.077.0100.0
F1-score70.068.071.048.077.075.0100.0
Accuracy74.071.059.073.069.086.092.0
ABC
Precision74.071.070.058.069.073.099.0
Recall76.072.068.061.068.071.0100.0
F1-score74.075.073.059.065.073.0100.0
Accuracy71.067.056.069.069.071.099.0
BSVM
Precision71.068.066.059.067.071.098.0
Recall72.072.064.061.064.075.0100.0
F1-score70.070.066.060.065.073.0100.0
Accuracy71.062.051.067.068.073.0100.0
BMLP
Precision70.066.064.057.066.067.093.0
Recall72.070.066.059.068.072.097.0
F1-score71.068.064.059.065.069.096.0
Accuracy56.052.049.065.054.064.085.0
Table 10. Percentage precision, recall and F1-score and accuracy on SAVEE HAF.
Table 10. Percentage precision, recall and F1-score and accuracy on SAVEE HAF.
Emotion
Classifier/MeasureAngryDisgustFearHappyNeutralSadSurprise
RDF
Precision100.0100.094.0100.0100.0100.0100.0
Recall100.094.0100.0100.0100.0100.0100.0
F1-score100.0100.0100.0100.0100.097.097.0
Accuracy100.099.097.0100.099.1100.0100.0
GBM
Precision100.096.0100.0100.0100.0100.0100.0
Recall100.0100.094.0100.0100.0100.0100.0
F1-score100.098.097.0100.0100.0100.0100.0
Accuracy100.098.199.0100.0100.0100.0100.0
ABC
Precision100.094.0100.0100.0100.0100.0100.0
Recall100.0100.096.0100.0100.0100.0100.0
F1-score100.094.091.0100.0100.0100.0100.0
Accuracy100.099.099.0100.0100.0100.0100.0
BSVM
Precision100.093.099.099.0100.0100.0100.0
Recall100.096.098.099.0100.0100.0100.0
F1-score100.096.099.0100.0100.0100.0100.0
Accuracy100.094.096.096.0100.0100.0100.0
BMLP
Precision100.090.098.097.0100.0100.0100.0
Recall100.093.097.099.0100.0100.0100.0
F1-score100.091.098.098.0100.0100.0100.0
Accuracy100.093.095.099.0100.0100.0100.0

Share and Cite

MDPI and ACS Style

Zvarevashe, K.; Olugbara, O. Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition. Algorithms 2020, 13, 70. https://doi.org/10.3390/a13030070

AMA Style

Zvarevashe K, Olugbara O. Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition. Algorithms. 2020; 13(3):70. https://doi.org/10.3390/a13030070

Chicago/Turabian Style

Zvarevashe, Kudakwashe, and Oludayo Olugbara. 2020. "Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition" Algorithms 13, no. 3: 70. https://doi.org/10.3390/a13030070

APA Style

Zvarevashe, K., & Olugbara, O. (2020). Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition. Algorithms, 13(3), 70. https://doi.org/10.3390/a13030070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop