Next Article in Journal
Evaluating Scoliosis Severity Based on Posturographic X-ray Images Using a Contrastive Language–Image Pretraining Model
Next Article in Special Issue
Identification of TLE Focus from EEG Signals by Using Deep Learning Approach
Previous Article in Journal / Special Issue
Constructing the Schizophrenia Recognition Method Employing GLCM Features from Multiple Brain Regions and Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of Emotions from Brain Signals on 3D VAD Space via Artificial Intelligence Techniques

1
Biomedical Engineering Graduate Program, Graduate School of Natural and Applied Sciences, Erciyes University, 38039 Kayseri, Türkiye
2
Department of Biomedical Engineering, Engineering Faculty, Erciyes University, 38039 Kayseri, Türkiye
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(13), 2141; https://doi.org/10.3390/diagnostics13132141
Submission received: 19 April 2023 / Revised: 12 June 2023 / Accepted: 14 June 2023 / Published: 22 June 2023
(This article belongs to the Special Issue Application of Deep Learning in the Diagnosis of Brain Diseases)

Abstract

:
Recent achievements have made emotion studies a rising field contributing to many areas, such as health technologies, brain–computer interfaces, psychology, etc. Emotional states can be evaluated in valence, arousal, and dominance (VAD) domains. Most of the work uses only VA due to the easiness of differentiation; however, very few studies use VAD like this study. Similarly, segment comparisons of emotion analysis with handcrafted features also use VA space. At this point, we primarily focused on VAD space to evaluate emotions and segmentations. The DEAP dataset is used in this study. A comprehensive analytical approach is implemented with two sub-studies: first, segmentation (Segments I–VIII), and second, binary cross-comparisons and evaluations of eight emotional states, in addition to comparisons of selected segments (III, IV, and V), class separation levels (5, 4–6, and 3–7), and unbalanced and balanced data with SMOTE. In both sub-studies, Wavelet Transform is applied to electroencephalography signals to separate the brain waves into their bands (α, β, γ, and θ bands), twenty-four attributes are extracted, and Sequential Minimum Optimization, K-Nearest Neighbors, Fuzzy Unordered Rule Induction Algorithm, Random Forest, Optimized Forest, Bagging, Random Committee, and Random Subspace are used for classification. In our study, we have obtained high accuracy results, which can be seen in the figures in the second part. The best accuracy result in this study for unbalanced data is obtained for Low Arousal–Low Valence–High Dominance and High Arousal–High Valence–Low Dominance emotion comparisons (Segment III and 4.5–5.5 class separation), and an accuracy rate of 98.94% is obtained with the IBk classifier. Data-balanced results mostly seem to outperform unbalanced results.

1. Introduction

Emotions emerge in our brains due to interactions caused by neural and hormonal systems [1]. If we look at emotions and their interactions in the brain in detail with examples, the brain controls emotions. The limbic system is important for emotion. The amygdala has a significant role, especially in influencing anger and aggression. Fear can appear due to the stimulated amygdala, i.e., with frightening memory flashbacks. The amygdala can enhance emotional stimuli perception. The septum influences anger and fear [2]. A phobia is an anxiety disorder occurring with extreme fear of the stimuli, such as a fear of heights (acrophobia), spiders (arachnophobia), crowds, etc. Physical and emotional symptoms appear with a phobia crisis, such as increased heart rate and breathing, sweating, and difficulty controlling emotions [3]. Besides its other vital functions, the hypothalamus is also crucial in consciousness control, regulating emotions, and responses to stress, as well as pleasure and pain [2]. Emotions affect perception, memory, creativity, behavior, etc. [1]. Perceiving and expressing emotions are abilities related to emotional intelligence. Problem-solving abilities can be positively affected by emotions. Autism also causes some difficulties in language and emotion. The internal context also affects memory retrieval of emotions, moods, consciousness, etc. [2]. Brain waves are separated into five bands: delta, theta, alpha, beta, and gamma. During a relaxed state, alpha waves dominate the cerebral cortex, gradually disappearing as emotional activity strengthens. During stressed conditions, alpha wave frequency increases and amplitude decreases, and alpha waves gradually turn to beta waves. Beta wave occurrence indicates an excited state in the cerebral cortex [4]. Chikara et al. [5] have used the fMRI–EEG system to implement a modified stop-signal task in a virtual battlefield scenario to study the neural mechanism of response inhibition under reward and punishment, which can modulate motivation and emotions and affect cognitive processing. According to their results, positive and negative monetary feedback made a difference for both behavioral indices of inhibition and brain activities.
Emotion analysis determines, compares, and evaluates emerging emotions, activating changes using various methods. Recently, emotion analysis has become increasingly attractive, and many intelligent applications for emotion recognition have been presented [6]. Artificial intelligence techniques allow computer programs to gain abilities, such as learning, solving problems, making decisions, etc. [7], and these intelligence techniques can also be used for emotion recognition despite some difficulties and limitations due to some factors, such as hiding feelings [8]. Even though emotions can be recognized by facial expressions, voice, or gesture recognition [6,9,10], these methods can be manipulated, preventing actual emotional states from being understood easily and clearly [9,10]. Because inner systems are affected less by human subjective consciousness, physiological signals can reflect emotions more truly [6]. EEG physiological brain signals remain neutral in this context and help to accurately analyze emotional states with more reliable data [9,10].
Visual, auditory, both visual and auditory, or memory-based stimuli can be used to obtain emotions. If we look at some definitions and some studies within the scope of mood analysis, two models (discrete and dimensional) are used in emotion prediction. The discrete model has six emotions, which are anger, fear, happiness, hatred, sadness, and surprise [11]. The dimensional emotion model includes 2D and 3D models. The 2D VA model has valence and arousal dimensions. The 3D VAD model has valence, arousal, and dominance dimensions [12]. The valence dimension ranges from negative emotions, such as unpleasant/unhappy/sad, to positive emotions, such as pleasant/happy/joyful. The arousal dimension ranges from calm/bored to stimulated/excited. The dominance dimension ranges from submissive/without control to dominant/in control and empowered [12,13]. The dimensional emotion model is preferred due to its international identification and acceptance [11]. The 3D Emotion Space Model can be seen in Figure 1, which shows a three-dimensional (valence, arousal, and dominance) space that occurs using the data in [14], taking the pleasure plane as valence. Verma and Tiwary [12] proposed a 3D emotion model on VAD space for describing various emotions, including fun, joy, cheerfulness, melancholy, love, mellow, shock, etc. In general, 3D space is more sufficient and accurate for representing all emotions than 2D space.
Many studies on emotion analysis appear in the literature; however, these studies generally focus on 2D valence and arousal space. Little work focuses on 3D valence, arousal, and dominance space. In the present study, we especially wanted to study 3D space to classify emotions and to compare emotions, segmentations, and classes. EEG is used in this study to conduct space classifications and comparisons. Verma and Tiwary [12] obtained a classification accuracy (%) of 63.47 (valence), 69.62 (arousal), and 63.57 (dominance) for a tree class classification (low, medium, and high) with the DEAP database besides their own experiment. Guendil et al. [15] used a feature-based wavelet and R-ELM (extreme learning machine) classifier (with physiological signals in the DEAP database). They obtained classification rates of 73.43% (V), 72.65% (A), and 69.3% (D) for the two classes. They obtained the best average classification accuracy of 53% for four class classifications in the VA quadrant. Özel et al. [16] studied multivariate synchro squeezing transform on EEG signals from the DEAP database for feature extraction to classify emotions. They used a 2D VA model for four emotional states and obtained the best accuracy of 79.1%. They used a 3D VAD model for eight emotional states with an SVM classifier and obtained the highest accuracy of 93%. Accuracies are higher for eight emotions than in the four emotion model.
The classification and comparison of emotions could contribute to many areas, such as intelligent health technologies, brain–computer interfaces (BCIs), psychology, patient robots, humanoid robot technologies, virtual reality, holograms, simulation systems, game technologies, interactive education systems, smart homes, workplaces, driver systems, etc. This study contributes to the literature with its comparisons and comprehensive analytical approach implemented with two sub-studies: first, a segmentation pre-study, and second, an emotion comparison and evaluation study, including comparisons of selected segments, class separation levels, and data balancing.
Our study’s general results show that accuracies obtained from balanced datasets mostly seem to outperform those obtained from unbalanced datasets. The accuracies of unbalanced data in the figures of the second part of this study generally seem to increase in the order of A&G, D&F, B&H, and C&E comparisons. We have obtained high accuracy results that can be seen directly in the spider web diagrams in the second part of this study. The best accuracy result in this study for unbalanced data is for the C&E comparison (Seg III and 4.5–5.5 class separation), and an accuracy rate of 98.94% CCI (F1: 98.9% and MCC: 97.9%) is obtained with the IBk classifier, which is greater than the compared studies from the literature which has a dominance dimension besides valence and arousal.
This study is divided into two sub-studies. In the first part of this study, the comparison of the VAD emotions via various segmentations of EEG signals is objected to. Segments I–VIII are obtained by applying the non-overlapping windowing method. In the second part of this study, Segments III, IV, and V are chosen, and the two class cross-comparisons of the VAD emotions, segments, and class levels are objected to. In both parts of this study, Discrete Wavelet Transform (DWT) is applied to EEG signals to separate the brain waves into their bands; d1, d2, d3, and a3 wavelet coefficients are obtained, which correspond to gamma, beta, alpha, and theta bands (γ, β, α, and θ bands), respectively. Attributes are also calculated for all channels for each of the 32 participants and for each of the 40 video pieces. After obtaining the feature datasets, classifications are conducted with classifiers, such as SMO, FURIA, etc. Because of the unbalanced classes, data balancing and comparing balanced and unbalanced results are also targeted in this study. In this context, the SMOTE method is used to pre-filter the dataset for oversampling of the minority classes. Classifications are conducted both with and without the SMOTE method in this study. This article is organized into five sections. Section 2 describes the materials and methods used in this study in detail with its subsections (Section 2.1. Dataset and Channels; Section 2.2. Overall Method and Classification; Section 2.3. Classes; Section 2.4. Attributes; Section 2.5. SMOTE; Section 2.6. Classifiers; and Section 2.7. Metrics). Section 3 presents the results for the two sub-studies (Section 3.1. Segmentation Pre-Study and Section 3.2. Cross-Comparisons of 3D VAD Emotions, Segments, and Classes). Section 4 includes a discussion of the results. Section 5 consists of the conclusion.

2. Materials and Methods

2.1. Dataset and Channels

The DEAP database is used in this study. It is a multimodal database for emotion analysis. The EEG and peripheral physiological signals of 32 participants (16 males and 16 females) aged between 19 and 37 (mean 26.9) and facial videos of 22 of the participants were recorded while they were watching 40 music videos, each of which was 1 min long. Self-assessments were also performed by the participants in addition to signal recordings. The participants rated each video according to their arousal, valence, liking, dominance, and familiarity levels (with continuous 1–9 scaled self-assessment manikins, except for familiarity, rated with a discrete 1–5 scaled manikin). The database contains both preprocessed and unprocessed data. EEG data were down-sampled to 128 Hz, EOG artifacts were removed, and a band-pass frequency filter of 4.0–45.0 Hz was applied. The data were also down-sampled to 128 Hz as a preprocessing measure for the other channels (EOG, EMG, GS, respiratory belt, plethysmography, and temperature) [17]. In the present study, preprocessed DEAP data were used. The 32 EEG channels used in this study are given in Figure 2.

2.2. Overall Method and Classification

The first three seconds of the EEG data from the DEAP dataset are subtracted from the signal, and the remaining one minute signal is used for further operations in this study. The windowing method is used for the segmentation of 60 s EEG signals. The flow chart for EEG signal analysis is given in Figure 3.
This study is divided into two parts. The flow chart of the overall method is given in Figure 4. The segmentation of EEG signals is studied in the first part. The 30, 20, 15, 10, 5, 3, 2, and 1 s non-overlapping windowing method is used to segment the signals. Here, time in seconds refers to the fact that the 60 s signal will be divided into segments. Windowing separations are titled with segment names, and the 2, 3, 4, 6, 12, 20, 30, and 60 signal pieces correspond to 30, 20, 15, 10, 5, 3, 2, and 1 s windowing, and these correspond to Segments I, II, III, IV, V, VI, VII, and VIII, respectively. For instance, Segment I means that the 60 s EEG signal is divided into 2 segments with 30 s windowing. N means no segmentation is applied to the signal (Section 3.1. Segmentation Pre-Study).
In the second part of this study, Segments III, IV, and V, which correspond to 15, 10, and 5 s windowing, are chosen for the 2 class cross-comparisons of the VAD areas (3.2. Cross-Comparisons of 3D VAD emotions, segments, and classes).
Wavelet Transform is a powerful statistical tool applied to various problems, and it has proven its effectiveness in various areas, such as time–frequency analysis of signals, image processing, denoising and extraction of weak signals, data compression, medical image technology, pattern recognition, speech recognition, fingerprint verification, etc. [18,19]. The local analysis of a signal is performed using the wavelets in this method. Unlike the Fourier Transform, which uses the time–frequency domain and constant window function, Wavelet Transform uses the time–scale domain and continuously varying window function [20]. Different mother wavelets, such as Harr, Daubechies, etc., can be used in Discrete Wavelet Transform (DWT) [18]. In both parts of this study, DWT is applied to the complete EEG signal as well as to all segments of the signal, and d1, d2, d3, and a3 wavelet coefficients are obtained. A Reverse Bior wavelet is used for the analysis.
To form the feature datasets for each signal or segment, the mean, Shannon entropy, variance, harmonic mean, mode, skewness, kurtosis, signal energy, Higuchi fractal dimension, mobility, complexity, LZ complexity, Lyapunov exponent, Hurst exponent, median, peak to peak distance, root mean square, root sum of squares, mean frequency, median frequency, power bandwidth, correlation dimension, mean absolute deviation, and interquartile range attributes are used in this study (24 attributes). They are calculated for all channels for each of the 32 participants and for each of the 40 video pieces.
After obtaining the feature datasets, Random Forest, SMO, IBk (1-Manh), FURIA, Bagging, Random Committee, and Random Subspace classifiers are studied and the results, such as accuracy rates (CCI %, F1 scores, and MCC values), are obtained for the data. In all classifications, 10-fold cross-validation is applied to the data.
A SMOTE filter is also applied to the data to balance the classes. The oversampling SMOTE method is applied, and synthetic samples are generated for the minority class. All minority classes are converged to the majority class for data balancing. Both filtered and unfiltered results are obtained for the comparison of emotions.
We generally compared eight three-dimensional emotion regions in this study: A: happy, joyful, useful, powerful, influential, friendly, excited, and inspired; B: relaxed, leisurely, untroubled, and satisfied; C: disdainful and unconcerned; D: angry, hostile, cruel, insolent, hate, and disgusted; E: impressed, surprised, thankful, and curious; F: protected, humble, and reverent. These emotion groups we compared are discussed in the discussion section in more detail. Besides the discussion section, cross-comparisons can also be seen in Figure 4.
The MATLAB program (R2022a) calculates the wavelets, 32 EEG channels, and attributes, and the WEKA program (Waikato Environment for Knowledge Analysis) version 3.9.6. is used for the classification [21].

2.3. Classes

In this study, classes are set by considering high and low values of valence, arousal, and dominance scores. Class placements on VAD space are shown in Figure 5, in which the emotions are placed in the eight emotion regions. The valence dimension is considered high and low (not as positive or negative), just like the other dimensions (valence and dominance) in this study, to simplify the understanding.
The letters in Figure 5 indicate the combined coordinates of arousal, valence, and dominance, respectively: A (HaHvHd); B (LaHvHd); C (LaLvHd); D (HaLvHd); E (HaHvLd); F (LaHvLd); G (LaLvLd); and H (HaLvLd). Four comparisons are conducted between the 3-dimensional regions as A&G (HaHvHd-LaLvLd), B&H (LaHvHd-HaLvLd), D&F (HaLvHd-LaHvLd), and C&E (LaLvHd-HaHvLd). In total, 5, 4–6, and 3–7 class separations are used for the A&G, B&H, and D&F emotion comparisons. Furthermore, 5, 4.5–5.5, and 4–6 class separations are used for the C&E emotion comparisons. 0-midpoint on VAD space in Figure 5 indicates 5-midpoint for the 1–9 scale for 5 class separation. For instance, for a 4–6 separation, a high value refers to ≥6, and a low value refers to ≤4.

2.4. Attributes

Brief definitions of the attributes used are given below.
The sample mean is the numerical average of the data. Variance can be defined as the average of the square of the differences from the mean [22]. The harmonic mean is defined as the reciprocal of the arithmetic mean of the reciprocals of the values in the observation, while the mode is the value that occurs most frequently, and the median is the value of the variable that divides the distribution into two parts [23]. Shannon entropy is proposed to measure how the signal information can be quantified with absolute precision [24]. Skewness is a 3rd order moment, which shows asymmetry of the data distribution about the mean, while kurtosis is a 4th order moment indicating the major peaks in the signal’s time domain [25]. If s(t) is a signal, the signal power is s”(t), and the signal energy is [26]
E = s 2 t dt
The interquartile range (IQR) is a statistical distribution measure, and it is defined as the difference between the 3rd–1st quartiles (Q3–Q1) [27]. The mean absolute deviation (MAD) is defined as the average of absolute distances from the data mean. For n number of observations (x1, x2…, xn)
MAD = k = 1 n x i x i n
where x is the mean of the data [28]. The mean frequency (MNF) is defined as the sum of the product of the signal power spectrum and the frequency divided by the total sum of the power spectrum.
MNF = j = 1 M f j P j / j = 1 M P j
Pj is the power spectrum at a frequency bin j, fj is the frequency of the spectrum at a frequency bin j, and M is the total number of frequency bins [29]. The median frequency (MDF) is a frequency at which the signal power spectrum is divided into two regions with equal integrated power [29].
MDF = j = 1 M P j = j = MDF M P j = 1 2 j = 1 M P j
The correlation dimension (CD) is a method for determining the correlation of a time domain signal sampled uniformly. It is a method to determine the dimension of a nonlinear signal [30].
C r = 2 N N 1 i j θ ( r | X i X j |
divided into two regions with an equal integrated power.
Root mean square (RMS) is the square root of the mean square value, which is defined for a specific interval of time. RSSQ is the root sum of squares [31]. The power bandwidth (PBW) is defined as the measure of the frequency difference between the points where the spectrum falls at least 3 dB below the reference level [32].
The Hjorth parameters (HP), extracted by Hjorth in 1970, are activity, mobility, and complexity. Activity can be defined as the information of the signal power; mobility is an estimation of the mean frequency; and complexity is the bandwidth, the change in frequency, or the standard deviation of the power spectrum. HPs are commonly used in biological signals and are suitable for non-stationary EEG signal analysis [25,33].
Activity = Var ( y t ) ;
Mobility = Var dy ( t ) dt Var y ( t )
Complexity = Mobility dy ( t ) dt Mobility y ( t )
The Lyapunov exponent (LE) measures the exponential convergence or divergence by considering the exponent to be lower or higher than 0, respectively [24]. lyap_r indicates the largest LE, while lyap_e indicates the whole spectrum of LE, whose positive exponents indicate chaos and unpredictability. LE measures total predictability [34]. It is a strong chaos indicator if it has a positive value, and as it increases, chaos increases while predictability decreases. The largest Lyapunov exponent (LLE) is used chiefly for the nonlinear analysis of physiological signals, and it is sufficient for chaos evaluation. Many algorithms can be used to calculate LE and LLE [30].
The Higuchi fractal dimension (HFD) method was first developed in 1988 by Higuchi et al. For a finite set of time series, if a new time series is obtained and the Lm(k) average value is calculated, this indicates the curve length (m is the initial time and k is the interval time). The HFD is then calculated in the slope between ln(L(k)) and ln(1/k) values [35].
First developed by Harold Edwin Hurst, the Hurst exponent (HE) statistically calculates the time series’ variability [24]. Considering a time series that increased previously, it is possible to identify if it is likely to increase more, less, or equally like the previous steps with the Hurst exponent [35]. The HE value can range from 0 to 1. If it is ranged as 0 < HE > 0.5, then the time series has long-ranged anti-correlations; if it is equal to 0.5, there are no correlations; if it is ranged as 0.5 < HE > 1, then the time series has long-ranged correlations; and, if the value equals to 1, then there is self-similarity in the time series [24]. Detrended fluctuation analysis (DFA) also uses a remarkably similar parameter to HE called the Hurst parameter H [35].
Lempel–Ziv (LZ) complexity is used frequently in the irregularity analysis of biomedical signals with its interpretability and suitability. LZ complexity measures the nonparametric complexity of the irregularity of time series [36]. LZ complexity evaluates the algorithmic complexity and determines the emergence rate of new patterns [37]. A greater LZ complexity means the new changes occur at a faster rate, indicating data change irregularly. In contrast, a lower LZ complexity means a slower rate of changes, and the data change regularly and periodically [36].

2.5. SMOTE

SMOTE is a data balancing method. If sample numbers of the classes in a dataset are not approximately equal, then there is an imbalance between the classes. If the data are imbalanced with different class sizes, using predictive accuracy would not be appropriate for classification [38]. Because there are imbalanced classes in our data from the DEAP dataset, we also used SMOTE to oversample our minority classes. Both classifications with and without the SMOTE method are performed in this study.
For the oversampling SMOTE method, synthetic samples are generated for the minority class. The sample and its nearest neighbor and the distance metric method are considered for creating a difference between them, which is multiplied by a random value at a [0, 1] interval and then added to the feature vector [39]. According to the results of Chawla et al. [38], for a minority class, SMOTE can improve accuracy, and it provides an innovative approach to oversampling. SMOTE also enhances the performance of the under-sampling of the majority class.

2.6. Classifiers

The classifiers used in this study are SMO from Function Classifiers, IBk from Lazy Classifiers, Random Forest and Optimized Forest from the Tree Structures, FURIA (Fuzzy Unordered Rule Induction Algorithm) from Rules, Bagging, Random Committee, and Random Subspace from Meta Classifiers. Classifiers are used with default parameters of the WEKA program, except for IBk. A 10-fold cross-validation is applied to the data during classifications.
IBk (Instance-Based Learning): the K-Nearest Neighbor (KNN) algorithm (lazy learning) is a supervised learning method used in a wide variety of fields in machine learning areas, such as pattern recognition, signal, image processing, etc. [25]. IBk is a KNN classifier. Several nearest neighbors can be explicitly specified or automatically determined. Different search algorithms can be used to speed up the finding process of nearest neighbors. IBk uses the same distance metric as KNN, and the distance function is a search method parameter [40]. The KNN classifier uses distances, such as Euclidean, Manhattan, Chebychev, Minkowski, etc., to determine the nearest neighbor in feature space [25]. This study uses Manhattan distance and 1 nearest neighbor parameters (K = 1) for classification.
SMO (Sequential Minimum Optimization): SVM is another supervised learning method [25], and an SVM quadratic programming problem can be easily solved with SMO. It is a simple algorithm for solving the problem by separating it into subproblems that are solved extremely fast, causing the overall problem to be solved too rapidly [41].
It is possible to achieve accurate classification and regression with suitable randomness. Random Forest (RF) is developed by Breiman, and it is an effective tool in prediction [42]. An ensemble classification method with high accuracy results, it randomly builds multiple trees in subspaces of the feature space [43]. The general error depends on the strength and correlation of the individual trees in the forest [42].
FURIA (Fuzzy Unordered Rule Induction Algorithm) is an advanced version of the RIPPER algorithm, which is a rule learner after some modifications and extensions. For instance, it uses fuzzy rules instead of conventional rules, it uses unordered rule sets instead of rule lists, and it uses a novel rule stretching technique. According to some experiments, FURIA outperforms the original RIPPER algorithm and some other fuzzy rule learning methods [44].
The Bagging method generates predictor versions by making bootstrap replicates of the learning set, and then it uses them to obtain an aggregated predictor. Classification and regression experiments with real and simulated datasets show that Bagging can significantly improve accuracy [45].
Random Committee (RC) is a classification method that builds an ensemble of randomizable base classifiers. Different random number seeds are used for each base classification, and predictions from each are averages to form the overall prediction value [21].
Random Subspace (RSS) is formed by multiple trees such that subsets will be selected pseudo-randomly from the feature vector. It forms a decision-tree-based classifier with the highest training data accuracy [21].
The Optimized Forest (OF) algorithm determines the number of trees for maintaining an optimal sub-forest using a Genetic-Algorithm-based technique. The method selects a small sub-forest from a large forest. An optimum number of diverse and accurate trees could improve the absolute accuracy besides its computational advantage, as many trees could be less valuable in a forest [46].

2.7. Metrics

The present article has provided % accuracies (Correctly Classified Instances—CCI (%)). Some other performance measures, which are F1 scores and MCC values, have also been calculated for all the results in this study. However, CCI values (%) have been used in most of the figures in this study because CCI is the most used performance metric in similar studies in the literature.
If the elements of the confusion matrix are Tp (True Positives), Tn (True Negatives), Fp (False Positives), and Fn (False Negatives), accuracy is defined as [47]:
Accuracy = Tp + Tn n + + n = Tp + Tn Tp + Tn + Fp + Fn
The F1 value is the harmonic mean of precision and sensitivity. Precision (Positive Predictive Value) is the ratio of true positives to all positives. Sensitivity (recall/true positive rate) is the ratio of positives to all true positives [47,48].
F 1   Score = 2 . Tp 2 . Tp + Fp + Fn = 2 Precision . Sensitivity Precision + Sensitivity
Matthew’s correlation coefficient (MCC) is a method of contingency matrix that calculates the Pearson product–moment correlation coefficient between actual and predicted values [47]:
MCC = Tp . Tn Fp . Fn Tp + Fp . Tp + Fn . Tn + Fp . ( Tn + Fn )
The F1 score differs from MCC, and accuracy performance measures for it are independent of True Negatives; it is also not symmetric for class swapping [47].

3. Results

This study is divided into two parts: the segmentation pre-study and cross-comparisons of 3D VAD emotions, segments, and classes. For the comparisons of the segmentations for emotion analysis, we used all valence, arousal, and dominance coordinates together for two class cross-comparisons in both parts of this study. Class comparisons are made for the class separation levels in this study, which are given in Table 1.

3.1. Segmentation Pre-Study

In this section of this study, the 30, 20, 15, 10, 5, 3, 2, and 1 s non-overlapping windowing method is used for the segmentation of EEG signals corresponding to Segments I, II, III, IV, V, VI, VII, and VIII, respectively, as given in Table 2. Feature sets are generated for all segmented data, and then these pieces are assembled to form the related segments’ sub-dataset for further use.
Correctly Classified Instances (CCI %) for VAD space binary emotion comparisons for evaluating the effect of eight segments are given in Figure 6. Accuracies with the balanced data with SMOTE are given in Figure 7. Only three classifiers (SMO, IBk, and Random Forest) are used to classify in this part of this study. Similarly, this part of this study considers only 4–6 class separation levels.

3.2. Cross-Comparisons of 3D VAD Emotions, Segments, and Classes

In the second part of this study, Segments III, IV, and V, corresponding to 15, 10, and 5 s windowing, are used for emotion comparisons. Here, Segments III, IV, and V mean that the 60 s EEG signal is divided into 4, 6, and 12 segments/pieces with 15, 10, and 5 s windowing, respectively. N means a complete 60 s signal processed without segmentation (Table 2).
In total, 5, 4–6, and 3–7 class separation levels are used for A&G, B&H, and D&F emotion comparisons. Furthermore, 5, 4.5–5.5, and 4–6 class separation levels are used for C&E emotion comparisons (Table 1).
The data balancing method with a SMOTE filter is also applied to the minority classes in this part of this study. Synthetic samples are generated for the minority class with oversampling. All minority classes for the binary cross-comparisons are converged to the majority class by increasing the applied percentages 5 by 5 until reaching a closer data size to the majority class (here, the aim is to achieve close to, but not to exceed, class size). Applied percentages for oversampling for data balancing are given in Figure 8.
All class sizes are not given here. However, some examples are given: for D&F SegD 4–6 separation, high and low classes have 288 and 150 samples, respectively, and after applying data balancing with a percentage of 90%, low-class samples increase to 285; for C&E SegD 5 separation, high and low classes have 480 and 348 samples, respectively, and after applying data balancing with a percentage of 35%, low-class increases to 469 samples.
Spider web graphics of VAD space emotion comparisons (A&G, B&H, D&F, and C&E) and the segment and class separation level comparisons of the emotions are given in Figure 9, Figure 10, Figure 11 and Figure 12, respectively. This part of this study uses SMO, IBk, FURIA, Random Forest, Optimized Forest, Bagging, Random Committee, and Random Subspace classifiers. The % accuracies (CCI %) obtained with all these classifiers are given in the figures below in this part of this study.
Both the filtered and unfiltered results are obtained for the comparison of emotions. All the related figures have unfiltered dataset results on the left sides, indicated with the letter, and filtered versions (balanced with SMOTE) on the right sides of the figures are indicated with the letter b.
CCI (%) values for A&G comparisons of VAD space for Segments III, IV, and V are given in Figure 9a,b for 5, 4–6, and 3–7 class separations to evaluate the emotions. Figure 9b is the SMOTE-applied version of the classification.
CCI (%) values of B&H comparisons of VAD space for Segments III, IV, and V are given in Figure 10a,b for 5, 4–6, and 3–7 class separations. Figure 10b is the SMOTE-applied version of the classification.
CCI (%) values of D&F comparisons of VAD space for Segments III, IV, and V are given in Figure 11a,b for 5, 4–6, and 3–7 class separations.
CCI (%) values for C&E comparisons of VAD space for segments III, IV, and V are given in Figure 12a,b. Here, 5, 4.5–5.5, and 4–6 class separations are used, which are different from the above comparisons in Figure 9, Figure 10 and Figure 11, which use 5, 4–6, and 3–7 class separations. The 3–7 class separation is not used here because it has so few sample sizes for minority classes for non-segmented data. As such, the 4.5–5.5 class separation level has been added only for comparisons evaluating C&E area emotions.
CCI, MCC values, and F1 scores (%) for the comparisons of VAD space and the comparison of performance metrics are given in Figure 13. Only Segment V and the 4–6 classification are used in the figure. The results are given for the Random Forest, SMO, IBk, FURIA, and Random Committee classifiers.
CCI, MCC values, and F1 scores (%) for the comparisons of VAD space and the comparison of performance metrics (data balanced with SMOTE) (Segment V, 4–6 classification) are given in Figure 14 for the Random Forest, SMO, IBk, FURIA, and Random Committee classifiers.

4. Discussion

In the literature, we generally see studies comparing emotions with binary classifications, such as high and low levels of valence and high and low levels of arousal, which divide the VA space into two pieces. Some studies also consider two-dimensional VA space, which divides the space into four pieces.
There are many studies on emotion analysis in the literature. However, these studies generally consider VA space. Little work seems to focus on the 3D space of valence, arousal, and dominance, which divides the VAD space into eight pieces weighing the dominance besides valence and arousal dimensions. Therefore, in this study, we especially wanted to focus on 3D VAD space to classify, compare, and evaluate emotions.
Similar to these comparisons of segmentations for emotion analysis using handcrafted feature sets that the researcher manually considers, we also see valence and arousal dimensions in the literature. In this context, we also wanted to focus on 3D VAD space to study the segmentations of EEG signals for emotion analysis.
There are few studies for emotion analysis using the SMOTE method. Because of the unbalanced classes for our comparisons, we also wanted to use this filter (oversampling of the minority class) to perform data balancing besides the unfiltered data.
This study is divided into two parts: a segmentation pre-study and cross-comparisons of 3D VAD emotions, segments, and classes. We used all valence, arousal, and dominance coordinates together for two class cross-comparisons in both studies. In the first part of this study, the 30, 20, 15, 10, 5, 3, 2, and 1 s non-overlapping windowing method is used for the segmentation of EEG signals, which correspond to Segments I, II, III, IV, V, VI, VII, and VIII, respectively (Table 2). The % CCI values are calculated for VAD space binary emotion comparisons (A&G, B&H, C&E, and D&F) for evaluating the segmentation effect. Eight segments (Segments I, II, III, IV, V, VI, VII, and VIII) are given in Figure 6 and Figure 7 (data-balanced version). Only three classifiers (SMO, IBk, and Random Forest) are used for the classification in this part of this study. Similarly, this part of this study considers only the 4–6 class separation level.
Ahmed et al. [49] studied a novel method (InvBase) for emotion classification after baseline removal. The results show that baseline removal gave higher accuracy with simple features, requiring less computation time. They also concluded in their study that the emotion features were dominant in a range of 1 s to 12 s time slots. Similarly to that study, we have subtracted the first three seconds of the preprocessed EEG data from the signal, and the remaining 60 s signal is used in this study.
From Figure 6, it can be seen that segmentation increases accuracy. The results show that accuracies generally grow to some extent as the segmentation increases; however, after some increase, they give similar results or even begin to decrease at the end parts for every three classifiers (SMO, IBk, and Random Forest). Segments III, IV, and V, which correspond to 15, 10, and 5 s windowing, are chosen to use in the second sub-study, which has more complex comparisons with the class separation levels, classifiers, and data balancing.
In the second part of this study, spider web diagrams compare the results for the binary cross-comparisons of 3D VAD emotions, segments, class separations, classifiers, and data balancing. Lines are drawn to ensure a more accessible and precise understanding of the results and the comparisons. Data points on the lines indicate the classifiers’ results (% accuracies). This part of this study uses SMO, IBk, FURIA, Random Forest, Optimized Forest, Bagging, Random Committee, and Random Subspace classifiers. Comparisons of segments III, IV, and V can be seen more clearly in this sub-study. Different colored lines, including data points of related classifiers in the figures, indicate the segmentations. Blue, green, and purple colored lines correspond to Segments III, IV, and V, respectively.
More opening parts in the middles of the figures mean the segment separations are closer. SMOTE versions of each figure in this part of this study show that those openings in the middle parts are increased compared to the unbalanced ones (Figure 9b, Figure 10b, Figure 11b, and Figure 12b). As such, the results of the segment classifications are closer to each other.
Class separations 5, 4–6, and 3–7 are used for the A&G, B&H, and D&F emotion comparisons. Class separations 5, 4.5–5.5, and 4–6 are used for the C&E emotion comparisons. Sizes of the same colored lines (segments) indicate the class separations in the figures. Balanced data results (SMOTE versions) use the same colors and sizes for segments and classes. Class separations with the related segments are stated on the right sides of the unbalanced results in Figure 9a, Figure 10a, Figure 11a, and Figure 12a.
It can be easily seen from those lines that data accuracy increases gradually as the class separations diverge from the midpoint (5) of the VAD space for most of the classifications. For instance, the 3–7 class separation has higher accuracy values than the 4–6 separation, and the 4–6 is bigger than the 5 separation for most classifications.
In this study, classes are set by considering the high and low values of the valence, arousal, and dominance scores. Class placements on VAD space are shown in Figure 15, in which the emotions are placed in the eight regions. To simplify understanding, the valence dimension is considered high and low (not as positive or negative), just like this study’s other dimensions (valence and dominance).
Bălan et al. [3] considered six basic emotions in their study: anger: low V, high A, D [6; 7]; joy: high V, high A, D [6; 7]; surprise: high V, high A, D [4; 5]; disgust: low V, high A, D [5; 6]; fear: low V, high A, D [3; 4]; and sadness: low V, low A, D [3; 4] for a 1–9 scale, as low indicates [1; 5] and high indicates [5; 9] intervals.
Looking mainly at Figure 15 and considering the 3D emotion space figures in [50] and [51] and the definitions in [3], we generally considered and inferred the eight three-dimensional emotion regions as:
  • A: Happy, joyful, useful, powerful, influential, friendly, excited, and inspired
  • B: Relaxed, leisurely, untroubled, and satisfied
  • C: Disdainful and unconcerned
  • D: Angry, hostile, cruel, insolent, hate, and disgusted
  • E: Impressed, surprised, thankful, and curious
  • F: Protected, humble, and reverent
  • G: Sad, depressed, bored, lonely, feeble, discouraged, and discontented
  • H: Fearful, frustrated, helpless, pain, humiliated, embarrassed, guilty, and confused
According to this information, we can state four cross-comparisons of VAD space in a briefer way as:
  • A&G (Happy, joyful, excited, and powerful and sad, depressed, and feeble)
  • B&H (Relaxed, leisurely, and satisfied and fear, pain, and guilt)
  • D&F (Angry, hate, and disgusted and protected, humble, and reverent)
  • C&E (Disdainful and unconcerned and impressed, surprised, and thankful).
Glancing at this inferred information, for the emotion regions with high valence values, which are A and B (with high dominance values) and E and F (with low dominance values), generally, we see positive emotions. On the other hand, for emotion regions with low valence values, which are G and H (with low dominance values) and C and D (with high dominance values), generally, we see negative emotions in accordance with the literature.
The best CCI, MCC values, and F1 scores (%) among Segments III, IV, and V for the cross-comparisons of VAD space are given in Table 3 for the unbalanced data and in Table 4 for the balanced data. It can be easily seen that Segment V gave mostly the best percentages for all CCI, MCC values, and F1 scores among Segments III, IV, and V. Besides these results, the RSS classifier gave the best percentages by a considerable amount with the unbalanced data, while the OF classifier gave similar results with the balanced data among the mentioned segments.
The best accuracies of the cross-comparisons for each class separation level can be easily seen in both tables. The SMOTE method considerably increases the percentages of the performance metrics, which can be seen in Table 4. As the class separations move away from the middle point (5), the accuracy values increase in general, which can also be seen in both of the tables.
For the A&G classes, which compare happy, joyful, excited, and powerful and sad, depressed, and feeble emotions, for the class separation levels 5 and 4–6, accuracy rates of 88.62% and 91.45% (CCI) are obtained, respectively, with the RF classifier and Seg V. For the 3–7 class separation level, an accuracy rate of 95.58% (CCI) is obtained for the SMO classifier according to the data-balanced results.
For the B&H classes, which compare relaxed, leisurely, and satisfied and fear, pain, and guilt emotions, for the class separation levels 5 and 4–6, accuracy rates of 82.60% and 93.37% (CCI) are obtained, respectively, with the OF classifier and Seg V. For the 3–7 class separation level, an accuracy rate of 100% (CCI) is obtained with both RF and OF classifiers according to the data-balanced results.
For the D&F classes, which compare angry, hate, and disgusted and protected, humble, and reverent emotions, for the class separation level 5, an accuracy rate of 89.94% (CCI) is obtained with the OF classifier and Seg V, and for the class separation level 4–6, an accuracy rate of 94.76% (CCI) is obtained for all IBk, SMO (Seg IV), and RC (Seg III and V) classifiers. For the class separation level 3–7, an accuracy rate of 95.05% (CCI) is obtained for the SMO classifier and Seg V according to the data-balanced results.
For the C&E classes, which compare disdainful and unconcerned and impressed, surprised, and thankful emotions, for the class separation level 5, an accuracy rate of 90.31% (CCI) is obtained with the OF classifier and Seg V, and for the class separation level 4–6, an accuracy rate of 97.92% (CCI) is obtained for the RC classifier and Seg III. For the class separation level 4.5–5.5, an accuracy rate of 97.92% (CCI) is obtained for the IBk classifier and Seg III according to the data-balanced results.
The best accuracy result in this study is for the C&E comparison for the unbalanced data (98.94% CCI with IBk classifier and Seg III for the 4.5–5.5 class separation). For the C&E comparison and the class separation level 4–6, an accuracy rate of 96.97% (CCI) is obtained with the FURIA classifier and Seg IV according to the unbalanced data results. For the B&H comparison and the class separation level 3–7, an accuracy rate of 96.03% (CCI) is obtained with the RC classifier and Seg V according to the unbalanced data results.
CCI, MCC values, and F1 scores (%) for the comparisons of VAD space for Segment V and the 4–6 classification are given in Figure 13 and Figure 14 (data-balanced version) for Random Forest, SMO, IBk, FURIA, and Random Committee classifiers. CCI (%) values generally gave close results to F1 scores (%); however, MCC values (%) gave relatively lower results, which can be seen in Figure 13 and Figure 14 and in Table 3 and Table 4.
Studies in the literature on emotion evaluation show that emotional states can be evaluated with discrete or dimensional emotion models. Dimensional models include Valence Arousal (VA) or Valence Arousal and Dominance (VAD) spaces. A comparison of our study with some studies that have a dominance dimension is given in Table 5. We can see that most of the studies in the literature use only VA due to the ease of differentiation; however, very few studies use VAD space like ours. Similarly, glancing over the studies in the literature for the segmentation of emotion analysis with handcrafted features, we also see that VA space is used. Considering those two issues, we especially wanted to focus on VAD space to compare and evaluate the emotional states and segmentations of EEG signals and segment comparisons in terms of emotion analysis. Based on this, we think this study will contribute to the literature.
In the literature, studies generally do not use so many comparison factors. In this context, the present study is critical because it offers an overall comprehensive and analytical approach with several different steps and complete comparisons applied for various factors for comparisons, such as binary cross-comparisons and evaluations of eight emotional states; comparisons of selected segments; comparisons of class separation levels (5, 4–6, and 3–7); and comparisons of unbalanced and balanced data with the SMOTE method. Giving detailed comparison results with the same or similar conditions is crucial. For instance, for these segmentation studies and binary comparisons of eight emotional states as well as other comparison factors, all the conditions are the same, including the same dataset, the same preprocessed EEG signal, the same channels, the same feature sets, the same class separations, the same classifiers, etc. As such, the results could be considered to show the differences in a more convenient and reliable way.
In our study, we have obtained high-accuracy results that can be seen directly in the spider web diagrams in the second part of this study and in Table 3 and Table 4. The best accuracy result in this study for unbalanced data is for the C&E comparison (Seg III and 4.5–5.5 class separation), for which an accuracy rate of 98.94% CCI (F1: 98.9% and MCC: 97.9%) is obtained with the IBk classifier, which is greater than the compared studies from the literature that have a dominance dimension besides valence and arousal, as shown in Table 5.
The overall analysis results conducted with VAD space show that accuracy values obtained from C&E comparisons were generally higher than all the other cross-comparisons for unbalanced data. From all the results in Figure 6 and Figure 7 (data-balanced version) in the first part of this study, accuracy rates generally seem to decrease gradually in the order of C&E, D&F, B&H, and A&G comparisons, respectively, for 4–6 class separations.
Besides these results, if we look at all the results and Figure 9a, Figure 10a, Figure 11a, and Figure 12a in the second part of this study, the accuracy rates generally seem to gradually decrease in the order of C&E, B&H, D&F, and A&G comparisons, respectively. Here, C&E and B&H seem similar at a glance; however, we should remember that C&E comparisons do not include the 3–7 class separation. As such, a middle-sized class separation (4–6) in B&H and other comparisons correspond to a thick-sized class separation (4–6) in C&E, as seen in Figure 12a,b. Data-balanced versions mostly seem to outperform the unbalanced versions in this study, as seen in the figures of the second part of this study.
A possible limitation of this study is that class-level separations could decrease the balance among data classes. Oversampling and down-sampling methods could not give sufficient repairments of the proposed methods. For instance, if the data have more unbalanced classes, this would require a more significant percentage of oversampling of the minority class or a more substantial percentage of down-sampling of the majority class.
We plan to apply the analysis to some other databases in future work. We plan to compare the emotions as changing parameters and emotions in VA/VAD spaces. We are also planning to try some other classifiers. In future applications, the obtained methods and information could be used with modifications for clinical research on emotion analysis. For example, this might involve evaluating a group’s emotions that appear due to stimuli applied according to the experimental design. Another example is BCI technologies, which have important contributions to medical applications. For example, BCI devices, such as wheelchairs, robotic arms, and drones, can easily turn left and right and move forward based on left and right hand motor imagery. Neural activities related to inhibition (P300) can be applied as a stop command [52]. Chikara et al. [52] developed a model for identifying the neural activities of human inhibition using a phase-locking value method. Their results imply that these activities can be used as a stop command in BCI technologies and to identify the symptoms of attention deficit hyperactivity disorder patients in clinical research. Furthermore, as future applications, the classification and comparison of emotions could be used in many areas, including smart health technologies, psychology, robot technologies, simulation systems, game technologies, smart homes, driver systems, etc.

5. Conclusions

A comprehensive analytical approach is implemented with two sub-studies: first, a segmentation study, and second, an emotion comparison and evaluation in addition to comparisons of selected segments in a more detailed way. This was performed in addition to comparing the effect of class separation levels and classification accuracies for both unbalanced and balanced feature sets. The results obtained from the second part of this study are compared, evaluated, and commented on with this information. Eight emotional states are considered and inferred for VAD space.
This study is critical because it compares segmentation, class separation levels, and the data balancing effect with SMOTE. The segmentation effect is significant because it is conducted with 3D VAD space with handcrafted features, including the dominance dimension in the evaluation, while also considering the impact of each of the three dimensions together. In conclusion, it is emphasized that emotion evaluation can be conducted with EEG signals successfully. It is believed that because facial expressions or speech could be misleading in some conditions that could change from person to person, a physiological signal obtained from the brain may give closer results to the emerging emotion. Consequently, the results emphasize that emotional states occur due to stimuli, such as music videos, which can change the physiological signals and can be predicted by analyzing the EEG data.

Author Contributions

Conceptualization, Ü.I. and A.G.; methodology, Ü.I. and A.G.; software, Ü.I. and T.B.; investigation, Ü.I., A.G. and T.B.; writing—original draft preparation, Ü.I.; writing—review and editing, A.G. and T.B.; supervision, A.G.; project administration, A.G. and Ü.I.; funding acquisition, TÜBİTAK. All authors have read and agreed to the published version of the manuscript.

Funding

This research is a part of an ongoing project related to a PhD thesis funded by TÜBİTAK (The Scientific and Technological Research Council of Türkiye), 1002A Short Term R&D Funding Program (Project No: 122E592, Project Title: Transfer Entropy, Moments & Artificial Intelligence Methods based Emotion Analysis).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The DEAP dataset used in this study is available at: http://www.eecs.qmul.ac.uk/mmv/datasets/deap/ and can be accessed upon approval (accessed on 14 September 2021).

Acknowledgments

We would like to express our gratitude to TÜBİTAK (Project No: 122E592) for the funding support and the DEAP dataset authorities for providing access to the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Paul, S.; Mazumder, A.; Ghosh, P.; Tibarewala, D.N.; Vimalarani, G. EEG Based Emotion Recognition System using MFDFA as Feature Extractor. In Proceedings of the International Conference on Robotics, Automation, Control and Embedded Systems–RACE 2015, Hindustan University, Chennai, India, 18–20 February 2015. [Google Scholar]
  2. Sternberg, R.J.; Sternberg, K. Cognitive Psychology, 6th ed.; Wadsworth Cengage Learning: Belmont, CA, USA, 2012. [Google Scholar]
  3. Bălan, O.; Moldoveanu, A.; Leordeanu, M. A Machine Learning Approach to Automatic Phobia Therapy with Virtual Reality. In Modern Approaches to Augmentation of Brain Function; Opris, I., Lebedev, M.A., Casanova, M.F., Eds.; Springer: Cham, Switzerland, 2021; pp. 607–636. [Google Scholar] [CrossRef]
  4. Zhang, J.; Yin, Z.; Chen, P.; Nichele, S. Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review. Inf. Fusion 2020, 59, 103–126. [Google Scholar] [CrossRef]
  5. Chikara, R.K.; Chang, E.C.; Lu, Y.C.; Lin, D.S.; Lin, C.T.; Ko, L.W. Monetary Reward and Punishment to Response Inhibition Modulate Activation and Synchronization within the Inhibitory Brain Network. Front. Hum. Neurosci. 2018, 12, 27. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Yu, W.; Ding, S.; Yue, Z.; Yang, S. Emotion Recognition from Facial Expressions and Contactless Heart Rate Using Knowledge Graph. In Proceedings of the IEEE International Conference on Knowledge Graph (ICKG), Nanjing, China, 9–11 August 2020. [Google Scholar] [CrossRef]
  7. Shabbir, J.; Anwer, T. Artificial Intelligence and its Role in Near Future. arXiv 2015, arXiv:1804.01396. [Google Scholar]
  8. Doma, V.; Pirouz, M. A comparative analysis of machine learning methods for emotion recognition using EEG and peripheral physiological signals. J. Big Data 2020, 7, 18. [Google Scholar] [CrossRef] [Green Version]
  9. Kumar, G.S.S.; Arun, A.; Sampathila, N.; Vinoth, R. Machine Learning Models for Classification of Human Emotions Using Multivariate Brain Signals. Computers 2022, 11, 152. [Google Scholar] [CrossRef]
  10. Abdulrahman, A.; Baykara, M.; Alakus, T.B. A Novel Approach for Emotion Recognition Based on EEG Signal Using Deep Learning. Appl. Sci. 2022, 12, 10028. [Google Scholar] [CrossRef]
  11. Alakuş, T.B.; Türkoğlu, İ. EEG Based Emotion Analysis Systems. In Türkiye Bilişim Vakfı Bilgisayar Bilimleri ve Mühendisliği Dergisi; Türkiye Bilişim Vakfı: Iistanbul, Turkey, 2018; Volume 15. [Google Scholar]
  12. Verma, G.K.; Tiwary, U.S. Affect representation and recognition in 3D continuous valence–arousal–dominance space. Multimed Tools Appl. 2017, 76, 2159–2183. [Google Scholar] [CrossRef]
  13. Verma, G.K.; Tiwary, U.S. Multimodal fusion framework: A multiresolution approach for emotion classification and recognition from physiological signals. NeuroImage 2014, 102, 162–172. [Google Scholar] [CrossRef]
  14. Russell, J.A.; Mehrabian, A. Evidence for A Three-Factor Theory of Emotions. J. Res. Personal. 1977, 11, 273–294. [Google Scholar] [CrossRef]
  15. Guendil, Z.; Lachiri, Z.; Maaoui, C. Computational framework for emotional VAD prediction using regularized Extreme Learning Machine. Int. J. Multimed Inf. Retr. 2017, 6, 251–261. [Google Scholar] [CrossRef]
  16. Özel, P.; Akan, A.; Yılmaz, B. Synchrosqueezing transform based feature extraction from EEG signals for emotional state prediction. Biomed. Signal Process. Control. 2019, 52, 152–161. [Google Scholar] [CrossRef]
  17. Koelstra, S.; Muehl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  18. Rhif, M.; Abbes, A.B.; Farah, I.R.; Martínez, B.; Sang, Y. Wavelet Transform Application for/in Non-Stationary Time-Series Analysis: A Review. Appl. Sci. 2019, 9, 1345. [Google Scholar] [CrossRef] [Green Version]
  19. Sifuzzaman, M.; Islam, M.R.; Ali, M.Z. Application of Wavelet Transform and its Advantages Compared to Fourier Transform. J. Phys. Sci. 2009, 13, 121–134. [Google Scholar]
  20. Öner, İ.V.; Yeşilyurt, M.K.; Yılmaz, E.Ç. Wavelet Analiz Tekniği Ve Uygulama Alanları. Ordu Univ. J. Sci. Tech. 2017, 7, 42–56. [Google Scholar]
  21. Frank, E.; Hall, M.A.; Witten, I.H. The WEKA Workbench. Online Appendix for “Data Mining: Practical Machine Learning Tools and Techniques (Version: 3.9.6)”, 4th ed.; Morgan Kaufmann: Burlington, MA, USA, 2016. [Google Scholar]
  22. Watkins, J.C. An Introduction to the Science of Statistics: From Theory to Implementation, Preliminary, ed.; 2016. Available online: http://www.freetechbooks.com/an-introduction-to-the-science-of-statistics-from-theory-to-implementation-preliminary-edition-t1277.html (accessed on 19 April 2023).
  23. Gupta, S.C.; Kapoor, V.K. Fundamentals of Mathematical Statistics (A Modern Approach), 10th ed.; Sultan Chand & Sons, Educational Publishers: New Delhi, India, 2000. [Google Scholar]
  24. Henriques, T.; Ribeiro, M.; Teixeira, A.; Castro, L.; Antunes, L.; Costa-Santos, C. Nonlinear Methods Most Applied to Heart-Rate Time Series: A Review. Entropy 2020, 22, 309. [Google Scholar] [CrossRef] [Green Version]
  25. Safi, M.S.; Safi, S.M.M. Early detection of Alzheimer’s disease from EEG signals using Hjorth parameters. Biomed. Signal Process. Control. 2021, 65, 102338. [Google Scholar] [CrossRef]
  26. Rihaczek, A.W. Signal Energy Distribution in Time and Frequency. IEEE Trans. Inf. Theory 1968, 14, 369–374. [Google Scholar] [CrossRef]
  27. Rafiuddin, N.; Khan, Y.U.; Farooq, O. Feature Extraction and Classification of EEG for Automatic Seizure Detection. In Proceedings of the International Conference on Multimedia, Signal Processing and Communication Technologies, Aligarh, India, 17–19 December 2011. [Google Scholar] [CrossRef]
  28. Rahman, A.; Hossain, F.; Hossain, M.; Ahmed, R. Employing PCA and t-statistical approach for feature extraction and classification of emotion from multichannel EEG signal. Egypt. Inform. J. 2020, 21, 23–35. [Google Scholar] [CrossRef]
  29. Thongpanja, S.; Phinyomark, A.; Limsakul, C.; Phukpattaranont, P. Application of Mean and Median Frequency Methods for Identification of Human Joint Angles Using EMG Signal. In Information Science and Applications, Lecture Notes in Electrical Engineering; Kim, K.J., Ed.; Springer: Berlin/Heidelberg, Germany, 2015; p. 339. [Google Scholar] [CrossRef]
  30. Minkowski, L.; Mai, K.V.; Gurve, D. Feature Extraction to Identify Depression and Anxiety Based on EEG. In Proceedings of the 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual Conference, 31 October–4 November 2021. [Google Scholar] [CrossRef]
  31. Mohamad Ismail, M.R.; Lam, C.K.; Cheng, F.Y.; Sundaraj, K.; Rahiman, M.H.F. Analysis and classification of muscle activity during biceps exercise using MMG signals. In Proceedings of the 2018 4th International Conference on Electrical, Electronics and System Engineering (ICEESE), Kuala Lumpur, Malaysia, 8–9 November 2018; pp. 12–16. [Google Scholar]
  32. Torun, Y.; Pazarci, Ö. Parametric Power Spectral Density Estimation-Based Breakthrough Detection for Orthopedic Bone Drilling with Acoustic Emission Signal Analysis. Acoust. Aust. 2020, 48, 221–231. [Google Scholar] [CrossRef]
  33. Li, X.; Song, D.; Zhang, P.; Zhang, Y.; Hou, Y.; Hu, B. Exploring EEG Features in Cross-Subject Emotion Recognition. Front. Neurosci. 2018, 12, 162. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Vijayakumar, S.; Flynn, R.; Murray, N. A Comparative Study of Machine Learning Techniques for Emotion Recognition from Peripheral Physiological Signals. In Proceedings of the 2020 31st Irish Signals and Systems Conference (ISSC), Letterkenny, Ireland, 11–12 June 2020. [Google Scholar]
  35. Wijayanto, I.; Hadiyoso, S.; Aulia, S.; Atmojo, B.S. Detecting Ictal and Interictal Condition of EEG Signal using Higuchi Fractal Dimension and Support Vector Machine. J. Phys. Conf. Ser. 2020, 1577, 012016. [Google Scholar] [CrossRef]
  36. Sun, Q.; Wang, Q.; Ji, B.; Wu, W.; Huang, W.; Wang, C. The Cardiodynamicsgram Based Early Detection of Myocardial Ischemia Using the Lempel-Ziv Complexity. IEEE Access 2020, 8, 207894–207904. [Google Scholar] [CrossRef]
  37. Šapina, M.; Karmakar, C.K.; Kramari, K.; Kosmider, M.; Garcin, M.; Brdari, D.; Milas, K.; Yearwood, J. Lempel-Ziv complexity of the pNN x statistics an application to neonatal stress. Chaos Solitons Fractals 2021, 146, 110703. [Google Scholar] [CrossRef]
  38. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  39. Rahman, M.; Sarkar, A.K.; Hossain, A.; Moni, M.A. EEG-based emotion analysis using non-linear features and ensemble learning approaches. Expert Syst. Appl. 2022, 207, 118025. [Google Scholar] [CrossRef]
  40. Vijayarani, S.; Muthulakshmi, M. Comparative Analysis of Bayes and Lazy Classification Algorithms. Int. J. Adv. Res. Comput. Commun. Eng. 2013, 2, 3118–3124. [Google Scholar]
  41. Platt, J.C. Sequential Minimal Optimization: A Fast Algorithm for Training Support Vector Machines; Microsoft Research, Technical Report MSR-TR-98-14; Microsoft: Redmond, WA, USA, 1998. [Google Scholar]
  42. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  43. Fawagreh, K.; Gaber, M.M.; Elyan, E. Random forests: From early developments to recent advancements. Syst. Sci. Control. Eng. 2014, 2, 602–609. [Google Scholar] [CrossRef] [Green Version]
  44. Hühn, J.; Hüllermeier, E. FURIA: An Algorithm for Unordered Fuzzy Rule Induction. Data Min. Knowl. Disc. 2009, 19, 293–319. [Google Scholar] [CrossRef] [Green Version]
  45. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  46. Adnan, N.; Islam, Z. Optimizing the Number of Trees in a Decision Forest to Discover a Subforest with High Ensemble Accuracy using a Genetic Algorithm. Knowl. Based Syst. 2016, 110, 86–97. [Google Scholar] [CrossRef]
  47. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [Green Version]
  48. Lipton, Z.C.; Elkan, C.; Naryanaswamy, B. Thresholding Classiffers to Maximize F1 Score. arXiv 2014, arXiv:1402.1892. Available online: https://deepai.org/publication/thresholding-classifiers-to-maximize-f1-score (accessed on 14 September 2022).
  49. Ahmed, M.Z.I.; Sinha, N.; Ghaderpour, E.; Phadikar, S.; Ghosh, R. A Novel Baseline Removal Paradigm for Subject-Independent Features in Emotion Classification Using EEG. Bioengineering 2023, 10, 54. [Google Scholar] [CrossRef] [PubMed]
  50. Almanza-Conejo, O.; Almanza-Ojeda, D.L.; Contreras-Hernandez, J.L.; Ibarra-Manzano, M.A. Emotion recognition in EEG signals using the continuous wavelet transform and CNNs. Neural Comput. Appl. 2023, 35, 1409–1422. [Google Scholar] [CrossRef]
  51. Shu, L.; Xie, J.; Yang, M.; Li, Z.; Li, Z.; Liao, D.; Xu, X.; Yang, X. A Review of Emotion Recognition Using Physiological Signals. Sensors 2018, 18, 2074. [Google Scholar] [CrossRef] [Green Version]
  52. Chikara, R.K.; Ko, L.-W. Neural Activities Classification of Human Inhibitory Control Using Hierarchical Model. Sensors 2019, 19, 3791. [Google Scholar] [CrossRef] [Green Version]
Figure 1. VAD space of emotions with class placements (VAD space is drawn by using the data in [14] and by taking the pleasure plane as valence).
Figure 1. VAD space of emotions with class placements (VAD space is drawn by using the data in [14] and by taking the pleasure plane as valence).
Diagnostics 13 02141 g001
Figure 2. 32 Channel placement of EEG according to the 10–20 electrode placement system (numbers indicate the channel numbers of EEG signals from the DEAP dataset).
Figure 2. 32 Channel placement of EEG according to the 10–20 electrode placement system (numbers indicate the channel numbers of EEG signals from the DEAP dataset).
Diagnostics 13 02141 g002
Figure 3. Flow chart for EEG signals analysis.
Figure 3. Flow chart for EEG signals analysis.
Diagnostics 13 02141 g003
Figure 4. Flow chart of the overall method.
Figure 4. Flow chart of the overall method.
Diagnostics 13 02141 g004
Figure 5. Class placements on VAD space (here, a 0-midpoint on VAD space indicates 5-midpoint for the 1–9 scale, and values below and above this scale also indicate the same logical separation related to the midpoint).
Figure 5. Class placements on VAD space (here, a 0-midpoint on VAD space indicates 5-midpoint for the 1–9 scale, and values below and above this scale also indicate the same logical separation related to the midpoint).
Diagnostics 13 02141 g005
Figure 6. CCI (%) values for VAD space for SMO, IBk, and Random Forest Classifiers for 4–6 classes for all segments (N: No Segment and Segments I–VIII).
Figure 6. CCI (%) values for VAD space for SMO, IBk, and Random Forest Classifiers for 4–6 classes for all segments (N: No Segment and Segments I–VIII).
Diagnostics 13 02141 g006
Figure 7. Balanced data CCI (%) values for VAD space for SMO, IBk, and Random Forest Classifiers for 4–6 classes for all segments (N: No Segment and Segments I–VIII).
Figure 7. Balanced data CCI (%) values for VAD space for SMO, IBk, and Random Forest Classifiers for 4–6 classes for all segments (N: No Segment and Segments I–VIII).
Diagnostics 13 02141 g007
Figure 8. Data balancing with SMOTE (applied percentages for oversampling).
Figure 8. Data balancing with SMOTE (applied percentages for oversampling).
Diagnostics 13 02141 g008
Figure 9. CCI (%) values for A&G comparisons of VAD space for Segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Figure 9. CCI (%) values for A&G comparisons of VAD space for Segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Diagnostics 13 02141 g009
Figure 10. CCI (%) values for B&H comparisons of VAD space for Segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Figure 10. CCI (%) values for B&H comparisons of VAD space for Segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Diagnostics 13 02141 g010
Figure 11. CCI (%) values for D&F comparisons of VAD space for segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Figure 11. CCI (%) values for D&F comparisons of VAD space for segments III, IV, and V for 5, 4–6, and 3–7 classifications by using: (a) Unbalanced data; (b) Balanced data.
Diagnostics 13 02141 g011
Figure 12. CCI (%) values for C&E comparisons of VAD space for segments III, IV, V for 5, 4.5–5.5, and 4–6 classifications by using: (a) Unbalanced data; (b) Balanced data.
Figure 12. CCI (%) values for C&E comparisons of VAD space for segments III, IV, V for 5, 4.5–5.5, and 4–6 classifications by using: (a) Unbalanced data; (b) Balanced data.
Diagnostics 13 02141 g012
Figure 13. CCI, MCC values, and F1 scores (%) for comparisons of VAD space (Segment V, 4–6 classification).
Figure 13. CCI, MCC values, and F1 scores (%) for comparisons of VAD space (Segment V, 4–6 classification).
Diagnostics 13 02141 g013
Figure 14. CCI, MCC values, and F1 scores (%) for the comparisons of VAD space (data balanced with SMOTE) (Segment V, 4–6 classification).
Figure 14. CCI, MCC values, and F1 scores (%) for the comparisons of VAD space (data balanced with SMOTE) (Segment V, 4–6 classification).
Diagnostics 13 02141 g014
Figure 15. VAD space of emotions with class placements (VAD space is drawn by using the data in [14] and taking the pleasure plane as valence). The class placements are marked on the figure.
Figure 15. VAD space of emotions with class placements (VAD space is drawn by using the data in [14] and taking the pleasure plane as valence). The class placements are marked on the figure.
Diagnostics 13 02141 g015
Table 1. Class separation levels.
Table 1. Class separation levels.
Class Separation LevelHigh (H) 1Low (L) 1
5≥5<5
4.5–5.5 2≥5.5≤4.5
4–6≥6≤4
3–7≥7≤3
1 For each 3 dimensions (valence, arousal, and dominance) 2 4.5–5.5 class separation level has only been used for C&E emotion comparisons.
Table 2. Segmentation of EEG signals.
Table 2. Segmentation of EEG signals.
Segment NameWindowing (Second)Segment/Piece
NNone1
Seg I302
Seg II203
Seg III154
Seg IV106
Seg V512
Seg VI320
Seg VII230
Seg VIII160
Table 3. Best CCI, MCC values, and F1 Scores (%) among Segments III, IV, and V for the cross-comparisons of VAD space.
Table 3. Best CCI, MCC values, and F1 Scores (%) among Segments III, IV, and V for the cross-comparisons of VAD space.
Sep. LevelClassCCI (%)F1 Score (%)MCC (%)
ValueClassifier *SegmentValueClassifierSegmentValueClassifierSegment
(5)A&G81.71RSSV80.5RSSV54.9RSSV
B&H81.98RSSV81.9RSSV63.8RSSV
D&F85.68RFV85.5RFV69.2RFV
C&E86.84RSSV86.8RSSV72.9RSSV
(4–6)A&G85.13RSSV84.3RSSV62.3RSSV
B&H91.01OFV90.9OFV80.7OFV
D&F92.35RFV92.3RFV82.9RFV
C&E96.97FURIAIV97.0FURIAIV93.9FURIAIV
(3–7)A&G90.08RSSV89.6RSSV72.2RSSV
B&H96.03RCV95.9RCV85.2RCV
D&F91.32RSSV91.4RSSV81.3RSSV
(4.5–5.5)C&E98.94IBkIII98.9IBkIII97.9IBkIII
* FURIA: Fuzzy Unordered Rule Induction Algorithm; IBk: Instance-Based Learning (it is a K-Nearest Neighbor classifier); OF: Optimized Forest; RC: Random Committee; RF: Random Forest; RSS: Random Subspace.
Table 4. Best CCI, MCC values, and F1 scores (%) among Segments III, IV, and V for the cross-comparisons of VAD space (balanced data with SMOTE).
Table 4. Best CCI, MCC values, and F1 scores (%) among Segments III, IV, and V for the cross-comparisons of VAD space (balanced data with SMOTE).
Sep. LevelClassCCI (%)F1 Score (%)MCC (%)
ValueClassifier *SegmentValueClassifierSegmentValueClassifierSegment
(5)A&G88.62RFV88.6RFIV, V77.4RFV
B&H82.60OFV82.6OFV65.2OFV
D&F89.94OFV89.9RF, OFV80.2OFV
C&E90.31OFV90.3OFV80.7OFV
(4–6)A&G91.45RFV91.4RFV83.1RFV
B&H93.37OFV93.4OFV90.8FURIAIII
D&F94.76IBk, RC, SMOIII, IV, V94.8IBk, RC, SMOIII, IV, V90.0IBkIV
C&E97.92RCIII97.9RCIII95.9RCIII
(3–7)A&G95.58SMOIV95.6SMOIV91.4SMOIV
B&H100.00RF, OFIII100.0RF, OFIII100.0RF, OFIII
D&F95.05SMOV95.0SMOV90.3SMOV
(4.5–5.5)C&E97.92IBkIII97.9IBkIII95.9IBkIII
* FURIA: Fuzzy Unordered Rule Induction Algorithm; IBk: Instance-Based Learning (it is a K-Nearest Neighbor classifier); OF: Optimized Forest; RC: Random Committee; RF: Random Forest; RSS: Random Subspace; SMO: Sequential Minimum Optimization.
Table 5. Comparison with some studies that use the dominance dimension.
Table 5. Comparison with some studies that use the dominance dimension.
StudyClassAccuracy 1
[12]3-class classification
(Low, medium, and high)
63.47 (V)
69.62 (A)
63.57 (D)
[15]2 classes73.43% (V)
72.65% (A)
69.3% (D)
[16]2D and 3D emotion models
4 emotional states (VA model)Best accuracy of 79.1%.
8 emotional states (VAD model)Highest accuracy of 93%
Our StudyBinary classification for cross-comparisons of 8 emotional states from VAD space (A&G; B&H; D&F; C&E) for unbalanced dataBest accuracy of 98.94% CCI
(F1: 98.9% and MCC: 97.9%) for C&E comparison (Seg III and 4.5–5.5 class separation)
Please see Table 3 for other results.
for balanced data with SMOTEBest accuracy of 100% CCI, F1 and MCC for B&H comparison (Seg III and 3–7 class separation)
Please see Table 4 for other results.
1 Valence: V; Arousal: A; Dominance: D.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Işık, Ü.; Güven, A.; Batbat, T. Evaluation of Emotions from Brain Signals on 3D VAD Space via Artificial Intelligence Techniques. Diagnostics 2023, 13, 2141. https://doi.org/10.3390/diagnostics13132141

AMA Style

Işık Ü, Güven A, Batbat T. Evaluation of Emotions from Brain Signals on 3D VAD Space via Artificial Intelligence Techniques. Diagnostics. 2023; 13(13):2141. https://doi.org/10.3390/diagnostics13132141

Chicago/Turabian Style

Işık, Ümran, Ayşegül Güven, and Turgay Batbat. 2023. "Evaluation of Emotions from Brain Signals on 3D VAD Space via Artificial Intelligence Techniques" Diagnostics 13, no. 13: 2141. https://doi.org/10.3390/diagnostics13132141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop