Next Article in Journal
Five-State Extended Kalman Filter for Estimation of Speed over Ground (SOG), Course over Ground (COG) and Course Rate of Unmanned Surface Vehicles (USVs): Experimental Results
Next Article in Special Issue
Sensors and Artificial Intelligence Methods and Algorithms for Human–Computer Intelligent Interaction: A Systematic Mapping Study
Previous Article in Journal
Plasmonic Field-Effect Transistors (TeraFETs) for 6G Communications
Previous Article in Special Issue
Learning Macromanagement in Starcraft by Deep Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Level-K Classification from EEG Signals Using Transfer Learning

Department of Industrial Engineering and Management, Ariel University, Ariel 4076414, Israel
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(23), 7908; https://doi.org/10.3390/s21237908
Submission received: 14 October 2021 / Revised: 23 November 2021 / Accepted: 24 November 2021 / Published: 27 November 2021

Abstract

:
Tacit coordination games are games in which communication between the players is not allowed or not possible. In these games, the more salient solutions, that are often perceived as more prominent, are referred to as focal points. The level-k model states that players’ decisions in tacit coordination games are a consequence of applying different decision rules at different depths of reasoning (level-k). A player at L k = 0 will randomly pick a solution, whereas a L k 1 player will apply their strategy based on their beliefs regarding the actions of the other players. The goal of this study was to examine, for the first time, the neural correlates of different reasoning levels in tacit coordination games. To that end, we have designed a combined behavioral-electrophysiological study with 3 different conditions, each resembling a different depth reasoning state: (1) resting state, (2) picking, and (3) coordination. By utilizing transfer learning and deep learning, we were able to achieve a precision of almost 100% (99.49%) for the resting-state condition, while for the picking and coordination conditions, the precision was 69.53% and 72.44%, respectively. The application of these findings and related future research options are discussed.

1. Introduction

In tacit coordination games, communication between the players is not allowed or not possible. In these games, the more salient solutions, that are often perceived as more prominent [1], are referred to as focal points. While many experiments showed that players are highly successful in converging on the same focal point (e.g., [1,2]), and even though several attempts were made to construct theoretical frameworks to explain this phenomenon (e.g., [3,4,5,6]), there is still no generally accepted explanation of how players manage to converge on the same solution [4].
One of the most prominent cognitive models that attempts to explain behavior in tacit coordination games is the level-k model, which is based on the cognitive hierarchy theory [4,7,8]. The model assumes that players’ reasoning depth relies on their subjective level of reasoning, k. For example, players in which k = 0 (sometimes referred to as L 0 players) will choose randomly between the available actions, while L 1 players assume that all other players are L 0 reasoners and will act according to this assumption. That is, L 0 players might utilize rules but will apply them randomly (picking), whereas L k 1 players will apply their strategy based on their beliefs regarding the actions of the other players (coordination).
In recent years, there has been a growing interest in examining how people reason in strategic situations. However, research regarding the level-k model has been conducted in the context of behavioral game theory, while in this study, we proposed to examine the neural correlates of different reasoning levels in tacit coordination games. To examine different levels of reasoning, we have designed a combined behavioral-electrophysiological study with three different conditions, each resembling a different depth of reasoning: (1) resting state, (2) picking, and (3) coordination. Each participant underwent the three conditions in this sequential order. First, the resting-state EEG was recorded from participants while they were requested to gaze at a cross in the center of the screen. In the second stage of the experiment, participants were engaged in a picking task. Participants were presented sequentially with different sets of four words appearing in Hebrew, and in each of the trials, they were asked to freely pick a word out of each set. In the third stage, participants were presented with the same sets of words that were presented in the picking condition. However, this time participants were engaged in a coordination task. That is, participants were instructed to coordinate their choice of a word with an unknown partner so that they would converge on the same word from the set presented to them in each of the trials. EEG was recorded from participants while they were performing each of the tasks. Before the start of the actual experiment, participants underwent a training session while wearing the EEG cap to get them familiar with the picking and coordination tasks.
The overarching goal of our study was to classify EEG continuous data into the level-k condition they were associated with. This will show that the level-k model can also be validated by electrophysiological correlates and not only by behavioral indices. This validation may potentially enable the construction of more accurate models for human–agent interactions [9]. To that end, we have first used methods of feature extraction and classification based on conventional machine learning techniques, such as computing the relative energy in each frequency band and applying standard predictive models such as random forest (see Appendix D). These techniques were not proven to be sufficient due to the complexity of the problem and the amount of data such models require. With that in mind, we have used the transfer learning technique (e.g., [10,11,12]) with pre-trained deep learning models trained on large datasets, which we have adjusted to the problem at hand. Specifically, the learning model was based on a continuous wavelet transform 2D image, which has been fed into a pre-trained network (VGG16 trained on ImageNet). Since determining the optimal weights of the different EEG channels is a computationally hard problem, we have defined a relative cost function and optimized the set of weights using a genetic algorithm. Our multi-channel deep learning method achieved a precision of almost 100% (99.49%) for the resting-state condition, while for the picking and coordination conditions, the precision was 69.53% and 72.44%, respectively.
The contribution of our study is three-fold. First, we have validated the level-k theory in the context of tacit coordination by using electrophysiological data. Second, we have demonstrated that the state-of-the-art transfer learning technique can be useful to cope with a complex classification problem with a low amount of electrophysiological data. Third, we have managed to predict the class label of EEG segments associated with different experimental conditions: picking (level-k = 0), tacit coordination (level-k > 0), or no-task (resting state). The implications of these findings and related future research options are discussed.

2. Materials and Methods

2.1. Participants

The participants were 10 students from Ariel University that were enrolled in one of the courses on campus (right-handed, mean age = ~26 years, SD = 4). Each task (picking and coordination) started with a verbal explanation followed by reading a written instruction file. Then, participants signed an informed consent form approved by the IRB of Ariel University. Participants were offered a reward based on the total number of points they earned in both tasks.

2.2. Experimental Design

Experimental conditions comprised resting-state EEG recordings with eyes open for two minutes while participants focused on a red cross on the screen overlayed over a grey background. The next two stages were based on the same set of stimuli and presentation scheme. The experiment consisted of two sets of twelve different trials each with a different set of words. For example, game board #1 displays a trial containing the set (“Water”, “Beer”, “Wine”, “Whisky”) appearing in Hebrew, respectively. Each set of words was displayed between two short vertical lines following a slide containing only the lines without the word set so that participants will focus their gaze at the center of the screen (Figure 1A,B) (e.g., [4,13]). In the first experimental condition, the task presented to the players was a picking task, i.e., participants were only required to freely pick a word out of each set of four words presented to them in each of the 12 trials. Subsequently, participants were presented with the coordination task, comprising the same set of 12 different trials. In the coordination condition, participants were instructed to coordinate their choice of a word with an unknown partner so that they would end up choosing the same word from the set. Participants were further informed that they will receive an amount of 100 points for each selection of a word in the picking task, and for each successful coordination in the coordination task. Each participant sat alone in front of the computer screen during the entire experimental session. It is important to note that no feedback was given between the games. That is, the participants were not informed whether they have coordinated successfully or not with their unknown coplayer. The individual accumulated reward for each of the participants was calculated by randomly matching each participant with a coplayer. The reward was presented to each of the participants only after the completion of the series of games.
Figure 2 portrays the outline of the experiment. Each slide containing the set of words (task trials) was preceded by a slide containing only the vertical lines without the word set (standby slides) to keep the gaze of participants at the middle of the screen throughout the experiment. Each of the standby slides was presented for U(2, 2.5) s, while each slide containing the set of words was presented for a maximal duration of 8 s. Following a task trial, participants could move to the next slide with a button press. The sequence of the task trials was randomized in each session.
The EEG was recorded from participants while they were performing the tasks. The EEG was recorded by a 16-channel g.USBAMP bio-signal amplifier (g.tec, Austria) at a sampling frequency of 512 Hz, and 16 active electrodes were used for collecting EEG signals from the scalp based on the international 10–20 system. Recording was performed by the OpenVibe (v3.2.0) [14] recording software. Impedance of all electrodes was kept below the threshold of 5 K (ohm) during all recording sessions.
Before performing the actual experiment, participants underwent a training session while wearing the EEG cap, to get them familiar with the application and task. The training task included a total of five trials (each including a different set of words), as displayed in Appendix C.

2.3. EEG Preprocessing and Feature Extraction Using CWT

Based on the literature (e.g., [15,16,17,18,19]), we focused on the following cluster of frontal and prefrontal electrodes (Fp1, F7, Fp2, F8, F3, and F4). The preprocessing pipeline (see Figure 3) consisted of finite impulse response (FIR), band-pass filtering (BPF) (1,32) Hz, and artifact removal following iCA. The data were re-referenced to the average reference and down-sampled from 512 to 64 Hz following baseline correction. Data were analyzed on a 1 s epoch window from the onset of each game. In the resting-state condition, a 30 to 90 s epoch was extracted from the entire 120 s from trial onset, resulting in 60 1 s epochs per participant. However, in the picking and coordination conditions, there was a total of 12 decision points per participant.
Then, to extract the features on which the processing will be performed, which in this case is an image, we performed the continuous wavelet transform (CWT) calculation. The CWT is a mathematical transformation that gives the signal a complete two-dimensional representation of time and scaling using a wavelet function that receives a continuously changing scale value [20]. We have used the Symlet (e.g., [21,22]) wavelet filter with a scaling factor ranging from 1 to 32. In addition, to obtain optimal results, we examined filters of the order 2, 4, 6, 8, and 10, and the best one according to the cross-entropy (CE) loss function (see Equations (3) and (4)) turned out to be a sixth-order Symlet filter (see Appendix A).

2.4. Software Tools and Work Environments

In this study, we used a variety of different tools and environments to perform data collection and analysis. The EEG was recorded by a 16-channel g.USBAMP bio-signal amplifier (g.tec, Austria) using 16 active electrodes based on the international 10–20 system. EEG was recorded by using OpenVibe [14]. The experimental application was developed in Java and included a communication interface based on the TCP-IP protocol to handle triggers.
The pre-processing pipeline (Figure 3) was implemented using EEGLAB [23] (v14.1.1). The CWT and DWT transformations were carried out by the Wavelet Toolbox and Signal Processing Toolbox in MATLAB 2016a. The transfer learning deep models were constructed by using the Keras python package. The optimization process was implemented using the NumPy package in python.

3. Results

Figure 4 shows the CWT conversion results of channel 1 (Fp1) of player #3 in all three experimental states (resting, picking, and coordination). The presented coordination and picking epoch were taken from the same experimental trial containing the same set of words. The x-axis represents the time of the epoch, [0, 1] (s), which is equivalent to 64 samples, while the y-axis represents the CWT value of the corresponding wavelet scaling factor in the range of (1, 32). The wavelet scale corresponds to the frequency of the signal. Note that the higher the scaling factor, the lower the corresponding wavelet frequency. The whole presented analysis can also be seen graphically in Figure 5, which shows the distribution of the average CWT values from Figure 4 according to the time (upper row) and CWT scale (lower row) variables.
Analyzing the differences between the various modes based on the timeline, it can be seen (Figure 4 and Figure 5, upper row) that in the resting-state mode, there is almost no energy increase throughout the epoch. However, in the picking task (level-k = 0), a prominent peak appears after 330 milliseconds (sample 21 out of 64), while in the coordination task (level-k > 0), several prominent peaks, indicating an increase in the signal energy, appear throughout the entire epoch. It can also be observed that the higher frequencies (which are related to smaller CWT scale values) are more dominant in the coordination tasks compared to the picking task and resting-state conditions. This result is consistent with previous studies (e.g., [24,25]) that show that the beta frequency domain (13–30 Hz), especially in the prefrontal brain area, is directly related to brain activity associated with top-down processes associated with prediction and expectation [26].

3.1. Cognitive Level Classification Using a Single EEG Channel

To classify the subject’s experimental condition (resting, picking, and coordination) based on the player CWT image during the cognitive task, we have constructed a classification model. However, the small amount of data we have, which includes 840 observations per EEG channel (120 picking epochs, 120 coordination epochs, 600 RS epochs), presents a challenge as it does not allow us to produce a highly complex model. To deal with the problem of the small number of observations in such a complex task, we will take two algorithmic-architectural steps. First, we will implement our classifier according to a one-versus-all approach [27,28]. That is, we will create three different classifiers, each for predicting one of the tree specific labels (resting, picking, and coordination). This approach reduces the complexity of each classifier because its purpose is to identify a single label, which is a more reasonable task given the small number of observations. Consequently, we are now required to train three models instead of one when each model receives a different dataset in which the target images are given the label “1” and all other images the label “0”. The final predicted label will be determined by the highest probability predicted by each of the three models (using an Argmax function). The complete classification architecture for a single electrode based on three one-versus-all classifiers is presented in Figure 6.
Second, we will use a transfer learning method (e.g., [29,30]) that focuses on the use of knowledge generated in a particular problem to solve another problem that has similar characteristics. This enables developing complex models with a low amount of observations [31]. The database used for transfer learning is taken from the ImageNet project [32,33], a visual database used for visual object recognition research that includes more than 14 million tagged images with over 20,000 possible classes.
To produce an optimal prediction model for our problem, classification of the cognitive level (level-k), we will examine three different pre-trained models: InceptionNet V3 [34], ResNet50 [35], and VGG16 [36]. The weights of each pre-trained network were originally calculated to detect one of each of the 1000 different classes (for example zebra, ox, submarine, ambulance, lemon, etc.). The training set on which the networks were trained included about 1.2 million images, while in addition, there were about 50,000 images for validation and about 100,000 images for testing.
In the model training process (Figure 7), we used the abovementioned pre-trained networks to extract the features from the CWT image. The weights of each network were frozen so that they would remain unchanged even at the end of the training process, after which we added a single neuron and a sigmoid activation function after the last pooling layer to obtain a prediction for our problem. The embedding features were taken out of the last pooling layer of the network and not from the output layer because of the difference between the testing set (images of EEG segments) and the transfer learning network training set (ImageNet). When the training set and the testing set comprise images of similar context, it is possible to take the features that are closer to the output layer. However, when the sets are from completely different domains, it is recommended to a take the features from pooling layers residing closer to the middle of the network, which represent basic shapes such as lines, circles, and trends [37]. We preferred using a single-neuron model to more complicated multi-layered architectures due to the size of the training set (for detailed results, see Appendix E). To avoid overfitting, we worked with a four-fold cross-validation method so that the training set included 630 samples at a time (three-fold) and the test set included 210 samples (one-fold). We repeated this process three times to obtain a reliable prediction of all the samples in the test group. The cost function used for the training process was binary cross-entropy (BCE) with balanced weights to overcome the amount of unbalanced observations between the different labels (as there is a 1 to 5 ratio between games and resting-state epochs).
As mentioned in Section 3.1, for each network, we performed five different training sessions. In each training session, we used different sets of CWT images as input. Each set was calculated using a Symlet wavelet of the following different orders: {2, 4, 6, 8, 10}. Therefore, in total, we had 15 configurations for the model (5 wavelets times 3 pre-trained networks). The configuration with the best classification performance for a single channel was obtained by the VGG16 network with a sixth-order Symlet wavelet.
Table 1 presents the classification accuracy results for the different channels. In all channels, the accuracy of the model was higher than the chance level (33.33%). The difference between the best (F3) and worst channels (F7) was 9.65% in the total level of accuracy (81 correctly predicted observations). Finally, there is symmetry between the left side of the scalp and the right side. For each homologous interhemispheric electrode pair (i.e., F7–F8, Fp1–Fp2, F3–F4), the same level of total prediction accuracy can be observed.

3.2. Cognitive Level Classification Using Multiple EEG Channels

In this section, we construct a model which combines the individual electrode models that were presented in the previous section. We define the model as a weighted linear combination of all the different single-electrode models.
M ( a , x ) = M a ( x ) = i     S a i M i ( x i ) i     S a i
where:
  • M i —The prediction model of the ith channel,
  • x i ith channel model input—CWT image of the EEG record from the ith channel,
  • a i —The model weight of the ith channel.
The result of the weighted model will provide the estimated probability of each of the three conditions, namely, resting state (L = (1;0;0)), picking task (L = (0;1;0)), or a coordination task (L = (0;0;1)). In order to measure the quality of the solution with the various weights, we use the cross-entropy (CE) cost function [38,39] which quantifies the difference between the actual label (i.e., the actual classification value) and the probabilities of the labels predicted by the model:
C O S T ( M a ( x ) , L ( x ) ) = x L ( x ) log ( M a ( x ) )
where:
  • M a ( x ) —The predicted probability vector by the weighted model for input x (which is a CWT image).
  • L (x)—Actual label of input x (CWT image).
To obtain an optimal model for the entire dataset of samples in our problem, we calculate the value of the average cost function of all the samples:
J ( a ) = 1 m i = 1 m C O S T ( M a ( x ( i ) ) , L ( x ( i ) ) ) = 1 m i = 1 m L ( x ( i ) ) log ( M a ( x ( i ) ) )
where:
  • m —the number of observations in the dataset.
With that in mind, we seek to find the set of weights that will bring the cost function to a minimum:
M I N a   J ( a )
In order to find the set of weights that brings the cost function to a minimum, we have used the genetic algorithm (GA) optimization process (e.g., [40,41]). To evaluate the improvement afforded by the GA model, we have compared it against a baseline model, which was an unweighted model comprising equal weights for each of the six electrodes. GA is an optimization method for hard combinatorial problems that uses a natural selection process, iteratively. In each iteration, the existing set of solutions gradually improves compared to the previous generation of the solution, and this is carried out by biologically inspired operators such as mutation, crossover, and selection.
Specifically, the optimization process started with 50,000 random solutions, where in each iteration, we kept the 15,000 best solutions (using the selection operator). In addition, we have created an additional 20,000 crossover solutions via a pairing of two previous generation solutions selected by the fitness values. Mutation was implemented as random changes in one of the weights in the range between 10% and 50%. In each iteration, we created 15,000 solutions in a mutation configuration, so that in total, we were left with the same amount of 50,000 solutions at the end of each iteration.
Several different and independent runs of the optimization algorithm determined the optimal weights (see Table 2), which resulted in a model with predictive accuracy at the level of 91.66% (770/840) (Table 3), but since our dataset is unbalanced, we will compare the accuracy levels (i.e., true positive rate) of each label, as presented in Table 3. This result represents an overall improved accuracy of 5.09% compared to the best single channel classifier (see Table 1).
Following the optimization process, several insights can be presented. First, the errors of the classifier are only first-order errors. That is, resting-state instances were only mislabeled as picking (level-k = 0) and not as coordination (level-k > 0). By the same token, coordination instances were only mislabeled as picking but not as resting-state instances. Second, the classifier precision (i.e., positive predicted value) is not equal for the different conditions. That is, while for the resting-state epochs the classifier precision is almost 100% (99.49%), for the picking and coordination segments, the precision is 69.53% and 72.44%, respectively. This is probably since the resting-state condition is associated with the spontaneous activity of the brain, whereas the picking and coordination conditions are associated with different levels of task states. It is noteworthy that the imbalance in the amount of data between the different conditions might also contribute to the differences in precision values among the conditions. Similarly, different levels of recall were also observed for the different conditions (98.17% for resting state, 74.16% for picking, and 76.67% for coordination).
Figure 8 shows the interpolation of the values of the optimal weights on the scalp according to the 10–20 system. The values of the unused electrodes were set to 0. In this way, it is possible to visually present the relative contribution of each electrode in the combined prediction model.
The level of accuracy of the unweighted model for all classes was lower than the best results of a single electrode. This indicates that the GA model performed better than the baseline model with equal weights.

4. Discussion

The overarching goal of our study was to classify EEG continuous data into the level-k condition they were associated with. In this study, we have presented a method to predict the class label of EEG segments taken from three different conditions: two cognitive tasks and a no-task condition (resting state). Each of the two cognitive tasks was associated with a different depth of reasoning, namely, picking (level-k = 0) and tacit coordination (level-k > 0). Classification relied on EEG measures using CWT transformation and transfer learning based on pre-existing state-of-the-art models for object recognition. We have constructed two classification models: the first classification model relied on single-electrode data as input, and the second model optimized the weights of the six frontal and pre-frontal electrodes using a genetic algorithm.
Additionally, we demonstrated that the state-of-the-art transfer learning technique [10,11,12] can be useful to cope with a complex classification problem with a low amount of data. Specifically, in this study, transfer learning was used to validate the level-k model by classifying electrophysiological data. We have used transfer learning since conventional machine learning models (e.g., random forest, see Appendix D) were not successful in differentiating between picking (level-k = 0) and coordination (level-k > 0). However, the conventional model was successful in differentiating between the no-task condition (resting state) and the two cognitive tasks (picking and coordination). The failure in distinguishing between the two cognitive states could be attributed to the fact the DWT, which is based on 400 ms time windows (with a 50% overlap), was not refined enough to capture the differences between the two conditions. In order to use a higher time resolution, we preferred using CWT. Since this computational method produces an image which requires a large amount of data to distinguish between different geometrical patterns, we opted for using a pre-trained network (VGG16 trained on ImageNet) to embed the CWT image to a feature vector. The feature vectors were used for training an additional neural network which classified between the different conditions.
Transfer learning has been previously used in EEG studies for classification purposes, mainly in the context of clinical research, e.g., [10,11,12]. In these studies, when the problem was relatively simple or in cases where a low amount of data was used, analysis was based on a single-electrode classification [10]. Otherwise, analysis was based on the embedding of multiple electrodes [11,12], to cope with the requirement for a large amount of data. The novelty in our study is that we have trained a separate classifier for each individual electrode and weighted the separate contribution of each electrode for optimal classification results. Training a network with an input layer of six neurons and an output layer of three neurons (resting, picking, and coordinating) with two hidden layers on a small dataset (120 examples of pickers and coordinators each) is not feasible. Therefore, since this study is a multi-class problem with a relatively small and unbalanced amount of data, and given our goal to maximize the accuracy of each classifier, we implemented each classifier based on an input from a single electrode in a one-versus-all architecture. Interestingly, in our study, the topographic distribution of the weights of different models was the strongest over frontal regions (Figure 8). This result is corroborated by previous findings showing a similar distribution, specifically, that cognitive load is enhanced in these regions during on-task periods [42,43]. Overall, the current study presented a novel method for differentiating between cognitive states associated with different depths of reasoning in the context of tacit coordination. Furthermore, to the best of our knowledge, this is the first study to corroborate the level-k theory based on electrophysiological measures.
The findings of our study suggest several avenues for future research. For example, previous studies have shown that various features such as culture [44,45], social value orientation [46,47], strategic profile [48,49], and loss aversion [50,51] might bias decision-making in tacit coordination games. Therefore, it will be interesting to investigate the effect of these parameters on EEG indices. In addition, EEG indices can be used to test assumptions associated with other behavioral economic models, such as team reasoning (e.g., [4,7,52,53]) or cognitive hierarchy theory (e.g., [4,7,8,54]). Furthermore, adding information based on brain sources associated with decision-making in tacit coordination might improve classification accuracy levels. To that end, in future studies, it is recommended to utilize inverse-problem techniques such as LORETA [55,56]. Finally, behavioral and electrophysiological data of human agents (e.g., [2,9,57,58,59]) gained from these studies might aid in constructing brain–computer interfaces as well as autonomous agents. In this study, we used transfer learning when the training set for the embedding network was a general set of images (ImageNet) that did not include EEG signals. It will be interesting to compare the performance of the network when the training set of the embedding network comprises only EEG signals or a mixture of EEG signals and other signal types.

Author Contributions

D.M., I.L. and I.Z. carried out the stages of conceptualization, design of methodology, data curation, formal analysis, data modeling, model validation, writing, drafting and editing. D.M. was also responsible for visualization and implementation of supporting algorithms. I.Z. and I.L. supervised the research activity. All authors discussed the results, read and approved the final manuscript and are accountable for all aspects of the work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of Ariel university (confirmation number: AU-SOC-SL-20190901, 1 September 2019 to 31 August 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Statements

The experimental protocols used in this work were evaluated and approved by the Ethics Committee of Ariel University (confirmation number: AU-SOC-SL-20190901). Permission to perform the electrophysiological recordings in the experiment was given from 1 September 2019 to 31 August 2020.

Appendix A. Symlet Wavelet Functions

In this appendix, we will present the Symlet wavelet function in different orders (2, 4, 6, and 8), as can be seen in Figure A1. Figure A1 shows the tradeoff in selecting the wavelet order. While using a higher order wavelet makes it possible to identify more complex patterns, there is a larger number of samples in the wavelet which impairs the temporary resolution of the transduction because now more specimens are involved in each forearm. The problem can also be looked at in reverse, a low order wavelet will provide good time resolution, but its structure is simple, and it will be difficult to find complex patterns in the signal.
Figure A1. Symlet wavelet function in different orders.
Figure A1. Symlet wavelet function in different orders.
Sensors 21 07908 g0a1
In this research, the optimal classified results on a single channel according to the CE cost function (see Equations (3) and (4)) were obtained by a sixth-order wavelet.

Appendix B. Tacit Coordination Game List

In this appendix, we will describe the set of tacit coordination games, which includes twelve games, that were designed in order to evaluate the individual coordination abilities of the various player together with their electrophysiological patterns in different cognitive hierarchy levels. The full game list is presented in Table A1. It should be noted that the words in the game boards appeared in the Hebrew, which is the native language of the participants.
Table A1. Tacit coordination game list.
Table A1. Tacit coordination game list.
Game NumberOption 1Option 2Option 3Option 4
1WaterBeerWine Whisky
2TennisVolleyballFootballChess
3BlueGrayGreenRed
4IronSteelPlasticBronze
5FordFerrariJaguarPorsche
618516
7HaifaTel-AvivJerusalemNetanya
8SpinachCarrotLettucePear
9LondonParisRomeMadrid
10HazelCashewAlmondPeanut
11StrawberryMelonBananaMango
12NoodlesPizzaHamburgerSushi
The position of the questions appearing on the game screen as can be seen in Figure 1 is fixed and follows the order of the lists shown in Table A1. This decision in the design of the experiment was made to create a uniform experimental set-up between the various actors and to neutralize the possible effect of spatial cues.

Appendix C. Training Tasks Game List

This appendix presents the training task, which was performed before the picking and coordination tasks. The purpose of these tasks is to verify the players’ technical understanding of the application before performing the actual experiment. From a review of Table A2, there is no overlap in the content of the training tasks with experiment tasks.
Table A2. Training game list.
Table A2. Training game list.
Game NumberOption 1Option 2Option 3Option 4
1SapphireGlassEmeraldDiamond
2LionPantherFrogTiger
3BoatHelicopterBicyclePlane
4ThursdayTuesdaySaturdaySunday
52019200019951997

Appendix D. Classification Using Classical Machine Learning Models

To estimate the performance of a conventional classifier, namely, random forest, we trained a model based on discrete wavelet transform (DWT) [60,61,62], which is the discrete equivalent of the CWT. The DWT was used to calculate the relative energy in each EEG band (Delta, Theta, Alpha, and Beta), as presented in Figure A2.
Figure A2. EEG relative band calculation using DWT.
Figure A2. EEG relative band calculation using DWT.
Sensors 21 07908 g0a2
Four hamming windows with 50% overlap were used. Each window was 400 ms long and contained 26 samples, as can be seen in Figure A3.
Figure A3. Baseline model windowing scheme.
Figure A3. Baseline model windowing scheme.
Sensors 21 07908 g0a3
Windowing and relative energy computation of each window resulted in 16 features (4 windows with 4 features per window). Based on these features, three classifiers were constructed, one for each class (resting state, picking, coordination) (see Figure 6). Each of the three classifiers was trained using a random forest model with 100 estimators (optimized using grid search) using 4-fold cross-validation. The model was trained on the F3 electrode, which provided the best classification results among all single electrodes.
As can be seen in Table A3, using the DWT and the random forest model, we have managed to differentiate between the no-task state (resting state) and the two cognitive states (picking and coordination), whereas we failed to distinguish between the latter two.
Table A3. Baseline model using single channel (F3)—confusion matrix.
Table A3. Baseline model using single channel (F3)—confusion matrix.
Predicted ClassesTrue Positive RateFalse Negative Rate
Resting State
(CHT Does Not Exist)
L = [1;0;0]
Picking
(CHT = 0)
L = [0;1;0]
Coordination
(CHT > 0)
L = [0;0;1]
True ClassesResting state
(CHT does not exist)
L = [1;0;0]
520493186.67%13.33%
Picking
(CHT = 0)
L = [0;1;0]
23653254.16%45.84%
Coordination
(CHT > 0)
L = [0;0;1]
5427360.83%39.17%
Positive Predicted Value94.89%41.67 %53.67 % Total Prediction Accuracy (658/820)
80.24%
False Discovery Rate1.51%58.33%46.33%

Appendix E. The Effect of Model Complexity on Classification Results

In this appendix we will examine the impact of more complex architectures on the classification results of a single electrode. In addition to the single-neuron architecture, we will examine three additional architectures, with 2, 3, and 4 layers. The structure of the four-layer network can be seen in Figure A4 (bias neurons are painted in blue).
Figure A4. Four-layer trainable neural network model.
Figure A4. Four-layer trainable neural network model.
Sensors 21 07908 g0a4
The training process was performed similarly to the process described in the manuscript by using a four-fold cross-validation method. We repeated this process three times to obtain a reliable prediction of all the samples in the test group. In order to fully evaluate the quality of the different models, we will use the F1-score index that weights the precision ( T P T P + F P ) and recall ( T P T P + F N ) by a harmonic mean. All the reported results in the appendix are based on using transfer learning with VGG16, that was trained on ImageNet. Since the results obtained for the picking and coordination conditions had similar trends, we present the results for a one-versus-all model that identifies the coordination condition based on the electrophysiological signal. The evaluation of the various models for each of the six frontal and prefrontal electrodes can be seen in Table A4.
Table A4. The effect of model complexity on classification results—coordination task (f1 score is the measure by which we evaluate the quality of the model best model colored in green. The least successful model is marked in red).
Table A4. The effect of model complexity on classification results—coordination task (f1 score is the measure by which we evaluate the quality of the model best model colored in green. The least successful model is marked in red).
Electrode/Architecture(Fp1)(F7)(Fp2)(F8)(F3)(F4)
1 layerprecisionprecisionprecisionprecisionprecisionprecision
56.10% (92/164)52.15% (85/163)61.36% (81/132)56.11% (78/139)68.18% (90/132)64.93% (87/134)
RecallRecallRecallRecallRecallRecall
76.66%—(92/120)68.33%—(82/120)67.50%—(81/120)65.00%—(78/120)75.00%—(90/120)72.50%—(87/120)
f1 score = 0.6479f1 score = 0.6007f1 score = 0.6429f1 score = 0.6023f1 score = 0.7143f1 score = 0.6851
2 layersprecisionprecisionprecisionprecisionprecisionprecision
58.17% (89/153)59.57% (84/141)59.71% (83/139)57.66% (79/137)66.66% (92/138)64.23% (88/137)
RecallRecallRecallRecallRecallRecall
74.17%—(89/120)70.00%—(84/120)69.17%—(83/120)65.83%—(79/120)76.66%—(92/120)73.33%—(88/120)
f1 score = 0.6520f1 score = 0.6437f1 score = 0.6409f1 score = 0.6148f1 score = 0.7131f1 score = 0.6848
3 layersprecisionprecisionprecisionprecisionprecisionprecision
49.67% (75/151)60.87% (84/138)59.29% (83/140)56.30% (76/135)66.66% (88/132)63.70% (86/135)
RecallRecallRecallRecallRecallRecall
62.50%—(75/120)70.00%—(84/120)69.17%—(83/120)63.33%—(76/120)73.33%—(88/120)71.66%—(86/120)
f1 score = 0.5535f1 score = 0.6512f1 score = 0.6385f1 score = 0.5961f1 score = 0.6984f1 score = 0.6745
4 layersprecisionprecisionprecisionprecisionprecisionprecision
55.88% (76/136)52.32% (79/151)57.35% (78/136)53.64% (81/151)70.43% (81/115)63.70%—(79/125)
RecallRecallRecallRecallRecallRecall
63.33%—(76/120)65.83%—(79/120)65.00%—(78/120)67.50%—(81/120)67.50%—(81/120)65.83%—(79/120)
f1 score = 0.5938f1 score = 0.5830f1 score = 0.6094f1 score = 0.5978f1 score = 0.6894f1 score = 0.6449
Table A4 displays precision and recall scores for each combination of electrodes and number of layers, as well as the F1 score (the weighted average of precision and recall) for each combination of electrodes and number of layers. The table shows that more complicated models with more layers do not improve the F1 score compared to the simplest single-neuron model. This finding is probably due to overfitting created by complicated models applied on a relatively small amount of information. Consequently, we decided to stay with the simplest single-neuron model, which is the most parsimonious one.

References

  1. Schelling, T.C. The Strategy of Conflict; Harvard University Press: Cambridge, MA, USA, 1960. [Google Scholar]
  2. Zuckerman, I.; Kraus, S.; Rosenschein, J.S. Using focal point learning to improve human-machine tacit coordination. Auton. Agents Multi. Agent Syst. 2011, 22, 289–316. [Google Scholar] [CrossRef] [Green Version]
  3. Bacharach, M.; Stahl, D.O. Variable-Frame Level-n Theory. Games Econ. Behav. 2000, 32, 220–246. [Google Scholar] [CrossRef]
  4. Bardsley, N.; Mehta, J.; Starmer, C.; Sugden, R. Explaining Focal Points: Cognitive Hierarchy Theory versus Team Reasoning. Econ. J. 2009, 120, 40–79. [Google Scholar] [CrossRef] [Green Version]
  5. Rascovsky, K.; Clark, R.; McMillan, C.T.; Khella, M.C.; Grossman, M. The neural basis for establishing a focal point in pure coordination games. Soc. Cogn. Affect. Neurosci. 2011, 7, 881–887. [Google Scholar]
  6. Sugden, R. A Theory of Focal Points. Econ. J. 1995, 105, 533–550. [Google Scholar] [CrossRef]
  7. Faillo, M.; Smerilli, A.; Sugden, R. The Roles of Level-k and Team Reasoning in Solving Coordination Games; University of Trento: Trento, Italy, 2013. [Google Scholar]
  8. Gold, N.; Colman, A.M. Team Reasoning and the Rational Choice of Payoff-Dominant Outcomes in Games. Topoi 2018, 39, 1–12. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Mizrahi, D.; Zuckerman, I.; Laufer, I. Using a Stochastic Agent Model to Optimize Performance in Divergent Interest Tacit Coordination Games. Sensors 2020, 20, 7026. [Google Scholar] [CrossRef]
  10. Jiang, Y.; Wu, D.; Deng, Z.; Qian, P.; Wang, J.; Wang, G.; Chung, F.-L.; Choi, K.-S.; Wang, S. Seizure classification from EEG signals using transfer learning, semi-supervised learning and TSK fuzzy system. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 2270–2284. [Google Scholar] [CrossRef]
  11. Lin, Y.-P.; Jung, T.-P. Improving EEG-based emotion classification using conditional transfer learning. Front. Hum. Neurosci. 2017, 11, 334. [Google Scholar] [CrossRef]
  12. Fahimi, F.; Zhang, Z.; Boon Goh, W.; Lee, T.-S.; Ang, K.K.; Guan, C. Inter-subject transfer learning with an end-to-end deep convolutional neural network for EEG-based BCI. J. Neural Eng. 2019, 16, 026007. [Google Scholar] [CrossRef] [Green Version]
  13. Mizrahi, D.; Laufer, I.; Zuckerman, I. The Effect of Individual Coordination Ability on Cognitive-Load in Tacit Coordination Games. In Proceedings of the NeuroIS Retreat 2020; Davis, F., Riedl, R., vom Brocke, J., Léger, P.-M., Randolph, A., Fischer, T., Eds.; Springer Cham: Manhattan, NY, USA; Vienna, Austria, 2020. [Google Scholar]
  14. Renard, Y.; Lotte, F.; Gibert, G.; Congedo, M.; Maby, E.; Delannoy, V.; Bertrand, O.; Le´cuyer, A. Openvibe: An open-source software platform to design, test, and use brain–computer interfaces in real and virtual environments. Presence Teleoperators Virtual Environ. 2010, 19, 35–53. [Google Scholar] [CrossRef] [Green Version]
  15. Gartner, M.; Grimm, S.; Bajbouj, M. Frontal midline theta oscillations during mental arithmetic: Effects of stress. Front. Behav. Neurosci. 2015, 9, 1–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. De Vico Fallani, F.; Nicosia, V.; Sinatra, R.; Astolfi, L.; Cincotti, F.; Mattia, D.; Wilke, C.; Doud, A.; Latora, V.; He, B.; et al. Defecting or not defecting: How to “read” human behavior during cooperative games by EEG measurements. PLoS ONE 2010, 5, e14187. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Boudewyn, M.; Roberts, B.M.; Mizrak, E.; Ranganath, C.; Carter, C.S. Prefrontal transcranial direct current stimulation (tDCS) enhances behavioral and EEG markers of proactive control. Cogn. Neurosci. 2019, 10, 57–65. [Google Scholar] [CrossRef]
  18. Moliadze, V.; Sierau, L.; Lyzhko, E.; Stenner, T.; Werchowski, M.; Siniatchkin, M.; Hartwigsen, G. After-effects of 10 Hz tACS over the prefrontal cortex on phonological word decisions. Brain Stimul. 2019, 12, 1464–1474. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Mizrahi, D.; Laufer, I.; Zuckerman, I. Topographic Analysis of Cognitive Load in Tacit Coordination Games Based on Electrophysiological Measurements. In Proceedings of the NeuroIS Retreat 2021; Springer: Manhattan, NY, USA; Vienna, Austria, 2021. [Google Scholar]
  20. Mallat, S. A Wavelet Tour of Signal Processing; Elsevier: Amsterdam, The Netherlands, 1999. [Google Scholar]
  21. Wang, X.; Guanghong, G.; Li, N. Automated recognition of epileptic EEG states using a combination of symlet wavelet processing, gradient boosting machine, and grid search optimizer. Sensors 2019, 19, 219. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Mahmoodin, Z.; Jalalludin, N.S.; Mansor, W.; Lee, K.Y.; Mohamad, N.B. Selection of Symlets wavelet function order for EEG signal feature extraction in children with dyslexia. In Proceedings of the 2015 IEEE Student Symposium in Biomedical Engineering & Sciences (ISSBES), Shah Alam, Malaysia, 4 November 2015; pp. 113–117. [Google Scholar]
  23. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Gross, D.W.; Gotman, J. Correlation of high-frequency oscillations with the sleep–wake cycle and cognitive activity in humans. Neuroscience 1999, 94, 1005–1018. [Google Scholar] [CrossRef]
  25. Ray, W.J.; Cole, H.W. EEG activity during cognitive processing: Influence of attentional factors. Int. J. Psychophysiol. 1985, 3, 43–48. [Google Scholar] [CrossRef]
  26. Tani, R.; Kashimori, Y. Coordination of top-down influence on V1 responses by interneurons and brain rhythms. Biosystems 2021, 104452. [Google Scholar] [CrossRef]
  27. Galar, M.; Fernández, A.; Barrenechea, E.; Bustince, H.; Herrera, F. An overview of ensemble methods for binary classifiers in multi-class problems: Experimental study on one-vs-one and one-vs-all schemes. Pattern Recognit. 2011, 44, 1761–1776. [Google Scholar] [CrossRef]
  28. Joutsijoki, H.; Juhola, M. Comparing the one-vs-one and one-vs-all methods in benthic macroinvertebrate image classification. In Proceedings of the International Workshop on Machine Learning and Data Mining in Pattern Recognition; Springer: Berlin, Heidelberg, 2011; pp. 399–413. [Google Scholar]
  29. Huang, Z.; Pan, Z.; Lei, B. Transfer learning with deep convolutional neural network for SAR target classification with limited labeled data. Remote Sens. 2017, 9, 907. [Google Scholar] [CrossRef] [Green Version]
  30. Ng, H.-W.; Nguyen, V.D.; Vonikakis, V.; Winkler, S. Deep learning for emotion recognition on small datasets using transfer learning. In Proceedings of the ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 443–449. [Google Scholar]
  31. Singla, A.; Bertino, E.; Verma, D. Overcoming the lack of labeled data: Training intrusion detection models using transfer learning. In Proceedings of the 2019 IEEE International Conference on Smart Computing, Washington, DC, USA, 12–15 June 2019; pp. 69–74. [Google Scholar]
  32. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2019; pp. 248–255. [Google Scholar]
  33. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  34. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Simonyan, K. Andrew Zisserman Very deep convolutional networks for large-scale image recognition. arXiv Prepr. 2014, 1409, 1556. [Google Scholar]
  37. Tsung, F.; Zhang, K.; Cheng, L.; Song, Z. Statistical transfer learning: A review and some extensions to statistical process control. Qual. Eng. 2018, 30, 115–128. [Google Scholar] [CrossRef]
  38. Ho, Y.; Wookey, S. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access 2019, 8, 4806–4813. [Google Scholar] [CrossRef]
  39. Muppidi, M.; Rad, P.; Agaian, S.S.; Jamshidi, M. Image segmentation by multi-level thresholding using genetic algorithm with fuzzy entropy cost functions. In Proceedings of the International Conference on Image Processing Theory, Tools and Applications (IPTA), Orleans, France, 9–13 November 2015; pp. 143–148. [Google Scholar]
  40. Mitchell, M. An Introduction to Genetic Algorithms; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  41. Harik, G.R.; Lobo, F.G.; Goldberg, D.E. The compact genetic algorithm. IEEE Trans. Evol. Comput. 1999, 3, 287–297. [Google Scholar] [CrossRef] [Green Version]
  42. van Son, D.; de Rover, M.; De Blasio, F.M.; van der Does, W.; Barry, R.J.; Putman, P. Electroencephalography theta/beta ratio covaries with mind wandering and functional connectivity in the executive control network. Ann. N. Y. Acad. Sci. 2019, 1452, 52–64. [Google Scholar] [CrossRef]
  43. van Son, D.; De Blasio, F.M.; Fogarty, J.S.; Angelidis, A.; Barry, R.J.; Putman, P. Frontal EEG theta/beta ratio during mind wandering episodes. Biol. Psychol. 2019, 140, 19–27. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Mizrahi, D.; Laufer, I.; Zuckerman, I. Collectivism-individualism: Strategic behavior in tacit coordination games. PLoS ONE 2020, 15, e0226929. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Cox, T.H.; Lobel, S.A.; Mcleod, P.L. Effects of ethnic group cultural differences on cooperative and competitive behavior on a group task. Acad. Manag. J. 1991, 34, 827–847. [Google Scholar]
  46. Mizrahi, D.; Laufer, I.; Zuckerman, I.; Zhang, T. The effect of culture and social orientation on Player’s performances in tacit coordination games. In Proceedings of the International Conference on Brain Informatics; Springer: Arlington, TX, USA, 2018; pp. 437–447. [Google Scholar]
  47. Mizrahi, D.; Laufer, I.; Zuckerman, I. The Effect of Expected Revenue Proportion and Social Value Orientation Index on Players’ Behavior in Divergent Interest Tacit Coordination Games. In Proceedings of the International Conference on Brain Informatics; Springer: Cham, Switzerland, 2021; pp. 25–34. [Google Scholar]
  48. Mizrahi, D.; Laufer, I.; Zuckerman, I. Individual strategic profiles in tacit coordination games. J. Exp. Theor. Artif. Intell. 2020, 33, 1–16. [Google Scholar] [CrossRef]
  49. Mizrahi, D.; Laufer, I.; Zuckerman, I. Modeling Individual Tacit Coordination Abilities. In Proceedings of the International Conference on Brain Informatics; Springer: Haikou, China, 2019; pp. 29–38. [Google Scholar]
  50. Mizrahi, D.; Laufer, I.; Zuckerman, I. The Effect of Loss-Aversion on Strategic Behaviour of Players in Divergent Interest Tacit Coordination Games. In Proceedings of the International Conference on Brain Informatics; Springer: Padova, Italy, 2020; pp. 41–49. [Google Scholar]
  51. Liu, W.; Song, S.; Wu, C. Impact of loss aversion on the newsvendor game with product substitution. Int. J. Prod. Econ. 2013, 141, 352–359. [Google Scholar] [CrossRef]
  52. Bacharach, M. Interactive team reasoning: A contribution to the theory of cooperation. Res. Econ. 1999, 53, 117–147. [Google Scholar] [CrossRef] [Green Version]
  53. Colman, A.M.; Gold, N. Team reasoning: Solving the puzzle of coordination. Psychon. Bull. Rev. 2018, 25, 1770–1783. [Google Scholar] [CrossRef] [Green Version]
  54. Costa-Gomes, M.A.; Crawford, V.P.; Iriberri, N. Comparing models of strategic thinking in Van Huyck, Battalio, and Beil’s coordination games. J. Eur. Econ. Assoc. 2009, 7, 365–376. [Google Scholar] [CrossRef]
  55. Michel, C.M.; Murray, M.M.; Lantz, G.; Gonzalez, S.; Spinelli, L.; Peralta, R.G. de EEG source imaging. Neurophysiology 2004, 115, 2195–2222. [Google Scholar] [CrossRef] [PubMed]
  56. Pascual-Marqui, R.D.; Michel, C.M.; Lehmann, D. Low resolution electromagnetic tomography: A new method for localizing electrical activity in the brain. Int. J. Psychophysiol. 1994, 18, 49–65. [Google Scholar] [CrossRef]
  57. Kraus, S. Predicting human decision-making: From prediction to action. In Proceedings of the 6th International Conference on Human-Agent Interaction, Southampton, UK, 15–18 December 2018; p. 1. [Google Scholar]
  58. Zuckerman, I.; Kraus, S.; Rosenschein, J.S. The adversarial activity model for bounded rational agents. Auton. Agents Multi. Agent Syst. 2012, 24, 374–409. [Google Scholar] [CrossRef]
  59. Mizrahi, D.; Laufer, I.; Zuckerman, I. Predicting focal point solution in divergent interest tacit coordination games. J. Exp. Theor. Artif. Intell. 2021, 1–21. [Google Scholar] [CrossRef]
  60. Shensa, M.J. The Discrete Wavelet Transform: Wedding the a Trous and Mallat Algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef] [Green Version]
  61. Rioul, O.; Duhamel, P. Fast algorithms for discrete and continuous wavelet transforms. IEEE Trans. Inf. theory 1992, 38, 569–586. [Google Scholar] [CrossRef] [Green Version]
  62. Jensen, A.; la Cour-Harbo, A. Ripples in Mathematics: The Discrete Wavelet Transform; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
Figure 1. (A) Standby screen. (B) Game board #1 {“Water”, “Beer”, “Wine”, “Whisky”}.
Figure 1. (A) Standby screen. (B) Game board #1 {“Water”, “Beer”, “Wine”, “Whisky”}.
Sensors 21 07908 g001
Figure 2. Experimental paradigm with timeline.
Figure 2. Experimental paradigm with timeline.
Sensors 21 07908 g002
Figure 3. Preprocess pipeline.
Figure 3. Preprocess pipeline.
Sensors 21 07908 g003
Figure 4. CWT results in different experimental states (resting, picking, and coordination).
Figure 4. CWT results in different experimental states (resting, picking, and coordination).
Sensors 21 07908 g004
Figure 5. Analysis of Player #3 CWT images as a function of time and CWT scale factor.
Figure 5. Analysis of Player #3 CWT images as a function of time and CWT scale factor.
Sensors 21 07908 g005
Figure 6. One-versus-all classifier architecture.
Figure 6. One-versus-all classifier architecture.
Sensors 21 07908 g006
Figure 7. Transfer learning scheme for binary classifier.
Figure 7. Transfer learning scheme for binary classifier.
Sensors 21 07908 g007
Figure 8. The optimal prediction weights in a 10–20 system.
Figure 8. The optimal prediction weights in a 10–20 system.
Sensors 21 07908 g008
Table 1. Classification accuracy as a function of channel number—VGG16 with symlet6 wavelet.
Table 1. Classification accuracy as a function of channel number—VGG16 with symlet6 wavelet.
Channel Number (Name)1 (Fp1)2 (F7)5 (Fp2)6 (F8)9 (F3)13 (F4)
Model precision—resting state94.62% (528/558)93.15% (517/555)92.02% (531/577)94.33% (516/547)98.62% (571/579)99.65 (576/578)
Model precision—picking (Level-K = 0)59% (70/118)51.63% (63/122)57.25% (75/131)53.90% (83/154)65.89% (85/129)64.06% (82/128)
Model precision—coordination (Level-K > 0)56.10% (92/164)52.15% (85/163)61.36% (81/132)56.11% (78/139)68.18% (90/132)64.93% (87/134)
Total model accuracy82.14% (690/840)79.16% (665/840)82.14% (690/840)80.59% (677/840)88.81% (746/840)88.69% (745/840)
Table 2. The weight values for the different channels in the weighted model after the optimization process.
Table 2. The weight values for the different channels in the weighted model after the optimization process.
Channel Notation(Fp1)(F7)(Fp2)(F8)(F3)(F4)
Calculated Weight0.12160.00130.15530.01080.41530.2957
Table 3. Optimal model using multiple channels—confusion matrix.
Table 3. Optimal model using multiple channels—confusion matrix.
Predicted ClassesTrue Positive RateFalse Negative Rate
Resting State
(CHT Does Not Exist)
L = (1;0;0)
Picking
(CHT = 0)
L = (0;1;0)
Coordination
(CHT > 0)
L = (0;0;1)
True ClassesResting state
(CHT does not exist)
L = (1;0;0)
58911098.17%1.83%
Picking
(CHT = 0)
L = (0;1;0)
3892874.16%25.84%
Coordination
(CHT > 0)
P = (0;0;1)
0289276.67%23.33%
Positive Predicted Value99.49%69.53%72.44 %Total Prediction Accuracy
(770/840)
91.66%
False Discovery Rate1.51%30.47%27.56%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mizrahi, D.; Zuckerman, I.; Laufer, I. Level-K Classification from EEG Signals Using Transfer Learning. Sensors 2021, 21, 7908. https://doi.org/10.3390/s21237908

AMA Style

Mizrahi D, Zuckerman I, Laufer I. Level-K Classification from EEG Signals Using Transfer Learning. Sensors. 2021; 21(23):7908. https://doi.org/10.3390/s21237908

Chicago/Turabian Style

Mizrahi, Dor, Inon Zuckerman, and Ilan Laufer. 2021. "Level-K Classification from EEG Signals Using Transfer Learning" Sensors 21, no. 23: 7908. https://doi.org/10.3390/s21237908

APA Style

Mizrahi, D., Zuckerman, I., & Laufer, I. (2021). Level-K Classification from EEG Signals Using Transfer Learning. Sensors, 21(23), 7908. https://doi.org/10.3390/s21237908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop