1. Introduction
Although psychology has taken longer than other fields to adopt machine learning models for the analysis of experimental results, an increasing number of research works have shown the effectiveness of these models and their benefits in complementing traditional statistical techniques [
1,
2,
3,
4,
5,
6,
7,
8] As a result, machine learning methods are becoming a valuable tool in differential statistics, supporting the development of more accurate forecasting models through the generalisation and evaluation of new data [
9,
10].
In [
1,
2,
3], several authors explored the potential of machine learning models in the analysis of psychological experimental data. In [
1], the authors presented a procedure that aimed to combine exploratory and predictive modelling to construct new psychometric questionnaires with psychological and neuroscientific theoretical grounding [
11]. They employed exploratory data analysis to examine the dimensional structure of the questionnaires and utilised artificial neural networks (ANNs) to predict the psychopathological diagnosis of clinical subjects. Similarly, Orrú et al. [
2] suggested the application of machine learning models to analyse a subset of items from the Toronto Alexithymia Scale, which was useful in accurately distinguishing patients with fibromyalgia from healthy controls. In [
3], the authors claimed that complementing the analytical workflow of psychological experiments with machine-learning-based analysis could maximise accuracy and minimise replicability issues. Additionally, machine learning analysis is model-agnostic and mainly focuses on prediction rather than inference.
Gonzalez et al. [
5] provided a review of four statistical approaches for item selection. The first approach presented was the response theory, which is used to build static short-forms. The second approach was computerised adaptive testing, followed by a genetic algorithm and regression trees. The authors discussed the theoretical strengths and weaknesses of these four statistical solutions and considered the overlap between both areas: psychometrics and machine learning. Their results suggest that machine learning models such as logistic regression or random forest can have comparable classification performance to the psychometric methods using estimated item response theory scores. Therefore, machine learning models can be a viable alternative for classification when psychometric methods are not feasible.
Traditionally, psychology has had a special interest in studying which factors improve the living standards of people [
12,
13,
14,
15]. Among these factors, we can find study of the term “Flow”, which has been widely researched by many authors in recent years [
16].
In psychology, the term “Flow” is defined as the mental state in which a person performing certain tasks is fully immersed in a feeling of energised focus and full involvement [
17,
18]. Initially, the studies were conducted using musicians, basketball players, surgeons, teachers, and businesspeople as these sorts of people seem to perform extremely motivating activities or, in other words, activities that they love to do [
19]. Later, these works were applied to people of different ages, cultures [
20], and jobs, finding that this optimal experience was also present in them.
Thus, this research has the potential to provide significant benefits to both individuals and organisations. By using AI techniques to estimate the state of Flow, we can offer a more objective way to quantify how efficient a person may be at both personal and professional levels. This information can be useful for individuals seeking to improve their performance or make better decisions about their career paths. Moreover, organisations can use this information to optimise work environments and tasks, leading to increased productivity and job satisfaction among employees. By incorporating personality variables that are not directly related to Flow, our approach also offers a novel perspective on the relationship between personality and optimal experience, providing new insights into human psychology.
Related Work
Csikszentmihaly [
17,
18] is the author who has studied this optimal experience or Flow state the most. He claimed that the Flow experience was characterised by a mental state in which a person is totally engrossed in some tasks with full involvement, focus, and enjoyment in the process of the activity. In other words, Flow would be the complete absorption of what a person does and even a results in a transformation of our sense of time. The Flow state is also linked to some activities which involve all of a person’s skills and abilities. According to the author, the Flow state occurs because the person’s full attention is on the task at hand: there is no more attention to be allocated. The concept of Flow has gradually evolved due to interviews and surveys conducted with individuals who described their emotional experiences while engaging in certain activities, resulting in emotions [
21] that they may not have previously experienced which are akin to happiness.
In addition to the Flow state, we can find the concept of the autotelic personality. People with this kind of personality possess some qualities such as curiosity, persistence, low levels of self-centeredness, and a high rate of performing activities for intrinsic reasons only. For this reason, these people are more likely to enter the Flow state than others.
As previously mentioned, the Flow state has been shown to have great importance in sports, music, and employment. Particularly, the latter is receiving extensive attention due to its connection with quality, job satisfaction, and productivity [
22,
23,
24,
25,
26]. Investigations have even detected factors that influence well-being and job satisfaction among employees [
27]. In addition, a positive relationship was found between emotional intelligence and psychological Flow, which turned out to prevent work stress [
28]. Indeed, Hirao and Kobayashi [
29] believe that unemployed people with an autotelic personality have a better quality of life than those without this kind of character.
In this work, several surveys were used to measure up to 26 factors related to various characteristics of individuals, e.g., personal, emotional intelligence, job satisfaction, Flow, etc. Our goal is to investigate whether it is possible to estimate the Flow variable using unrelated variables by implementing several machine learning models to predict and classify the Flow variable, using the rest of the items from different questionnaires.
We emphasise the importance of the Flow state and increasing Flow among employees as it can improve the work environment and job performance [
30]. This could help companies enhance their productivity, organisational quality, and other important factors. Accordingly, it seems to be important to study whether it is possible to predict the Flow score of an individual from other personality characteristics or emotional intelligence, among other factors, i.e., data that companies frequently possess with respect to their workers. In that connection and on the basis of artificial intelligence techniques, we develop the following in this study: provided a set of different variables evaluated in diverse people, we implement machine learning models to predict and classify the Flow variable. Eventually, after conducting the experiments, we draw a conclusion about what the best model is and what the pros and cons of each model are.
To this end, the present manuscript is structured as follows:
Section 2 introduces our methodology, participants, and procedures in addition to the machine learning methods employed.
Section 3 describes the experimentation conducted.
Section 4 includes the results obtained from our experiments. Finally,
Section 5 summarises the main conclusions achieved.
2. Methodology
This study aimed to utilise artificial intelligence techniques to model the Flow variable [
31], a significant factor in the field of psychology. To accomplish this objective, data mining methods were necessary to interpret the vast amount of data gathered from various psychological tests. These tests were designed to measure the participants’ different characteristics.
This task was divided into two different parts. The first was to examine all the variables involved in the problem, while the second was to determine which parameters influence the Flow variable. To achieve this, a predictive correlation study was conducted.
As we have just mentioned, the first step was to analyse the data collected to obtain statistical relationships among the variables of our problem. At this point, our endeavours were focused on linking the Flow variable with other kinds of psychological or personal variables.
Two different approaches were used to forecast the Flow variable for a particular person, and each approach used a different method. The first option was to use regression models since the psychological variables can be treated as continuous variables, and thus each variable can be used in a specific range of applications. The second approach was to employ supervised algorithms in order to classify labelled samples for the Flow variable.
2.1. Method
Participants and Procedures
In our study, we conducted surveys with a total of 856 participants. It is worth noting that we intentionally chose Spanish speakers from around the world to avoid focusing on a specific demographic sample that could potentially bias our study. By selecting participants from a diverse range of backgrounds and geographic locations, we aimed to increase the generalizability of our findings to broader populations. This approach allowed us to obtain a more representative sample. There were 36 variables to be analysed, one of which was the variable to be predicted: Flow.
Each person has a set of 11 variables that are personal and business-oriented with respect to their current employment situation, family, etc. In addition to this, there are 25 other psychologically related variables obtained from the applied surveys: (1) Flow [
32]; (2) self-efficacy [
33]; (3) self-esteem [
34]; (4) extraversion; (5) emotional stability; (6) responsibility; (7) kindness; (8) open-mindedness to experience; (9) flourishing; engagement, [
34] which is split into four variables: (10) vigour, (11) dedication, (12) absorption, and (13) total engagement; (14) satisfaction with life [
35,
36,
37]; emotional intelligence [
38] also divided into four items: (15) perception; (16) comprehension; (17) regulation; (18) total emotional intelligence; and finally, personal and organisational quality [
39], which involves (19) emotional vitality, (20) organisational stress, (21) emotional stress, (22) physical stress, (23) abandonment, (24) health risk and (25) total.
To facilitate understanding for readers, we have summarised each variable and its corresponding test in
Table 1.
All variables might be related to the variables to be predicted, and this was the main issue to solve: to analyse the different relationships among the variables.
These data have a wide variety of information. Furthermore, since they were filled in online, some of them might present several mistakes. Some people may not have answered properly, or there may be null values in the questionnaires. Therefore, we needed to pre-process these flawed values due to the potential influence they could have on our models.
First, we analysed the data provided by the participants to subsequently create social groups and to study links among attributes. The personal information variables of the participants were the following: gender, age, marital and employment status, educational background, business unit, seniority in the company and position, number of dependents, experience, professional group, sector, number of personnel, annual turnover, scope, and recent 12-month situation of their company.
According to the data, the gender distribution was 40% men and 60% women, which is not a significant imbalance. The age of the participants ranged from 16 to 76 years old, with an average age of 44.34 and a standard deviation of 11.29. The majority of participants were married and employed and had at least a Bachelor, Master’s, or PhD degree. Only a few participants had a primary school certificate or non-certificated studies.
After examining all the previous variables, we focused on the psychological variable named Flow. Flow is a state of consciousness which appears when people experience optimal situations; as we mentioned before, it is usually associated with specific tasks in which one needs to involve their attention capacity and other skills. These experiences were defined by scientific research as a state in which a person is utterly absorbed in an assignment [
41]. With the goal of measuring this feature, Likert’s proposal, commonly known as the EBF 9 scale, was adopted. This scale is an abbreviated version comprising nine items for each dimension proposed in the Flow Theory. It draws on a study conducted by Godoy-Izquierdo [
42] in which the authors established a range between 9 and 45 to evaluate this state.
To ensure accurate results, a three-step procedure was employed. Firstly, the data were split into training and validation sets with an 80:20 ratio. Secondly, the models and their parameters were established using the training set and a grid search procedure with cross-validation. Lastly, the trained models were validated, and the results were extracted.
Figure 1 provides an illustration of this process.
Note that all variables were standardised so that they could be applied to distance-based models and, in this way, they were not influenced by the variables whose scales were uneven. Moreover, having all variables on the same scale provides a better understanding of the outcomes and a better classification and prediction.
2.2. Measures
The main metric we used in this study to evaluate the performance of our models was the mean squared error (MSE). Nonetheless, we also used the explained variance and the R2 score to support our statements.
In addition to the root mean squared error (RMSE), the MSE [
43] is one of the most commonly used techniques for evaluating regression models. These two metrics have been adopted in a large number of regression problems. They rely on the error made by each estimated sample
compared with its actual counterpart
. For a dataset whose number of samples is
, this metric can be calculated as follows:
The explained variance (EV) [
42] score allowed us to indicate the importance of our statistical findings. It is quantified by the variance. In this event, the best result would be 1, and it can be formulated according to the next equation:
The last measure employed was the R2 metric [
43], which is also known as the coefficient of determination and is used to measure the predicting success. This metric also has a top limit of 1, and it might be negative too due to the fact that the model can be significantly worse. On the other hand, a constant model, which predicts the expected value of
regardless of the input vector, would obtain an R2 score of 0. The formula to compute this metric is defined below:
These were the metrics adopted in our first approach. The following are the metrics needed to quantify the quality of our results in the second approach. As a consequence, some classification metrics must be defined. The first is accuracy, which is defined as the number of correct predictions made divided by the total number of samples [
44]:
where
and
stand for true positives and negatives, respectively, and
are their false counterparts. Recall [
45] is also used to obtain the number of correct positives divided by the number of all samples that should have been identified as positive:
Precision is the number of correct positives divided by the number of positive results obtained by the model [
45]:
The last measure is the Hamming loss [
46], which is the portion of labels that are incorrectly predicted.
2.3. Machine Learning Techniques
This section is intended to provide a brief introduction to the machine learning method applied in this work. Specifically, we used linear regression, k-nearest neighbours, support vector machine, trees, random forest and artificial neural networks. While these models are typically used to solve classification problems [
47], in this work, we utilized their regression adaptation as the problem we wanted to solve required the analysis of a high range of possible values for the variables.
Our proposed method involved testing several machine learning algorithms for predicting the Flow variable. Each algorithm has its own strengths and weaknesses and may be more suitable for different types of data. By testing multiple algorithms, we can determine which one is the most effective for our specific dataset. We trained and tested each algorithm using a cross-validation technique, ensuring that our results were robust and not dependent on any particular training or testing set. We then compared the performance of each algorithm using the various evaluation metrics previously described. This allowed us to identify the most effective algorithm for predicting the Flow variable and to gain insights into which factors may be driving its predictions.
2.3.1. Linear Regression
Linear regression (LR) is an essential model used in statistics identify the relationship between variables and to analyse their dependency [
48].
Here, this model was used as a reference for the rest of the regression models so as to compare their results. In doing so, we obtained the lower limit that the rest of the models needed to improve. Since this is a simple model, it did not have any parameters to analyse or improve. Briefly, LR is represented by a line which must minimise the error between the training points and the line which is subsequently employed to predict.
In the field of psychology, traditional approaches to modelling the relationship between psychological variables often rely on linear regression analysis. As such, we included LR to ensure that our proposed method could be compared to traditionally oriented methods.
2.3.2. Nearest Neighbours
In this work, we made use of the popular nearest neighbour (k-NN) implementation, based on the relative distances between each data point to compute the expected class [
49]. This algorithm has a parameter k that is set by the user to obtain the best number of neighbours and to obtain a mean value for that value.
This algorithm is widely used in many problems due to its ease of understanding and low computational cost. Although k-NN is mainly adopted in classification problems, this technique can be extended to regression problems. It is based on the sliced inverse regression technique; in this way, local regression using neighbourhoods is carried out.
2.3.3. Support Vector Machine
The third model we applied is the regression version of the well-known support vector machine. This model is designed to solve regression problems by finding the hyperplane that best fits the data [
50], maximising the margin between the hyperplane and the data points. It works by transforming the data into a higher-dimensional space in which a linear hyperplane can be used to separate the data into different regions. The goal is to find the hyperplane that maximises the margin between the hyperplane and the data points while minimising the error.
However, the training process takes a long time in this technique due to the high number of calculations needed. On the basis of SVM models, the support vector regression (SVR) model has benefits in high dimensionality spaces as it is characterised by the optimization-margin algorithm to model different data points.
2.3.4. Tree-Based Models
Another two state-of-the-art methods were used in this study: decision tree (DT) or regression tree (RT) and its ensemble variant, random forest (RF) [
51]. In addition to being widely used in hundreds of fields, it is one of the most descriptive models as it can provide a wealth of information on the data and the relationships between input and output. Its tree-based structure provides a hierarchical representation of the knowledge of the data, which can be used both to predict and to describe how the data behave.
In this case, there were two important parameters to be tested: the depth of the tree and the minimum number of samples required to split an internal node. Additionally, an extra parameter was required for the RF model. RF comprises a set of DTs which combine their output to refine the expected value. As a result, the additional parameter for RF was the number trees used in the ensemble.
2.3.5. Artificial Neural Networks
An artificial neural network (ANN) is a bio-inspired system that builds a computing system by means of nodes, or neurons, and connections, or weights, among those nodes. By combining different values of the weights, an ANN is capable of modelling dependencies between inputs and outputs and predicting future values.
The simplest model is known as a multi-layer perceptron (MLP) network [
52] and has a feedforward architecture. MLPs commonly consist of three main layers: input, hidden, and output layers. Firstly, all the model weights are randomly initialised. Eventually, they change in order to correct the error achieved by the model. Consequently, several iterations are needed so that all the weights can converge until an optimal solution. In our problem, several numbers of layers and neurons were be tested to obtain the best network.
3. Experiments
The data were extracted from the doctoral thesis of Pérez-Moreiras [
53]. First and foremost, we verified that our participants were in the range we highlighted above. Considering
Table 2, we observed that our data fulfilled this requirement. Most were grouped within the higher values, between 30 and 36. This is a good score for this psychological aspect. It is important to note that a higher value indicates better performance for this variable and in our case, very few data points presented a low value of this psychological component.
Once this was achieved, the next point to study was the correlation between variables. To do so, our first steps were to focus on using RF to provide some knowledge about these correlations. The nature of this model allowed us to extract this information. In order to obtain the correlation degree between predictors and independent variables, we analysed how many times these variables were chosen as root nodes. Notice that this node is considered the most important one for our classifier as it is the node with the highest capabilities of slicing and predicting the data.
The variable we analysed was the Flow variable. The remaining variables were used as inputs for our models. All experiments were performed using the RF model, with 1000 trees per classifier to ensure the validity of the results.
As can be seen in
Figure 2, there were three most relevant variables in this problem: V16 (comprehension), V2 (self-efficacy), and V5 (emotional stability). Therefore, we can conclude that these three variables are related to the Flow state, as one needs to have good emotional stability and be quite efficient in order to enter this state. An aspect worthy of mention is that variables such as personal character and business-related variables barely influenced our problem. For this reason, we did not make use of these variables in our models as inputs and we skipped them as our first solution. In this way, it was easier for the machine learning methods to adjust their results as potential noise variables were removed.
Thus far, our efforts had been focused on analysing the problem. Since our variables had a wide range of possible values, the best way to proceed with predicting them was by using regression models. Therefore, this was our first approach to solving the problem.
Each model was tested with different parameters. In the case of kNN, the number of neighbours was verified from 1 to 20. Several kernels were examined in the case of SVR, such as RBF, polynomial, sigmoidal, and linear; the regularisation parameter was set to . The criterion to measure the quality of a split of RT was tested using the MSE, the Friedman version, and the mean absolute error. The RF experiments focused on the number of estimators, setting them up to . For the MLP, several hidden layer sizes were tested, from 1 to 5 layers, and each of them had neurons from 2 to 100; the activation functions were identity, logistic, hyperbolic tan function, and the rectified linear unit function. The optimization methods tested were the stochastic gradient descendent (SGD), a family of the quasi-Newton optimizer (LBFGS) and the Adam optimizer.
Our second approach was to employ multiclass classifiers to categorise different ranges of our variable. In doing so, we transformed our regression problem into a multi-classification problem. There were five intervals or classes. The first one, from 0 to 0.2, was labelled 0 (whose meaning is very low); the second, from 0.2 to 0.4, was labelled 1 (low); the third, from 0.4 to 0.6, was labelled 2 (medium); the fourth, from 0.6 to 0.8, was labelled 3 (high); and the last one, from 0.8 to 1, was labelled 4 (very high). In this way, our psychological variables would have a logical meaning as they measured the intensity of the variables in a specific individual. The greater value of a variable, the higher its strength in that field.
4. Results
The outcomes of the first approach, which used regression techniques to predict the Flow variable, are presented in
Table 3. The table shows that the MLP model performed best, followed by RF in both training and testing, although the RF model performed slightly better in terms of EV, with a difference of eight units to the fourth decimal place. Another remarkable point in this table is the low error achieved by the LR model because, despite being the simplest model, this technique has better prediction than RT.
In order to not hamper the readability of this paper, we will not show all the experiments conducted and will instead endeavour to explain the most important ones. Therefore, we will skip the intermediate tests in order to obtain the best parameters, and we will only list the optimal values attained. For instance, the grid search for the MLP’s parameters chose a single hidden layer with only one neuron and the logistic activation function. It was an unexpected discovery that MLPs obtained better results in training as the number of neurons increased; however, the test error also increased significantly. The notable aspect of this finding was that the threshold in which this event occurs was a single neuron. Considering these results, we can observe a small difference between the R2 metrics for the MLP in the first place and the RF in the second and vice versa for the EV; however, this was true not only for this model but also in comparison with the rest of the models. This dissimilarity was approximately eight-hundredths.
There is a clear difference among the models tested in this study. The best-performing techniques were MLP and RF, with the MLP method achieving the best results in the test stage, indicating the ability of our models to generalise. Notice that ANN training is a computationally costly task as it could overfit easily. For this reason, MLP may have had the lowest MSE in training but not in testing for the EV metric. Nonetheless, it achieved a similar error as the tree ensemble. Although the error in most models is comparatively low, it can be seen that virtually all models obtained a slightly lower error than LR. Having noted this, we obtained another interesting and unexpected finding because the same cannot be said of kNN, SVR, or RT, which achieved worse results despite their higher computational cost and complexity.
In summary, the MLP model was shown to be the most accurate model for predicting the Flow variable. Nevertheless, it is also recommended to use RF due to its low computational cost and high interpretability as it is based on trees.
In our continued efforts to address the problem, we proposed a second approach in which we aimed to classify the different ranges of our variable using a supervised algorithm. As a multi-class problem cannot be solved by a linear model, this technique was removed from our experiments. Therefore, we used SVM, kNN, DT, RF, and ANN as classifiers in this problem. Similar to our previous experiments, our models were tested with several parameters, and the best ones were as follows: linear kernel and a regularisation parameter of 10 for SVM; 17 neighbours for kNN; Gini criterion in the case of DT; entropy criterion and 1000 estimators for RF; and five layers with 200 neurons, using the Adam optimizer, for MLP.
The results of these models can be seen in
Table 4. The resultant metrics are quite impressive. In the training stage, SVM is the second-best technique, according to three out of four metrics, accuracy, recall, and precision, and it obtained a poor Hamming loss. This tells us that if SVM fails, then the difference between the misclassification and its actual value is greater than other models, such as kNN. However, kNN was one of the worst methods for all the metrics in the test stage. This is because kNN is a technique based on distances, and hence it is not rare that it achieved a good score in training by trying to minimise distances. The Hamming loss, in this case, indicates how poor the classification from the model was. Therefore, according to the rest of the measures, kNN does not classify properly either in training or testing, but its approximation was better than DT in some statistics.
Another aspect of these results can be seen with the SVM model. Since it accomplished the second-best precision in training, we can suppose that SVM classified minority classes in a better way than the others, although this fact is not reflected in the test.
Finally, we can observe that MLP achieved the best scores for all metrics in both the test and training stages. MLP’s high accuracy indicates that a greater proportion of its predictions were correct. In addition to this, MLP obtained the best Hamming loss during testing, which means that its incorrect classifications were closer to the correct class than those of the other models. MLP also achieved the best recall score, indicating that MLP correctly classified more positive samples out of all positive classifications. A similar trend was observed with SVM, which had the second-best recall and precision but obtained a worse accuracy, as it seems to have focused on a specific fraction of the data. In conclusion, MLP outperformed the other models overall, although some models that specialised in a portion of the data achieved interesting scores in the training stage.
After obtaining the best model for predicting the Flow variable, the next step was to analyse the specific results of the chosen model which, in this case, was MLP. To do so, we computed the confusion matrix. An illustration of the results can be found in
Figure 3. This figure illustrates the points in which the classifier was not able to predict properly. In most cases, the classes with more misclassified points are classes 2 and 3, and sometimes class 4; seldom do we find class 0 or 1 misplaced. This is a direct result of the oversampling method carried out. In fact, if this technique was not applied, those minority classes would have been misclassified.
At first glance, this fact seems to be caused by an overlapping effect of the adjacent classes, as the predicted label and the true value are placed one unit from its actual value; that is to say, if one focuses on label 2, it can be observed that the failures are obtained by classifying this class in class 1 once and class 3 13 times. This is why the Hamming loss was low for this model; this is noteworthy, as we transformed the resultant values of the Flow variable into different ranges to obtain classes.
We can refer to
Figure 4 to observe the relationship between the predicted labels and the true label. The abscissa axis is the associated label, the ordinate axis displays the normalised value of the Flow variable, and the colour and the horizontal lines define each label within the same range. As mentioned previously, the MLP’s misclassified points in
Figure 4 are closer to the boundaries. For instance, there are 13 points labelled as a large yellow dot in class 2, which is very close to the boundary between class 2 and class 3 and were classified as a label 3. Similarly, the small yellow point indicates that it is near class 2 and was incorrectly classified with that label. The situation is also similar for the purple circles of label 3. Among the eight points that were misclassified, five out of eight are in the limits between the two classes, 3 and 4. This is the highest point. The second point, which is slightly smaller, had three erroneous predictions. This is not located exactly on the class boundary but rather the second-closest location, as it is the case for the last point, which represents one instance classified as a class 2.
These two figures,
Figure 3 and
Figure 4, represent the typical scenario encountered while training models to predict the Flow variable. It is not rare for misclassified points to fall in the middle of the defined interval. For this reason, intelligent models achieved high scores but not perfect marks: it was due to instances that are extremely close to one another.
While our results demonstrate that the MLP model outperformed other models for predicting the Flow variable in our dataset, it is important to consider the theoretical implications of our approach. Our oversampling method, which increased the representation of minority classes in the dataset, allowed our models to better capture the subtle difference in the data and avoid bias towards the majority classes. Furthermore, the MLP’s ability to identify complex relationships between input features and target variables is a key strength of neural networks and is likely a major factor contributing to its superior performance. It is worth noting, however, that the proposed method is not without limitations. Oversampling can lead to overfitting if not applied properly, and ANNs can be prone to overfitting as well. Therefore, it is important to continue refining and validating our approach to ensure its effectiveness and generalisability.
5. Conclusions
In this study, we proposed a machine-learning-based method to analyse a set of psychological tests in order to determine several characteristics in people. We used several surveys to measure up to 26 factors related to multiple psychological features. In particular, we focused on the Flow variable, and we developed models to estimate this item using not-related variables. We implemented linear regression, regression tree, random forest, support vector machine, and artificial neural networks.
In our study, MLP demonstrated the best performance in predicting the Flow variable, with an MSE of 0.007122, an R2 of 0.604601, and an EV of 0.607038, outperforming RF in both the learning and test phases. Despite its simplicity, LR also showed a low prediction error. In terms of classification, MLP was the most accurate model, with a precision of 88.58% and similar performance in recall and precision metrics.
On the other hand, we observed that the misclassified samples were only those located at the very limit of their classes. This suggests that our solution works as a new tool for estimating Flow behaviour in individuals, using variables measured by other tests that are not directly correlated with the Flow measure with high accuracy and precision.
In future work, it would be interesting to investigate the potential application of advanced mathematical tools [
54,
55] in order to improve real-world decision-making problems such as the one presented in this study.