Next Article in Journal
Evaluating Different TiO2 Nanoflower-Based Composites for Humidity Detection
Previous Article in Journal
Assessing Electroencephalography as a Stress Indicator: A VR High-Altitude Scenario Monitored through EEG and ECG
Previous Article in Special Issue
Classification Learning of Latent Bruise Damage to Apples Using Shortwave Infrared Hyperspectral Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessment of Various Machine Learning Models for Peach Maturity Prediction Using Non-Destructive Sensor Data

by
Dejan Ljubobratović
1,
Marko Vuković
2,
Marija Brkić Bakarić
1,
Tomislav Jemrić
2,* and
Maja Matetić
1
1
Faculty of Informatics and Digital Technologies, University of Rijeka, Radmile Matejčić 2, 51000 Rijeka, Croatia
2
Division of Horticulture and Landscape Architecture, Department of Pomology, Svetošimunska cesta 25, University of Zagreb Faculty of Agriculture, 10000 Zagreb, Croatia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5791; https://doi.org/10.3390/s22155791
Submission received: 5 July 2022 / Revised: 30 July 2022 / Accepted: 1 August 2022 / Published: 3 August 2022

Abstract

:
To date, many machine learning models have been used for peach maturity prediction using non-destructive data, but no performance comparison of the models on these datasets has been conducted. In this study, eight machine learning models were trained on a dataset containing data from 180 ‘Suncrest’ peaches. Before the models were trained, the dataset was subjected to dimensionality reduction using the least absolute shrinkage and selection operator (LASSO) regularization, and 8 input variables (out of 29) were chosen. At the same time, a subgroup consisting of the peach ground color measurements was singled out by dividing the set of variables into three subgroups and by using group LASSO regularization. This type of variable subgroup selection provided valuable information on the contribution of specific groups of peach traits to the maturity prediction. The area under the receiver operating characteristic curve (AUC) values of the selected models were compared, and the artificial neural network (ANN) model achieved the best performance, with an average AUC of 0.782. The second-best machine learning model was linear discriminant analysis with an AUC of 0.766, followed by logistic regression, gradient boosting machine, random forest, support vector machines, a classification and regression trees model, and k-nearest neighbors. Although the primary parameter used to determine the performance of the model was AUC, accuracy, F1 score, and kappa served as control parameters and ultimately confirmed the obtained results. By outperforming other models, ANN proved to be the most accurate model for peach maturity prediction on the given dataset.

1. Introduction

In the peach (Prunus persica (L.) Batsch) industry, proper fruit maturity determination at harvest is of prime importance for proper post-harvest manipulation [1,2,3], ensuring quality and consumer satisfaction [1,2,4,5]. Peaches are classified as climacteric fruits, characterized by a rapid increase in ethylene emission and respiration at the onset of ripening, accompanied by color, texture, aroma, and other biochemical changes [6]. Firmness, soluble solids concentration (SSC), and ground color changes are usually the most important methods used by producers for the determination of the harvest date with respect to appropriate maturity. Peach fruits ripen rapidly and have a short postharvest life, usually limited to 3–4 weeks depending on storage conditions [7]. Therefore, the supply chain is primarily focused on prolonging their storability to allow long distance export [8]. The post-harvest performance of peaches is mainly determined by flesh firmness [2,3]. Since melting peaches are very susceptible to rapid flesh firmness decline during the end of ripening and in the postharvest period [6,9], they are usually harvested at early maturity stages. There is a close link between “on-tree physiological maturity” and the evolution of key traits responsible for peach quality during the postharvest phase [1]. Fruits harvested at an unripe stage are more prone to shriveling, internal breakdown, and mechanical damage, and they are of inferior quality when ripe [4]. Consequently, the most important peach producing countries in Europe have lost considerable market shares mainly due to excessive early harvesting [1], while bad organoleptic attributes [5] are the main reasons why consumers do not eat more stone fruit.
The determination of peach maturity using destructive measurements is a slow process that results in fruit destruction, and it is usually conducted only on a certain (smaller) number of fruits, which can affect accuracy. On the other hand, manual sorting according to maturity fruit stage is tedious and time-consuming, and susceptible to discrepancies and inaccuracies if fruits are sorted by different human experts. In the light of that, peach maturity prediction based on sensory non-destructive data would present a notable improvement of various processes. Consequently, fruit maturity could be assessed.
Sensors could be used on a factory line to provide retailers with “up to date” information of a general peach maturity stage. This would help retailers to determine the further course of action in a more precise manner, e.g., penalize the producer with insufficiently ripe peaches, or return the shipment. In order to achieve this, machine learning models are used for data processing.
The application of machine learning to sensory data has already been successfully applied in agriculture. Such management systems, based on machine learning models, provide farmers with real-time recommendations and insights to assist them in the decision-making process [10]. One of the most common applications of machine learning in agriculture is the prediction of fruit maturity. To date, many studies have been conducted to predict fruit maturity using various machine learning models, and machine learning implementation in agriculture has been extensively researched. A random forest (RF) algorithm in combination with explainable machine learning methods was used by Ljubobratović et al. [11] to develop a machine learning model that identifies the most important features for predicting the maturity of peaches to detect nonlinear (and linear) relationships between them. In their study, Scalisi et al. [12] used partial least square (PLS) regression and linear discriminant analysis (LDA) algorithms for peach maturity prediction in different configurations of the spectrometer (fluorescence, near infrared spectroscopy (NIR), and RGB color model). In a study conducted by Sohaib et al. [13], spectral information was used to develop an NIR-based maturity estimator of various fruits (apple, mango, grapes, peaches, pears, and melons) using least squares support vector machine learning techniques. The RF machine learning algorithm was used by Ljubobratović et al. [14] for the prediction of ‘Spring Belle’ peach maturity, while RF and KNN models were successfully established to predict the maturity of peaches during shelf-life in another study [15]. Voss et al. [16] used three machine learning models, i.e., extreme learning machine, KNN, and support vector machines (SVM), for the prediction of peach fruit growth and maturation based on data collected using the E-nose prototype. Artificial neural network (ANN) models were used for fruit maturity prediction and classification in several studies [17,18,19]. However, up to our knowledge, a more detailed analysis of the fruit ripening prediction models and a comparison of their performance in this area has not yet been made. Thus, the aim of this study is to determine the best machine learning model for predicting the maturity of fruits, i.e., in this particular case, predicting the maturity of peaches.
Researchers often encounter a high dimensionality of the dataset, i.e., a large number of predictors, in their studies. To predict the maturity of fruits, Brezmes et al. [20] used the outputs of a large number of electronic nose sensors and then used principal component analysis (PCA) for reducing the dimensionality of results. The PCA method has also been used by Rajkumar et al. [21] to test the variability of the observed data in the studies related to banana fruit quality and maturity stages by using hyperspectral imaging.
Although the PCA method has already been used to reduce the dimensionality of datasets in predicting fruit maturity, one of the main disadvantages of this method is that the learned projective axes are actually linear combinations of all the original features. In this way, it is difficult to give a reasonable interpretation of which features play an important role in prediction [22]. The dataset used in this study included 30 nondestructive measurements on 180 peaches. The measurements are mainly related to peach dimensions and various color indices (Appendix A). Due to the large number of variables, dimensionality reduction was applied to the dataset to remove irrelevant features, as irrelevant features in the data can reduce the accuracy of the model by introducing model overfitting and cause the model to learn based on irrelevant features. To reduce the dimensionality of the dataset, least absolute shrinkage and selection operator method (LASSO) is used.
The group LASSO regularization method, which selects a subset of variables, was also used, and served not as a tool to reduce the dimensionality of the dataset, but as an indicator of certain peaches properties essential for the accurate prediction of peach maturity. Measurements of the ground color of peaches have been shown to have the greatest influence on the prediction of its maturity. Although this is not related to the selection of the best model directly, it helps in understanding the results and indicates a possible direction of future research.
Later, eight machine learning models were trained on the dimensionality reduced set of sensory data in order to predict the maturity of peach. The performance of the proposed models was compared and the model that gave the best results in predicting the maturity of peaches on the given dataset was selected. The models trained and compared in this study are: LDA, logistic regression (LR), classification and regression trees (CART), KNN, SVM, RF, gradient boosting machine (GBM), and ANN.
Therefore, the main objective of this study was to identify a machine learning model from the proposed models that has the best performance in predicting peach maturity using a set of non-destructive input parameters.
The later sections are organized as follows. The methods for measuring peaches and the description of the measured properties are described in Section 2.1. Section 2.2 describes the dataset and the procedure used to reduce its dimensionality (LASSO and group LASSO). Section 2.3 describes the machine learning models and the methods used to select the best model. The results are explained in the third section, and the best model, i.e., ANN, is presented. In the same section, the results are compared to those obtained with a dataset without dimensionality reduction. Section 4 and Section 5 contain a discussion and a conclusion. A complete list of the measured variables can be found in Appendix A.

2. Materials and Methods

‘Suncrest’ peaches of different maturity stages were harvested at the onset of August in a commercial orchard located near the city of Čakovec (Northern part of Croatia). In total, 180 peach fruits were harvested. The peaches were raised as an open vase on vineyard peach used as rootstock. The spacing was 4 m between peach rows and 3 m within rows. In the orchard, standard agro- and pomo-tehnical measures were regularly applied. ‘Suncrest’ peach is a late-maturing variety [23] originating from the USA (CA, USA) [24]. When ripe, it develops intense yellow (ground) and intense bright red skin colour (additional colour), while its flesh is yellow coloured. Additional colour overlays from 50 to 90% of its fruit surface [25].

2.1. Physico-Chemical Properties of Fruits

Immediately after the harvest, fruits were transferred to the laboratory of Department of Pomology at the Faculty of Agriculture of the University of Zagreb in Croatia, where all physicochemical analyses have been conducted.

2.1.1. Ground (GC) and Additional (AC) Fruit Skin Color

On each fruit, ground and additional fruit skin color parameters were measured separately using a colorimeter (ColorTec PCM; ColorTec Associates Inc., Clinton, NJ, USA), according to the CIE L*a*b* and CIE L*C*h° systems (Commission Internationale d’eclairage).
The measurements with the colorimeter were made under laboratory conditions by using instruments and reading the displayed values. Since the measurement conditions were the same for all the samples, no preprocessing of the data was necessary.
In the CIE L*a*b* color space, the L* value corresponds to a dark-bright scale and represents the relative lightness of colors with a range from 0 to 100 (0 = black, 100 = white) [26]. The a* and b* scales extend from −60 to 60, where a* is negative for green and positive for red and b* is negative for blue and positive for yellow [26].
According to Carreño et al. [27], the hue angle (h°) and the chroma (C*) are calculated as given in Equations (1) and (2).
h ° = tan 1   ( b * a * )
C = [ ( a * ) 2 + ( b * ) 2 ] 0.5
where: a* and b*—variables in the CIE L*a*b system.
The hue angle (h°) describes the relative amounts of redness and yellowness, where 0°/360° is defined for red/magenta, 90° for yellow, 180° for green, and 270° for the blue color [28].
From the obtained color values, various ground and additional fruit color indexes were subsequently calculated:
(a) a/b color index.
The a/b ratio is used as a color index for tomatoes, citrus, red grapes, etc., [27,29,30,31]. It is calculated according to Equation (3).
a b = a * b *
where: a* and b*—variables in the CIE L*a*b system.
(b) Citrus color index (CCI).
The CCI color index is described by Jimenez-Cuesta et al. [32], and it is used for de-greening of citrus fruits. It is calculated according to Equation (4).
C C I = 1000 × a * L * × b *
where: L*, a*, and b*—variables in the CIE L*a*b* system.
(c) Tomato color index (COL).
The COL index, described by Hobson [33], is calculated by Equation (5).
C O L = 2000 × a * L * × c *
where: L*, a* and C*—variables in the CIE L*a*b* and CIE L*C*h° systems.
(d) Red grape color index (CIRG1).
This index is designed by Carreño et al. [27] by modifying the index reported in [31]. It is calculated according to Equation (6).
C I R G 1 = 180 h ° L * + C *
where: L*, C*, and h°—variables in the CIE L*a*b* and CIE L*C*h° systems.
(e) Red grape color index (CIRG2).
This index is designed by Carreño et al. [27] by modifying the index reported in [31]. It is calculated according to Equation (7).
C I R G 2 = 180 h ° L * × C *
where: L*, C*, and h°—variables in the CIE L*a*b* and CIE L*C*h° systems.

2.1.2. Fruit Weight, Width, Length, Shape Index, Diameter, Volume, and Density

Fruit weight was measured using a digital analytical balance (OHAUS Adventurer AX2202, Ohaus Corporation Parsippany, Parsipanny, NJ, USA) with an accuracy of 0.01 g. Fruit length and width (mm) were measured with a digital scrolling scale Prowin HMTY0006 on two fruit sides. The fruit shape index was calculated by Equation (8).
Fruit   shape   index = fruit   length fruit   width
Fruit radius was calculated as an average of fruit length and width values. Fruit volume was calculated by Equation (9).
Fruit   volume   ( cm 3 ) = ( 4 3 ×   π ) ×   fruit   diameter 3  
Fruit density was calculated according to Equation (10).
Fruit   density   ( g   cm 3 ) = fruit   mass fruit   volume

2.2. Dataset and LASSO

The main goal of this study was to find the best machine learning peach maturity predicting model for a dataset with 180 observations. The original dataset included 30 nondestructive variables mainly related to peach morphological characteristics and measured by sensors (Appendix A).
According to minimal instrumental parameters, peaches at harvest should have firmness no more than 4.59 kg cm−2 ([7] according to [34]). Hence, in this study, this value was adopted as the firmness threshold. The output (predicting) variable was therefore binary variable ripe derived from peach firmness, representing peaches that have firmness no more than 4.59 kg cm−2.
In addition to the aforementioned ripe output variable, 7 of the remaining 29 variables in this dataset relate to peach weight, density, and dimensions, while the other 22 variables relate to peach ground and additional colors, as described in the previous section.

2.2.1. LASSO

Measurements in this study included 30 variables, which is a lot in relation to the number of measurements. Too many variables can reduce the accuracy of the model and cause overfitting and learning based on irrelevant features [22]. Therefore, the LASSO method was used for feature subset selection in order to increase the accuracy of the models. LASSO is a machine learning technique for selecting a subset of relevant features or variables for constructing a model and eliminating redundant or irrelevant or highly correlated features without much loss of information [22]. LASSO was first introduced by Tibshirani [35] for parameter estimation and variable selection in regression analysis. It is a particular case of the penalized least squares regression with L1-penalty. LASSO, as a feature selection method, focuses on deleting irrelevant or redundant features as opposed to the PCA method that reduces dimensionality by combining features into a smaller number of new, derived features [36]. All calculations, predictions, visualizations, and the LASSO regularization coefficients were performed using the R programming language version 4.13 (R Foundation for Statistical Computing, Vienna, Austria) with the caret, neuralnet, ggplot2, and glmnet packages.
The LASSO regularization uses a modified least squares method in which the regression coefficients are calculated by minimizing the residual sum of squares increased by the sum of the absolute values of the coefficients multiplied by lambda (11) [22,37].
R S S + λ j = 1 p | β j |
Lambda is a tuning parameter and setting it to zero reduces the problem to the least squares method, while a sufficiently large value of lambda yields the null model, i.e., all regression coefficients are zero. The idea is to find a lambda that minimizes the mean squared error (MSE) for the mentioned function. For this purpose, a cross-validation is performed where the lambda value varies between 1010 and 10−2. In this study, the lambda for which MSE was lowest was 0.014 (Figure 1).
Using this method and the selected lambda value, 21 input variables were eliminated from the original dataset, leaving only 8 input variables for model training (Table 1), and predicting the output variable ripe.

2.2.2. Dimension Reduction Using Group LASSO

The LASSO regularization reduces the number of variables in the dataset and retains only those variables that most contribute to the accurate prediction of the output variable.
There is also a regularization that selects specific groups of variables called the group LASSO. First introduced by Yuan et al. [38], the group LASSO is a type of the LASSO regularization for performing variable selection on predetermined groups of variables. In this study, this type of variable subgroup selection provided valuable information on the contribution of specific groups of peach properties to the prediction of maturity.
The data used in this study that consists of 29 variables were divided into 3 subgroups, with the first 7 variables forming Group 1 (morphological properties), the next 11 variables forming Group 2 (ground color properties), and the remaining variables forming Group 3 (additional color properties). List of variables and subgroups are shown in Table A1.
In the same way as for the LASSO regularization, the cross-validation (from the R package gglasso) was used to find the tuning parameter lambda for which MSE is the smallest. The group LASSO regularization was performed using the obtained lambda and it was found that Group 3 (ground color) is the most important for correctly predicting peach maturity. The results of the group LASSO regularization are shown in Table 2.

2.3. Training the Machine Learning Models

Eight different machine learning models trained on 8 input variables obtained by the LASSO regularization were used to predict the maturity of peaches. The output (predicting) variable was the derived binary variable ripe with a value of 1 for the measurements where the firmness of the peach was ≤4.59 kg cm−2 and a value of 0 for the measurements where it was not. The R packages used for prediction computations and visualizations were caret, neuralnet, and ggplot2. All models were trained on an HP ProDesk 600 computer with an i7 (3.2 GHz) processor, 16 GB RAM, and an Intel UHD 630 graphics card.
Before training, the dataset was divided into a training set containing 75% of the data and a test set containing the remaining 25% of the data. Thus, the training set contained 135 measurements and the test set contained 45 measurements.
As a training control mechanism, 7-fold cross validation repeated 3 times was used. Cross-validation is a statistical method for evaluating a learning algorithm in which the data are split into two segments, one used for learning and the other for validation [39]. The basic form of cross-validation is k-fold cross-validation. In k-fold cross-validation, the dataset is divided into k subsets of equal size, one of which is excluded for validation, while the other k−1 subsets are used for model training. Next another subset is taken for validation, while training is done with all the other subsets. After k iterations, each subset was used exactly once as a validation set and the performance of each iteration was determined. Various methods such as averaging can be used to obtain a cumulative result based on the performance of all iterations [39].
Since the training set contained 135 measurements, 10-fold cross-validation would test on subsets with 13 or 14 items, thus yielding less reliable results than 7-fold cross-validation performed on subsets with 19 or 20 measurements.
The selected machine learning models were trained on the given dataset and their performances were compared. The area under the receiver operating characteristic curve (AUC), accuracy, F1 score, and kappa parameters of given machine learning models were compared to select the model with the best results.
Model accuracy is defined as the percentage of correct predictions for the test data.
The F1 score is defined as the harmonic mean of recall and precision [40] as shown in (12), where the best value of the F1 score is one, and zero represents the worst value. According to Sasaki et al. [41], the F1 score was first introduced at the Fourth Message Understanding Conference (MUC-4) in 1992 [40], and its name was derived from van Rijsbergen’s book as the definition of the “F-measure” [42].
F 1 = 2 × P × R P + R
Precision (P) is given in Formula (13), where TP represents true positives (correctly predicted positive outcome) and FP represents false positives (falsely predicted positive outcome). The calculation of recall is shown in Formula (14), where FN stands for false negatives (falsely predicted negative outcome).
Kappa parameter (Cohen’s Kappa) is a measure of model reliability, and it is a useful evaluation metric. Kappa is calculated as given in (13).
kappa   = total   accuracy     random   accuracy 1   random   accuracy
It tries to correct the evaluation bias by considering the correct classification obtained by a random guess. Kappa is within the range [−1, +1], where values closer to one indicate a more precise model.
The area under the receiver operating characteristic (ROC) curve is a two-dimensional measure of classification efficiency. AUC is a scalar measure that shows one aspect of its performance [43]. According to Bradley et al. [44], AUC is one of the best methods for evaluating the performance of a model when a “singular” evaluation is required. The authors in [45] find it to be a better measure of model performance than accuracy. AUC is therefore used as the primary performance measure for the selected machine learning algorithms.
Due to the diversity, models of varying complexity were used:
  • LR and LDA (simple linear models),
  • KNN and CART (nonlinear models),
  • SVM, RF, and GBM (complex nonlinear models), and
  • ANN (deep learning model).
The first two models are simple linear models. The LR model is an extension of the linear regression model in which, instead of fitting the line, the logistic function is used to fit the results of a linear Equation between 0 and 1 [46]. LR is considered a linear model because the boundary of the decision it generates is linear, which can be used for classification purposes [46,47], as is the case in this study. In this study, the generalized linear model from the Caret R package was used to train the LR model.
LDA is linear model used for dimensionality reduction and classification [48]. The LDA technique reduces dimensionality by transforming features into lower-dimensional space, maximizing the ratio of variance between classes and variance within a class, and thus maximizing the distance between classes [49]. According to the same authors, predictions are made by estimating the probability that a new set of input data belongs to each class, where the output class is the one that has the highest probability.
KNN method is a simple classification and regression method that classifies an object by finding the k nearest training examples in a dataset and forms its neighborhood [50]. The output is a class determined by a plurality vote of its neighbors. The object is therefore assigned to the class that is the most common among its k nearest neighbors [51].
CART is a classification technique that creates decision trees from input data, which can then be used to classify new observations [52]. It can also be used for regression, but that was not the case in this study.
SVMs are classifiers that distinguish data objects from two categories, where each object is represented by an n-dimensional vector and belongs to only one of the two classes [53]. The linear classifier separates them by a hyperplane, so the SVM selects the hyperplane with the largest margin to maximize the separation of the two classes [53]. The margin is the sum of the shortest distance between the separating hyperplane and the closest object from the two categories. This classification is then applied when predicting “unseen” or test objects.
RF is an ensemble machine learning model for classification and regression first introduced by Breiman in 2001 [54]. It creates a large number of decision trees by using bagging and randomness of features in the creation of each tree and resulting in an uncorrelated forest of trees whose overall prediction is more accurate than that of any individual tree [54].
GBM is also an ensemble machine learning model that sequentially fits new models to obtain a more accurate estimate of the response variable by converting weak learners (weak models) into strong learners (strong models) [55,56].
ANN is a machine learning model whose basic idea is to simulate the function of the human brain and its basic unit, a neuron [57]. According to Mohammadhassani et al. [57], as in the real world, the ANN model consists of many neurons, each of which generates a set of activations with real value. Although similar, the idea of ANN is not to replicate the work of biological systems, but to use what is known about how biological networks work to solve complex problems [58].
The way ANN works is that the artificial neuron sums the weighted inputs and passes the result to the transfer function to produce the output [58]. This output is then sent to another neuron as input or used directly as a result of the network. Some inputs may be more important than others, and therefore, weights are used that correspond to the importance of each input and provide an effective way to generate ideal outputs.
All models used the 7-fold cross-validation technique repeated three times. To ensure that each algorithm was evaluated using the same data divisions, the value of the random seed number was initially set to the same value before running the models. This resulted in model accuracies that varied widely depending on the value of the seed variable, partly due to the size of the test set. Table 3 shows the AUC results of all models when the seed changes from 1 to 5. The differences are large, e.g., greater than 20% for some models.
To avoid this variability, the models were trained with the seed values set from 1 to 100, and their averages were taken as the performance measure of each model. Thus, the fact that the models were trained 100 times with seven-fold cross-validation repeated three times, making a total of 16,800 training sessions, gives us confidence that the results are not accidental or biased.
After calculating the average AUC, accuracy, F1 score, and kappa value of each model, the one that came closest to these averages was selected, and this model was then used as a representative model.

3. Results

After training the individual models and comparing the average AUC, accuracy, F1 score, and kappa values, the ANN model proved to be the best model with the highest AUC (0.782), accuracy (0.738), F1 score (0.765), and kappa coefficient (0.468), followed by the LDA model with an AUC 0.766, accuracy of 0.730, and F1 score (0.765). The KNN model had the weakest performance with the lowest AUC (0.626), accuracy (0.605), and F1 score (0.653), while the other models AUCs ranged from 0.670 to 0.765. Table 4 shows the averaged AUC, accuracy, F1 score, and kappa values from 100 model trainings with different seed values, sorted by AUC.
The predictive models average AUC and accuracy boxplot comparisons are shown in Figure 2. The graphs show how much the model results vary for different seed values. For example, the KNN model, which proved to be the weakest on average, gave an AUC value of 0.778 and an accuracy of 77.8% for the seed value 56, which can clearly be seen as an anomaly in the graph. In this way, without considering other seed values, one could incorrectly conclude that the KNN model works well on that particular dataset.
The density curves of AUC and accuracy parameters of the trained models are shown in Figure 3. The density curves of the ANN model are evidently left-skewed and the narrowest due to the highest scores, which means that ANN performs consistently well for most seed values.

3.1. Representative Models

The results given in Table 4 are average results per model obtained by 100 training runs with different seed values. In order to plot ROC curves and analyze individual models, for each model a seed that gave the most similar results to the average result was determined. Thus, for the average AUC, accuracy, F1 score, and kappa of each model, a corresponding representative model was found. Because the testing set contained only 45 measurements, the AUC and accuracies of the representative models changed in increments of 0.022. Therefore, for example, the AUC of a representative ANN model was reported as 0.778, although its average value was higher (0.782).
Table 5 shows average score values for individual models and the results of the corresponding representative models for the chosen seed values that most closely match these values.
Based on the results of the representative models, ROC curves were generated for each model (Figure 4). If AUC is approximately 0.5, it means that the model has no discrimination ability, and it is represented by a straight diagonal line. On the other hand, the maximum value for AUC is 1.0, indicating a theoretically perfect model [59].
It is noticeable that the curves of the CART and KNN models are “flatter” and closer to the diagonal, unlike those of the ANN or LDA models that form a larger arc and thus give a larger area underneath, i.e., a larger AUC.

3.2. The Best Model—ANN

This model had an average AUC of 0.782, an accuracy of 73.8%, an F1 score of 0.765, and a corresponding kappa of 0.468 (Table 4). The model that best represents these average results is the representative ANN model with two hidden layers (Figure 5).

3.3. Training the Model on the Entire Dataset

To justify the use of regularization, the predictions of the three models with the best performance were trained on all the available features of the dataset. The machine learning model results trained on the dataset with all 29 input variables compared to the results for the same algorithms trained on the lasso-reduced dataset are shown in Table 6.
All three models showed an increase in performance in all three measured parameters, with the largest increase in LR, where the obtained AUC is 7.14% better when using the LASSO regularization.
Figure 6 shows a graphical comparison of the performances of the best learning models trained on the full dataset and on the dataset reduced with the LASSO regularization.

4. Discussion

After training several models and comparing their average parameters, in this study, the model with the highest AUC, accuracy, F1 score, and kappa coefficient was determined. Although four parameters were considered (AUC, accuracy, F1 score, and kappa), AUC was used as the primary comparison parameter for all eight machine learning algorithms, since it is a better measure of model performance, as reported by Ling et al. [45]. A similar study was conducted by Bradley [44] comparing six machine learning algorithms with six sets of “real-world” medical diagnostic data in order to determine the one with the best performance. In the aforementioned study, AUC proves to be one of the best methods to evaluate the performance of a model on a dataset when a “single number” evaluation is required.
Although the original dataset contained 30 variables, using the LASSO regularization, only eight features were selected to be used for training the models. The main features used for model training were fruit length, fruit shape index, a*-AC, C*-AC, dE2000-AC, L*-GC, a*-GC, and C*-GC. These features are important fruit characteristics that are genetically controlled and thus cultivar-specific (fruit length, fruit shape index) [60] or reflect important characteristics that are highly correlated to fruit maturity (color parameters) [4]. On the other hand, group LASSO found that Group 3 (ground color) was the most important for the correct prediction of peach maturity. This was expected since peach skin ground color is an important maturity prediction tool as it changes along with other important parameters (soluble solids, flesh firmness, and volatile compounds) [2,7]. According to Nascimento Nunes [61], the development of peach blush color is related to the light exposure rather than to the fruit maturation. The fact that peaches can be harvested from different canopy positions and orchards with or without applied nets (different light growing conditions), as indicated in the previous study by Ljubobratović et al. [14], explains why Group 2 (additional color) was not the most important for the correct prediction of peach maturity.
It has been shown that the results of the experiments vary greatly depending on the seed value, and a solution to this problem has been proposed. Models were trained with seed values set from 1 to 100, and average results were taken as the performance of each model.
In many statistical programs, random numbers are “calculated” using the so-called pseudo-random number generators—a recursive method that starts from the initial value determined by an input number called the “seed”. The random number generator in R (the statistical language used in this study) is based on the Mersenne Twister algorithm MT19937 [62]. Using the same seed makes it possible to reproduce the same results in calculations with a random element, e.g., randomly selecting elements from a set. Due to the small size of the dataset, the performance of the models varied widely for different seed values. For example, the AUC value of the CART model varied from 0.489 to 0.822, as shown in Figure 2. Therefore, the use of the mentioned method gave reliable average results that were not random or biased.
In this study, the model with the best results was the ANN model. This model had an average AUC of 0.782, an accuracy of 73.8%, an F1 score of 0.765, and a corresponding kappa of 0.468. The ANN model was also the most complex model in this study. The second-best model (LDA) was a linear model with an average AUC of 0.766. Although this model was one of the simplest, it provided the second-best result in this study. The simplest model in this study, LR, ranked very high in the table with an average AUC of 0.765, slightly less than LDA. The SVM, RF, and GBM models are complex nonlinear models but yielded intermediate results. Although more complex, they did not produce better results than the much simpler LR model. The GBM model was one of the most complex in the comparison, but its accuracy was not the best. This model was obviously not a good fit for the small dataset used in this study. The RF algorithm had an average model AUC of 0.708, placing it in the lower half of our model accuracy table. CART was the second to the worst model at the bottom of the table, i.e., to the KNN model, which gave the worst results in this study with an average AUC of 0.626. All these results are shown in Table 4. The fact that the accuracy, F1 score, and kappa values gave almost identical results for the model ranking confirms the justification of using the AUC parameter as the primary criterion for comparison. Moreover, the ANN model achieved the best results for all three measured parameters.
An ANNs AUC value of 0.782 represents an “acceptable” predictive result, according to Hosmer et al. [47], but to achieve “excellent” or “outstanding” results, it is necessary to add new input parameters or to increase the dataset. In this study, only non-destructive variables were used for peach maturity prediction, given the fact that they have the possibility of being implemented in post-harvest processes with minimal deceleration.
The implementation of these non-destructive measurements in post-harvest processes could be achieved by adding sensors. To automate and speed up the data collection process, color computer vision with ANN could be used to detect the ground color from an image taken with an industrial camera. A similar method was used by Patel et al. [63] to detect bloodstains or dirt stains on poultry eggs. In a study conducted by Jiang et al. [64], neural networks were used to detect five common apple leaf diseases. With the development of convolutional neural networks, visual recognition, such as image classification, localization, and detection has led to excellent performance [65] and would be very suitable for peach ground color recognition. However, this could be applied only to peach varieties that have ground color (as the one in this study), while new non-destructive measurements should be studied for peach varieties with extremely little or no ground color.

5. Conclusions

The comparison of machine learning model training results showed that the ANN model had the best predictive performance. AUCs ranged from the weakest KNN model at the value of 0.626 to 0.782 for the best ANN model. The values of the accuracy, F1 score, and kappa parameters were also considered, producing almost identical ranking results.
In this work, the primary concern was not only to predict fruit maturity, but to find the model that gives the best results on the given dataset, which consisted mainly of the color measurements of peaches. The original dataset contained 29 input variables, and the LASSO regularization method reduced their number to only eight. The measurement showed that this method, in addition to reducing the dimensionality of the set, simultaneously increased the accuracy of the model by more than 2% for the best model, i.e., ANN, and even more for the other models. Since the LASSO regularization proved useful, the group LASSO method was also used. The group containing the measurements of the ground color was selected as the most relevant for the successful prediction of peach maturity from the three pre-determined subsets of variables.
However, including measurements of other non-destructive parameters, such as peach electrical impedance, NIR, spectroscopy, or ‘electrical nose’ might give much better results. In our future research, we will therefore strive to increase the dataset and include other non-destructive parameters.

Author Contributions

Conceptualization, D.L., M.V., T.J. and M.M.; data curation, D.L. and M.V.; formal analysis, D.L. and M.V.; funding acquisition, M.M.; investigation, D.L., M.V. and T.J.; methodology, D.L., M.V., M.B.B., T.J. and M.M.; project administration, M.B.B. and M.M.; resources, T.J. and M.M.; software, D.L.; supervision, M.B.B., T.J. and M.M.; validation, M.B.B., T.J. and M.M.; visualization, D.L., M.V. and M.B.B.; writing—original draft, D.L. and M.V.; writing—review and editing, D.L., M.V., M.B.B., T.J. and M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the University of Rijeka, grant number uniri-drustv-18-122.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. A dataset with the list of variables it contained before the dimension was reduced.
Table A1. A dataset with the list of variables it contained before the dimension was reduced.
FeatureVariable NameDescriptionGroup
fruit firmnessfirmnesspeach firmnessoutput var.
fruit weightfruit_weightpeach weight1
fruit widthfruit_widthpeach width1
fruit lengthfruit_lengthpeach length1
fruit shape indexfruit_shape_indexpeach shape index1
fruit diameterfruit_diameterpeach diameter1
fruit volumefruit_volumepeach volume1
fruit densityfruit_densitypeach density1
L*-ACL_ACL* variable of additional fruit color2
a*-ACa_ACa* variable of additional fruit color2
b*-ACb_ACb* variable of additional fruit color2
C*-ACC_ACC* variable of additional fruit color2
h°-ACh_ACh° variable of additional fruit color2
a*/b*-ACa.b_ACa*/b* additional color index2
CCI-ACCCI_ACCCL additional color index2
COL-ACCOL_ACCOL additional color index2
CIRG1-ACCIRG1_ACCIRG1 additional color index2
CIRG2-ACCIRG2_ACCIRG2 additional color index2
dE2000-ACdE2000_ACdE2000 for additional color2
L*-GCL_GCL* variable of ground fruit color3
a*-GCa_GCa* variable of ground fruit color3
b*-GCb_GCb* variable of ground fruit color3
C*-GCc_GCC* variable of ground fruit color3
h°-GC h_GCh° variable of ground fruit color3
a*/b*-GCa.b_GCa*/b* ground color index3
CCI-GCCCI_GCCCL ground color index3
COL-GCCOL_GCCOL ground color index3
CIRG1-GCCIRG1_GCCIRG1 ground color index3
CIRG2-GCCIRG2_GCCIRG2 ground color index3
dE2000-GCdE2000_GCdE2000 for ground color3

References

  1. Crisosto, C.H.H.; Costa, G. Preharvest factors affecting peach quality. In The Peach: Botany, Production and Uses; Layne, D.R., Bassi, D., Eds.; CAB International: Oxford, UK, 2008; pp. 536–549. [Google Scholar] [CrossRef] [Green Version]
  2. Shinya, P.; Contador, L.; Predieri, S.; Rubio, P.; Infante, R. Peach ripening: Segregation at harvest and postharvest flesh softening. Postharvest Biol. Technol. 2013, 86, 472–478. [Google Scholar] [CrossRef]
  3. Infante, R.; Aros, D.; Contador, L.; Rubio, P. Does the maturity at harvest affect quality and sensory attributes of peaches and nectarines? N. Z. J. Crop Hortic. Sci. 2012, 40, 103–113. [Google Scholar] [CrossRef]
  4. Ferrer, A.; Remón, S.; Negueruela, A.I.; Oria, R. Changes during the ripening of the very late season Spanish peach cultivar Calanda: Feasibility of using CIELAB coordinates as maturity indices. Sci. Hortic. 2005, 105, 435–446. [Google Scholar] [CrossRef]
  5. Crisosto, C.H. How do we increase peach consumption? Acta Hortic. 2002, 592, 601–605. [Google Scholar] [CrossRef]
  6. Minas, I.S.; Tanou, G.; Molassiotis, A. Environmental and orchard bases of peach fruit quality. Sci. Hortic. 2018, 235, 307–322. [Google Scholar] [CrossRef]
  7. Ramina, A.; Tonutti, P.; McGlasson, W.; McGlasson, B. Ripening, nutrition and postharvest physiology. In The Peach, Botany, Production and Uses; Layne, D.R., Bassi, D., Eds.; CAB International: Oxford, UK, 2008; pp. 550–574. [Google Scholar] [CrossRef]
  8. Ceccarelli, A.; Farneti, B.; Frisina, C.; Allen, D.; Donati, I.; Cellini, A.; Costa, G.; Spinelli, F.; Stefanelli, D. Harvest maturity stage and cold storage length influence on flavour development in peach fruit. Agronomy 2019, 9, 10. [Google Scholar] [CrossRef] [Green Version]
  9. Kao, M.W.S.; Brecht, J.K.; Williamson, J.G. Optimum harvest of low-chill melting and non-melting flesh peach cultivars for direct ripening and ripening following low temperature storage. HortScience 2020, 55, 487–495. [Google Scholar] [CrossRef]
  10. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine learning in agriculture: A review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef] [Green Version]
  11. Ljubobratović, D.; Matetić, M.; Vuković, M.; Brkić Bakarić, M.; Jemrić, T. Utilization of Explainable Machine Learning Algorithms for Determination of Important Features in ‘Suncrest’ Peach. Electronics 2021, 10, 3115. [Google Scholar] [CrossRef]
  12. Scalisi, A.; Pelliccia, D.; O’connell, M.G. Maturity prediction in yellow peach (Prunus persica l.) cultivars using a fluorescence spectrometer. Sensors 2020, 20, 6555. [Google Scholar] [CrossRef]
  13. Shah, A.S.S.; Zeb, A.; Qureshi, W.S.; Arslan, M.; Malik, A.U.; Alasmary, W.; Alanazi, E. Towards fruit maturity estimation using NIR spectroscopy. Infrared Phys. Technol. 2020, 111, 103479. [Google Scholar] [CrossRef]
  14. Ljubobratović, D.; Zhang, G.; Brkić Bakarić, M.; Jemrić, T.; Matetić, M. Predicting peach fruit ripeness using explainable machine learning. In Proceedings of the 31st International DAAAM Symposium, Mostar, Bosnia and Herzegovina, 21–24 October 2020; pp. 717–723. [Google Scholar] [CrossRef]
  15. Zhong, Y.; Bao, Y.; Ye, J.; Liu, J.; Liu, H. Combination of unsupervised and supervised models to predict the maturity of peaches during shelf-life. J. Food Process. Preserv. 2021, 45, e15624. [Google Scholar] [CrossRef]
  16. Voss, H.G.J.; Ayub, R.A.; Stevan, S.L. E-nose Prototype to Monitoring the Growth and Maturation of Peaches in the Orchard. IEEE Sens. J. 2020, 20, 11741–11750. [Google Scholar] [CrossRef]
  17. Furferi, R.; Governi, L.; Volpe, Y. ANN-based method for olive Ripening Index automatic prediction. J. Food Eng. 2010, 101, 318–328. [Google Scholar] [CrossRef] [Green Version]
  18. Mazen, F.M.A.; Nashat, A.A. Ripeness Classification of Bananas Using an Artificial Neural Network. Arab. J. Sci. Eng. 2019, 44, 6901–6910. [Google Scholar] [CrossRef]
  19. Hambali, H.A.; Abdullah, S.L.S.; Jamil, N.; Harun, H. Fruit classification using neural network model. J. Telecommun. Electron. Comput. Eng. 2017, 9, 43–46. [Google Scholar]
  20. Brezmes, J.; Fructuoso, M.L.L.; Llobet, E.; Vilanova, X.; Recasens, I.; Orts, J.; Saiz, G.; Correig, X. Evaluation of an electronic nose to assess fruit ripeness. IEEE Sens. J. 2005, 5, 97–108. [Google Scholar] [CrossRef] [Green Version]
  21. Rajkumar, P.; Wang, N.; Imasry, G.E.; Raghavan, G.S.V.; Gariepy, Y. Studies on banana fruit quality and maturity stages using hyperspectral imaging. J. Food Eng. 2012, 108, 194–200. [Google Scholar] [CrossRef]
  22. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning with Applications in R, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2021; 621p. [Google Scholar]
  23. Versari, A.; Castellari, M.; Parpinello, G.P.; Riponi, C.; Galassi, S. Characterisation of peach juices obtained from cultivars Redhaven, Suncrest and Maria Marta grown in Italy. Food Chem. 2002, 76, 181–185. [Google Scholar] [CrossRef]
  24. Krpina, I. Voćarstvo; Nakladni zavod Globus: Zagreb, Croatia, 2004. [Google Scholar]
  25. Miljković, I. Suvremeno Voćarstvo; Nakladni zavod Znanje: Zagreb, Croatia, 1991. [Google Scholar]
  26. Fruk, G.; Fruk, M.; Vuković, M.; Buhin, J.; Jatoi, M.A.; Jemrić, T. Colouration of apple cv. ‘Braeburn’ grown under anti-hail nets in Croatia. Acta Hortic. Et Regiotect. 2016, 19, 1–4. [Google Scholar] [CrossRef] [Green Version]
  27. Carreño, J.; Martínez, A.; Almela, L.; Fernández-López, J.A. Proposal of an index for the objective evaluation of the colour of red table grapes. Food Res. Int. 1995, 28, 373–377. [Google Scholar] [CrossRef]
  28. Pedisić, S.; Levaj, B.; Verica, D.U.; Škevin, D.; Babojelić, M.S. Color parameters and total anthocyanins of sour cherries (Prunus Cerasus L.) during ripening. Agric. Conspec. Sci. 2009, 74, 259–262. [Google Scholar]
  29. Gao, Y.; Liu, Y.; Kan, C.; Chen, M.; Chen, J. Changes of peel color and fruit quality in navel orange fruits under different storage methods. Sci. Hortic. 2019, 256, 108522. [Google Scholar] [CrossRef]
  30. Camelo, A.F.L.; Gómez, P.A. Comparison of color indexes for tomato ripening. Hortic. Bras. 2004, 22, 534–537. [Google Scholar] [CrossRef]
  31. Little, A.C. A Research note: Off on a Tangent. J. Food Sci. 1975, 40, 410–411. [Google Scholar] [CrossRef]
  32. Jimenez-Cuesta, M.; Cuquerella, J.; Martinez-Javaga, J.M. Determination of a color index for citrus fruit degreening. In Proceedings of the International Society of Citriculture, Tokyo, Japan, 9–12 November 1981; pp. 750–753. [Google Scholar]
  33. Hobson, G.E. Low-temperature injury and the storage of ripening tomatoes. J. Hortic. Sci. 1987, 62, 55–62. [Google Scholar] [CrossRef]
  34. Neri, F.; Brigati, S. Sensory and objective evaluation of peaches. In Cost 94: The Postharvest Treatment of Fruit and Vegetables; Commission of the European Communities: Brussels, Belgium, 1994; pp. 107–115. [Google Scholar]
  35. Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  36. Huang, H.D.S. Scalability LASSO & PCA. In Data Analytics—A Small Data Approach; Chapman and Hall/CRC: Boca Raton, FL, USA, 2021; p. 26. [Google Scholar] [CrossRef]
  37. Muthukrishnan, R.; Rohini, R. LASSO: A feature selection technique in predictive modeling for machine learning. In Proceedings of the 2016 IEEE International Conference on Advances in Computer Applications, ICACA 2016, Coimbatore, India, 24 October 2016; IEEE: New York, NY, USA, 2016; pp. 18–20. [Google Scholar] [CrossRef]
  38. Yuan, M.; Lin, Y. Model Selection and Estimation in Regression with Grouped Variables. J. R. Stat. Soc. Ser. B 2006, 68, 49–67. [Google Scholar] [CrossRef]
  39. Refaeilzadeh, P.; Tang, L.; Liu, H.; Angeles, L.; Scientist, C.D. Cross validation. In Encyclopedia of Database Systems; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar] [CrossRef]
  40. Chinchor, N. MUC-4 Evaluation Metrics. In Proceedings of the 4th Conference on Message Understanding, McLean, VA, USA, 16–18 June 1992; pp. 22–29. [Google Scholar] [CrossRef] [Green Version]
  41. Sasaki, Y.; Fellow, R. The truth of the F-measure. Teach Tutor Mater 2007, 1, 1–5. [Google Scholar]
  42. van Rijsbergen, C.J. Information Retrieval; Butterworths: Oxford, UK, 1975; Available online: https://books.google.hr/books?id=EJ2PQgAACAAJ (accessed on 12 March 2022).
  43. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  44. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef] [Green Version]
  45. Ling, C.; Huang, J.; Zhang, H. AUC: A Better Measure Than Accuracy in Comparing Learning Algorithms; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2671, pp. 29–341. [Google Scholar] [CrossRef]
  46. Menard, S. Applied Logistic Regression Analysis; Sage: Thousand Oaks, CA, USA, 2002; Volume 106. [Google Scholar]
  47. Hosmer, D.W., Jr.; Lemeshow, S.; Sturdivant, X.R. Assessing the Fit of the Model. In Applied Logistic Regression, 3rd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2013; pp. 153–225. [Google Scholar] [CrossRef]
  48. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef] [Green Version]
  49. Xanthopoulos, P.; Pardalos, P.M.; Trafalis, T.B. Linear discriminant analysis. In Robust Data Mining; Springer: Berlin/Heidelberg, Germany, 2013; pp. 27–33. [Google Scholar]
  50. Zhang, M.L.; Zhou, Z.H. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef] [Green Version]
  51. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN Model-Based Approach in Classification; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2888, pp. 86–996. [Google Scholar] [CrossRef]
  52. Timofeev, R. Classification and Regression Trees (CART) Theory and Applications Ferda; Humboldt University: Berlin, Germany, 2004. [Google Scholar]
  53. Yu, H.; Kim, S. SVM tutorial-classification, regression and ranking. Handb. Nat. Comput. 2012, 1–4, 479–506. [Google Scholar] [CrossRef]
  54. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  55. Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobot. 2013, 7, 21. [Google Scholar] [CrossRef] [Green Version]
  56. Huang, Y.; Kangas, L.J.; Rasco, B.A. Applications of Artificial Neural Networks (ANNs) in food science. Crit. Rev. Food Sci. Nutr. 2007, 47, 113–126. [Google Scholar] [CrossRef]
  57. Mohammadhassani, M.; Nezamabadi-Pour, H.; Jumaat, M.Z.; Jameel, M.; Arumugam, A.M.S. Application of artificial neural networks (ANNs) and linear regressions (LR) to predict the deflection of concrete deep beams. Comput. Concr. 2013, 11, 237–252. [Google Scholar] [CrossRef]
  58. Jain, A.K.; Mao, J.; Mohiuddin, K.M. Artificial neural networks: A tutorial. Computer 1996, 29, 31–44. [Google Scholar] [CrossRef] [Green Version]
  59. Fan, J.; Upadhye, S.; Worster, A. Understanding receiver operating characteristic (ROC) curves. Can. J. Emerg. Med. 2006, 8, 19–20. [Google Scholar] [CrossRef]
  60. Cirilli, M.; Baccichet, I.; Chiozzotto, R.; Silvestri, C.; Rossini, L.; Bassi, D. Genetic and phenotypic analyses reveal major quantitative loci associated to fruit size and shape traits in a non-flat peach collection (P. persica L. Batsch). Hortic. Res. 2021, 8, 232. [Google Scholar] [CrossRef]
  61. do Nunes, M.C.N. Color Atlas of Postharvest Quality of Fruits and Vegetables; Blackwell Pub: Hoboken, NJ, USA, 2008. [Google Scholar]
  62. Matsumoto, M.; Nishimura, T. Mersenne Twister: A 623-Dimensionally Equidistributed Uniform Pseudo-Random Number Generator. ACM Trans. Modeling Comput. Simul. 1998, 8, 3–30. [Google Scholar] [CrossRef] [Green Version]
  63. Patel, V.C.; McClendon, R.W.; Goodrum, J.W. Color Computer Vision and Artificial Neural Networks for the Detection of Defects in Poultry Eggs. Artif. Intell. Rev. 1998, 12, 163–176. [Google Scholar] [CrossRef]
  64. Jiang, P.; Chen, Y.; Liu, B.; He, D.; Liang, C. Real-Time Detection of Apple Leaf Diseases Using Deep Learning Approach Based on Improved Convolutional Neural Networks. IEEE Access 2019, 7, 59069–59080. [Google Scholar] [CrossRef]
  65. Khan, S.; Rahmani, H.; Shah, S.A.A.; Bennamoun, M. A Guide to Convolutional Neural Networks for Computer Vision. In Synthesis Lectures on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2018; pp. 1–207. [Google Scholar] [CrossRef]
Figure 1. Graphical representation of MSE values obtained by cross-validation used to select the best lambda value.
Figure 1. Graphical representation of MSE values obtained by cross-validation used to select the best lambda value.
Sensors 22 05791 g001
Figure 2. Comparison of AUC and accuracies for all eight models from 100 model trainings with different seed values.
Figure 2. Comparison of AUC and accuracies for all eight models from 100 model trainings with different seed values.
Sensors 22 05791 g002
Figure 3. AUC and accuracy density distributions of compared models.
Figure 3. AUC and accuracy density distributions of compared models.
Sensors 22 05791 g003
Figure 4. ROC curves for representing models.
Figure 4. ROC curves for representing models.
Sensors 22 05791 g004aSensors 22 05791 g004b
Figure 5. Representation of the ANN network with 8 input variables and 2 hidden layers with the output variable ripe.
Figure 5. Representation of the ANN network with 8 input variables and 2 hidden layers with the output variable ripe.
Sensors 22 05791 g005
Figure 6. The graphical comparison of model performance shows an increase in all measured parameters for the models trained on the dataset to which LASSO was applied.
Figure 6. The graphical comparison of model performance shows an increase in all measured parameters for the models trained on the dataset to which LASSO was applied.
Sensors 22 05791 g006
Table 1. A dataset with the list of variables used in model training.
Table 1. A dataset with the list of variables used in model training.
FeatureVariable NameDescription
fruit maturityripepeach maturity (output binary variable)
fruit lengthfruit_lengthpeach length
fruit shape indexfruit_shape_indexpeach shape index
a*-ACa_ACa* variable of additional fruit color
C*-ACC_ACC* variable of additional fruit color
dE2000-ACdE2000_ACdE2000 for additional color
L*-GCL_GCL* variable of ground fruit color
a*-GCa_GCa* variable of ground fruit color
C*-GCc_GCC* variable of ground fruit color
Table 2. Group LASSO regularization preserving the coefficients next to the variables in Group 3.
Table 2. Group LASSO regularization preserving the coefficients next to the variables in Group 3.
VariableGroupGroup Lasso
fruit_weight10.000000000
fruit_width10.000000000
fruit_length10.000000000
fruit_shape_index10.000000000
fruit_diameter10.000000000
fruit_volume10.000000000
fruit_density10.000000000
L_AC20.000000000
a_AC20.000000000
b_AC20.000000000
C_AC20.000000000
h_AC20.000000000
a.b_AC20.000000000
CCI_AC20.000000000
COL_AC20.000000000
CIRG1_AC20.000000000
CIRG2_AC20.000000000
dE2000_AC20.000000000
L_GC3−0.003395380
a_GC30.029737581
b_GC30.005994080
c_GC30.014852482
h_GC3−0.025684634
a.b_GC30.024926216
CCI_GC30.022079283
COL_GC30.022994860
CIRG1_GC30.011642995
CIRG2_GC30.004742768
dE2000_GC30.008545801
Table 3. Different models AUC results for seed values from 1 to 5.
Table 3. Different models AUC results for seed values from 1 to 5.
SeedANNCARTGBMLDALRKNNRFSVM
10.7560.7560.8220.8440.8670.6000.8000.778
20.7780.6670.7560.7560.7330.6440.7110.733
30.7110.7110.7330.7560.7560.6890.7330.689
40.8440.7560.7560.8440.8440.6440.6670.756
50.8000.6000.7560.6890.6890.5780.7110.644
Table 4. Model’s AUC, accuracy, F1 score, and kappa averages.
Table 4. Model’s AUC, accuracy, F1 score, and kappa averages.
ModelAUCAccuracyF1 ScoreKappa
ANN0.7820.7380.7650.468
LDA0.7660.7300.7650.448
LR0.7650.7320.7650.453
GBM0.7140.6750.7240.333
RF0.7080.6750.7220.332
SVM0.6910.6420.6880.267
CART0.6700.6630.7190.301
KNN0.6260.6050.6530.197
Table 5. Comparison of averaged model scores and representative model scores based on the chosen seed values.
Table 5. Comparison of averaged model scores and representative model scores based on the chosen seed values.
ModelRepresentative Model SeedAverage AUCRepresentative Model AUCAverage AccuracyRepresentative Model AccuracyAverage
Kappa
Representative Model Kappa
ANN60.7820.7780.7380.7330.4680.467
LDA580.7660.7560.7300.7330.4480.460
LR30.7650.7560.7320.7330.4530.449
GBM290.7140.7110.6750.6670.3330.322
RF350.7080.7110.6750.6670.3320.328
SVM630.6910.6890.6420.6440.2670.273
CART180.6700.6670.6630.6670.3010.301
KNN290.6260.6220.6050.6000.1970.182
Table 6. Results of the best performing models trained on the full dataset compared to the results of a model trained on the dataset with only 8 input variables (LASSO).
Table 6. Results of the best performing models trained on the full dataset compared to the results of a model trained on the dataset with only 8 input variables (LASSO).
ModelAUC (Lasso)AUC (Full)AUC IncreaseAcc. (Lasso)Acc. (Full)Acc. IncreaseKappa (Lasso)Kappa (Full)Kappa Increase
ANN0.7820.7632.49%0.7380.7182.79%0.4680.4308.84%
LDA0.7660.7314.79%0.7300.6836.88%0.4480.36024.4%
LR0.7650.7147.14%0.7320.6719.09%0.4530.33535.2%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ljubobratović, D.; Vuković, M.; Brkić Bakarić, M.; Jemrić, T.; Matetić, M. Assessment of Various Machine Learning Models for Peach Maturity Prediction Using Non-Destructive Sensor Data. Sensors 2022, 22, 5791. https://doi.org/10.3390/s22155791

AMA Style

Ljubobratović D, Vuković M, Brkić Bakarić M, Jemrić T, Matetić M. Assessment of Various Machine Learning Models for Peach Maturity Prediction Using Non-Destructive Sensor Data. Sensors. 2022; 22(15):5791. https://doi.org/10.3390/s22155791

Chicago/Turabian Style

Ljubobratović, Dejan, Marko Vuković, Marija Brkić Bakarić, Tomislav Jemrić, and Maja Matetić. 2022. "Assessment of Various Machine Learning Models for Peach Maturity Prediction Using Non-Destructive Sensor Data" Sensors 22, no. 15: 5791. https://doi.org/10.3390/s22155791

APA Style

Ljubobratović, D., Vuković, M., Brkić Bakarić, M., Jemrić, T., & Matetić, M. (2022). Assessment of Various Machine Learning Models for Peach Maturity Prediction Using Non-Destructive Sensor Data. Sensors, 22(15), 5791. https://doi.org/10.3390/s22155791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop