Next Article in Journal
Scenic Influences on Walking Preferences in Urban Forest Parks from Top-View and Eye-Level Perspectives
Previous Article in Journal
Reconstruction of Minimum May Temperatures in Northeast China Since 1797 AD Based on Tree Ring Width in Pinus sylvestris var. mongolica
Previous Article in Special Issue
Loading Round Wood in Forestry Trucks and Forestry Platforms: A Case Study for Romania
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Operational Events in Mechanized Weed Control Operations by Offline Multi-Modal Data and Machine Learning Provides Highly Accurate Classification in Time Domain

by
Stelian Alexandru Borz
1 and
Andrea Rosario Proto
2,*
1
Department of Forest Engineering, Forest Management Planning and Terrestrial Measurements, Faculty of Silviculture and Forest Engineering, Transilvania University of Brasov, Şirul Beethoven 1, 500123 Brasov, Romania
2
Department of AGRARIA, University of Reggio Calabria, Feo di Vito snc, 89122 Reggio Calabria, Italy
*
Author to whom correspondence should be addressed.
Forests 2024, 15(11), 2019; https://doi.org/10.3390/f15112019
Submission received: 7 August 2024 / Revised: 9 October 2024 / Accepted: 13 November 2024 / Published: 15 November 2024
(This article belongs to the Special Issue Sustainable Forest Operations Planning and Management)

Abstract

:
Monitoring of operations has become a critical activity in forestry, aiming to provide the data required by planning and production management. Conventional methods, on the other hand, come at a high expense of resources. A neural network was trained, validated, and tested in this study based on multi-modal data to classify relevant operational events in mechanized weed control operations. The architecture of a neural network was tuned in terms of the number of hidden layers and neurons, and the regularization term was set at various values to obtain optimally tuned models for three data modalities: triaxial acceleration data coupled with speed extracted from GNSS signals (AS), triaxial acceleration (A), and speed alone (S). In the training and validation phase, the models based on AS and A achieved a very high classification accuracy, accounting for 92 to 93% when considering four relevant events. In the testing phase, which was run on unseen data, the classification accuracy reached figures of 91 to 92%, indicating a good generalization ability of the models. The results point out that multimodal data are able to provide the features for distinguishing events and add spatial context to the monitored operations, standing as a suitable solution for offline, partly automated monitoring. Future studies are required to see how the capabilities of online, real-time technologies such as deep learning coupled with computer vision can add more context and improve classification performance.

1. Introduction

The management of forests for timber production requires a series of operations aimed at establishing, tending, and harvesting the wood [1,2]. Poplar forests are highly valued worldwide, mainly due to their shorter production cycles, fast growing capacity, and the high-quality wood they provide [3,4], whereas poplar wood has several well-established and emergent uses in industries such as plywood and panel production, replacement of plastic as a material, outdoor construction applications, agri-food, medical, and packaging industries [5]. However, the management of poplar forests typically requires a higher number of operations, compared to conventional forestry [6]. Young poplar forests, for instance, require weed control operations to enhance the ability of trees to compete with the surrounding vegetation [3]. Typically, such operations are implemented several times in the youth stages of the poplar forest stands.
Modern weed control operations are supported by a full mechanization, although some plot-level smaller areas may require additional manual work [7]. For mechanical weed control, commonly used equipment comprises a farm tractor fitted with active devices such as plows or rotary tillers [8]. Of these, rotary tillers have been increasingly used in Romania, due to their capability to comminute and mix the weeds with the soil.
On the other hand, forest managers charged with weed control operations often lack a decision support system that can estimate the time needed for operations, provide the spatial context for planning, allocate resources such as money and fuel, and provide updates on the level of operational completeness. As a consequence, the planning and monitoring of operations is usually performed using conventional methods and protocols, which require manual inputs of data, incur a low level of data processing automation, and necessitate extensive paperwork.
The developments in machine learning for forest operations activity recognition has provided new ways to collect, process, and interpret the data, by removing the human effort and expertise required to a large extent [9,10]. These advancements are welcome as forest management is increasingly focusing its resources on production and activities that directly contribute to income. However, this brings a lack of well-trained personnel that need to make an effort to develop statistics and models to support the planning and implementation of the operations. Moreover, the improvements in operational equipment, as well as the implementation of different technologies in new areas [11,12], are progressing at an unprecedented rate. This makes it very difficult to keep the pace with studies that evaluate the operational performance, whereas the knowledge provided by the existing studies has only a limited applicability in new contexts [13]. Commonly used algorithms such as neural networks, random forests, and support vector machines are now available for implementation in several freeware software platforms [14,15,16]. Such algorithms may be trained to work with multi-modal data [17], and significant progress has been made in terms of using these data in recognizing activities with high accuracy and adding temporal and spatial context to classification problems, which are essential for operational planning, implementation, and monitoring [18,19].
Since farm tractors enable multifunctionality, and they are a common choice due to cost affordability [20,21,22]; developing and using systems able to collect, pre-process, and transfer the information for decision making is less likely to happen for these machines. Developing such systems could provide contextualized operational classifications in the time and space domains but will also increase the machine cost. In other words, algorithms that may correctly detect the tasks forming an operation with a given active device such as a plow could be less effective, for instance, when using a rotary tiller, mainly due to the mechanical behavior of the two. In addition, farm tractors are typically used for a long time with the aim of recovering the investments in machines, and many owners prefer to keep them in use at the expense of higher maintenance and repair costs, which means that an important share of the fleet is still dominated by old machines. Commonly, these machines are not equipped with devices able to collect useful data on their state such as CAN-BUS [23], and among the possible solutions for activity recognition and monitoring are those making use of external sensors.
On the other hand, the task of collecting and classifying time study data is typically transferred to the science of forest operations, since there is a lack of qualified personnel to undertake such jobs. Performing it conventionally, this activity is prone to error [24] and uses up important resources [25]. As a consequence, at least in the short term, a system able to collect and process data will be limited to offline applications able to prove concepts and, eventually, to produce data and statistics that support a full-scale, online deployment. To this end, conventional machine leaning models can be used in an offline approach to make accurate and contextualized predictions on relevant operational events based on multi-modal data such as that combining GNSS, which is able to contextualize location and movement parameters, and accelerometer data, which are able to document, by magnitude, specific events showing contrasting patterns in the time domain [26,27,28].
The goal of this study was to check whether the combined use of GNSS and triaxial accelerometer signals can provide a suitable input for an accurate neural network classification of relevant operational events in the time domain for a rotary tiller-based weed control operation. Two objectives were set for this study: (i) to run tests by trial-and-error, aiming to see which set of signals (modality) works best in producing an overall highly accurate classification, which was performed by iteratively training and validating the neural network on various sets of input signals while varying its architecture and tuning its parameters; and (ii) to test the best performing neural network models on unseen data, so as to check their generalization ability, which was achieved by saving the best performing model of each set of input signals and feeding it with unseen data.

2. Materials and Methods

2.1. Data Sourcing and Processing

The data required by this study were collected in Romania in June of 2018. In this country, most of the poplar forests are concentrated along the Danube River [29], and their management involves operations such as preparing the soil, planting, cultivation and weed management, thinning, and final harvesting by clearcutting [30,31]. The field study was carried out in the forests managed by Poiana Mare Forest District, which is under the administration of Dolj Forest Directorate of the National Forest Administration–RNP Romsilva (Figure 1). According to the Romanian management of poplar forests, final harvesting is performed at an age of 25–30 years by clear-cutting implemented in areas that do not exceed 5 hectares. Once harvested, these areas require stump removal, soil preparation, planting, which currently is achieved partly mechanized [32], and, at given time intervals, weed control operations, which can be implemented manually, mechanically, or as a combination of both [7,30]. The decision on the degree of mechanization in weed control operations depends on the size of the plots, availability of machines, and manual labor requirements. In Romania, the poplar stands are harvested by clear cutting at the end of rotation, and the harvested plots have a limited size of up to 5 hectares.
A farm tractor equipped with a rotary tiller was taken into study, as shown in Figure 2, since this equipment configuration is being increasingly used in the country, mainly due to the operational performance and quality of the work. The tractor was a Romanian-made U650 model (UTB—Uzina Tractorul Brașov, Brașov, Romania) manufactured in 2004, equipped with a diesel engine and featuring an engine output of 47.8 kW. A Maschio Gaspardo W 145 (Maschio Gaspardo S.p.A., Campodarsego, Italy) rotary tiller was attached to the tractor using the power take-off. It was manufactured in 2016, and designed to operate on strips with a width of 145 cm at a depth of 18 cm. Its operation is sustained by a power range of 18–30 kW.
The data collectors used were a GNSS handheld unit which was placed on the tractor cab and set to collect data at a rate of one second, a triaxial accelerometer which was placed on the rotary tiller and set to collect data at a rate of one second, and a miniaturized camera that was placed in the cab with the field of view oriented towards the rotary tiller so as to also see the activity of the driver, and set to collect video data continuously. The technical features of the equipment used to collect the data, as well as the protocols used to collect, pair, and process data are similar to those described in [26,27,32].
At the office, GNSS and accelerometer data were paired in a Microsoft Excel sheet based on the time labels collected by both devices. Since both devices provided samples at a rate of one second along with their corresponding time labels standing for the date, hour, minute, and second of sampling, these data were used to pair the samples collected by the two devices. Then, video data were carefully analyzed to document the database by the main operational events. The observed work, which is typical to weed control operations, includes the advancement of the farm tractor equipped with the tiller on the space between two tree rows. After arriving at the plot’s headland, maneuvers are taken to exit a given operated strip and to enter on a new one. Besides these types of events, other maneuvers can be taken in the plot, as well as full stops with the engine of the tractor running without power transmission to the tiller, or full stops with the tractor’s engine off.
Based on the video data visualization, the work was divided into relevant operational events. These included (Table 1): (i) the equipment completely stopped with the engine turned off, which is typical to various delays, equipment preparation for work, meal time and other breaks, (ii) the tractor with the engine on, but with no movement of the equipment, which is typical to short breaks taken in the plot or at the headland, (iii) maneuvering, which is required at the headlands to exit an operated strip and to enter on a new one, or in the plot to pass over various obstacles and to change the row of operation, as well as (iv) active work, during which the tractor moves while powering the tiller which operates the soil. As observed in the field and in the video files, the maneuvers were taken with disengagement of the power provided to the tiller while the latter was lifted from the ground when maneuvering.
According to the time consumption classification in forest operations [19], the events described in Table 1 stand, by cumulation in a given day, for the gross time consumption. This time category along with those operated during that time may provide important figures on the gross productivity. Taken together, active work and maneuvering provide a good approximation of the net time consumption which, used along with the operated area, may provide estimates on the net productivity. The time spent by the equipment in events with no movement can be used as a good approximation of other categories of time consumption which are relevant to understanding the operational performance, such as repair and maintenance time, meal and other breaks time, and delays, respectively. However, to plan the operations, science and practice are typically interested in the net time, since it is used to develop piece-rate systems used to evaluate and pay for the work being performed [33], whereas accurately documenting different kind of delays typically requires extensive field data [34].
In total, the collected dataset had 15,688 records taken at a rate of one second. Before data annotation, and based on the analysis of video files, 1092 records were removed since they covered the time used to prepare the study. Although this time could be documented separately by conventional studies [18], it is rather irrelevant when monitoring the operations. The annotated dataset used for machine learning contained a number of 14,595 records.

2.2. Prediction of Operational Events by Machine Learning

The annotated dataset was arbitrarily divided in two parts, using the first 10,003 records to support the first objective of this study (hereafter called dataset 1–DS1), and using the last 4592 records to support the second objective of this study (hereafter called dataset 2–DS2). The division of the initial dataset was conducted whilst having in mind that each dataset should contain data for all of the documented events, as well as the fact that a given sequence of records belonging to a given event in the time frame should not be split and allocated to different datasets, this way preserving the original pattern in data.
GNSS data provide several attributes such as the location, sampling rate, heading, speed, and sampling date and time. These were extracted from GPX files with the use of Garmin BaseCamp (Version 4.7.0) software, in a way which was similar to [27,28]. For activity recognition and monitoring, besides the location, movement speed is important to distinguish between given operational events. As such, this feature was used as an input signal for machine learning, and was called S (km/h). Triaxial accelerometer data provide readings on the three axes (x, y and z), as well as an aggregated measure called the vector magnitude, which is computed from the triaxial measurements, as generically described in Equation (1). An important feature of vector magnitude is that it accommodates the effects of a sensor’s orientation; therefore, it virtually removes the inconvenience related to the location at which the sensor is placed on different machines, or on the same machine on different days. Vector magnitude was used as an input signal for machine learning, and it was called A (g). In total, three approaches were possible to support the first objective of the study: using the A and S data as features taken together (hereafter called modality AS), using solely A data (hereafter called modality A), or using S data (hereafter called modality S).
A = x 2 + y 2 + z 2 2
where A is the vector magnitude, and x, y, and z are magnitudes measured on the x, y, and z axes, respectively.
Three main steps were considered in this study, which were adapted to the common language used in machine learning. Accordingly, training and validation are typically the steps used to check the performance of models before feeding them with unseen data, whereas later, testing is conducted to check for their generalization ability. As a first step, the three modalities were checked for accuracy by training and validation on DS1. For that, the functionalities of the Orange Visual Programming [14] (Version 3.31.1) software were used. The workflow used to test and validate the best modalities in terms of classification performance is described in Figure 3. The neural network functionalities supported by the used software can be configured in terms of architecture and tuned for hyperparameters. For each modality, the architecture was configured so as to have 1, 5, and 10 hidden layers. For each option in terms of the number of hidden layers, the number of neurons was set at 10, 100, and 1000, respectively. The Rectified Linear Unit (ReLU) as well as the stochastic gradient descent optimizer (ADAM) were used for all these options as a standard activation function and solver, respectively, due to their performance and popularity in machine learning [35,36,37]. The number of iterations was set to the maximum enabled by the software (1,000,000) and, for each option the regularization term was tuned at α = 0.0001, 0.001, 0.01, 0.1, 1, and 10. By the number of modalities, number of hidden layers, number of neurons per layer and regularization term values, a total of 162 models were trained and validated (54 models for each modality).
To select the best model, the performance of classification was evaluated based on the overall classification accuracy (CA). Using the “Test and Score” widget, the software enables the calculation of the commonly used metrics for classification performance, most of which are described, for instance, in [38,39]. They include the classification accuracy (CA), precision (PREC), recall (REC), and cross-entropy (LOSS). In addition, the software enables the computation of training and validation time. Classification accuracy was used as the main metric to characterize the classification performance because it indicates the share of correctly predicted instances of the total dataset [39]. In the steps of testing and validation, the overall classification accuracy was used as a criterion to select the best performing models.
To supplementarily document the overall classification performance, training (TT) and validation (VT) time, the area under the ROC curve (AUC), classification accuracy (CA), F1 score (F1), precision (PREC), recall (REC), specificity (SPEC), and cross-entropy (LOSS) were computed and reported for each model as well. The steps of training and validation were performed by a stratified cross-validation using 10 folds. Cross-validation is a technique used to prevent overfitting by dividing the dataset into a number of folds, of which one is retained for validating the results trained on the rest; the procedure is repeated multiple times by using a different fold for validation each time, and then the results are averaged to produce more robust figures on a model’s performance [40]. Training models by cross-validation is also a good strategy to deal with datasets that are limited in size [41].
Procedurally, all 162 models were first trained and validated. Based on the figures on classification accuracy (CA), a final model was retained and saved for each modality. Then, these three models were used to make predictions on unseen, although annotated, data. The workflow used for making predictions on unseen data is described in Figure 4.” The workflow interconnects the data loaded by a “Data” widget with a given model loaded by the “Load Model widget and with a “Predictions” widget, which pairs the true classes of unseen data with those predicted by the model and returns the main metrics of classification performance. In addition, the workflow supports the computation of a confusion matrix, which may be used to map the absolute or relative frequency of actual instances against the predicted ones.

2.3. Computer Architecture and Data Visualization

All the tasks involving the training and validation of the models were carried out on a Dell Alienware 17 R3 machine equipped with an Intel © Core TM i7–6700HQ CPU, featuring 2.60 GHz, 4 cores, and 8 Logical Processors, with an installed physical memory of 16 GB, under a Microsoft Windows 10 Home edition operating system (Microsoft, Redmond, WA, USA). Effects brought on by the modality, network architecture, and hyperparameters were visually assessed using the “Scatter Plot” widget connected to a “Data” widget which was used to load the necessary information from a Microsoft Excel file. The steps of pairing the GNSS and accelerometer data, as well as coding of data according to the events seen in the video files, were supported by Microsoft Excel.

3. Results

3.1. Training and Validation

The main results characterizing the classification performance during the training and validation of the models are included for the three modalities in Table 2, where the optimal models are described in terms of architecture and hyperparameters. Using AS and A modalities to train and validate the models provided similar classification performances. For these two modalities, the overall classification accuracy (CA, OVERALL, Table 2) reached values of 92.8 (modality A) to 93% (modality AS), indicating the benefits of using the acceleration as data input for classification. When using both acceleration and GNSS speed, the classification accuracy improved for maneuvering (MAN, 98.2%), stop (STOP, 94.2%), and off (OFF, 97%) events, emphasizing the effect of class imbalance in DS1, where “Work” events accounted for 7261 observations (73% of DS1), as well as the increased power of the A signal to discriminate the events. In the same dataset (DS1), maneuvering (MAN, 2046, 20%), stop (STOP, 299, 3%), and off (OFF, 397, 4%) events accounted for less than 25%. Overall, the classification accuracy for modality A was lower, accounting for 92.8%. However, “WORK” and “STOP” events were better classified, accounting for classification accuracies of 97.8 and 97%, respectively. For these models, the multi-class cross-entropy (LOSS, Table 2) had values of 0.19 and 0.23, respectively, pointing out that working with AS modality may provide lower errors during generalization.
Using the S data alone led to an important drop in classification accuracy (OVERALL, CA = 84.8%, Table 2), as well as to a high increment in multi-class cross-entropy (OVERALL, LOSS = 0.463, Table 2), indicating a limited ability of the GNSS speed alone to accurately classify the events taken into study. At the event class level, the classification accuracy was better for events that did not involve a machine’s movement such as “STOP” (CA = 97%, Table 2) and “OFF” (CA = 96%, Table 2), indicating that lower ranges of GNSS speed may accurately predict such events. “WORK” events were still highly accurately classified (CA = 91.1%, Table 2), maybe due to a differentiation in the magnitude of speed, but the overall result in classification accuracy may be the consequence of “MAN” events being misclassified based on speed, since this event class had a classification accuracy of 85.5%.
There were differences in the classification performance of the trained and validated models, as shown in Table A1, Table A2, Table A3 and Figure A1. Training and validation time also increased in direct relation to the complexity of the neural network’s architecture (Figure A2) and the value used for the regularization term. Training and validating all the models presented in Table A1, Table A2 and Table A3 took about 92 h, a time which was quite equally shared between the models using AS data (approximately 30 h), A (approximately 29 h), and S (approximately 33 h) modalities. Of these, working with complex architectures (five or ten layers containing 1000 neurons each) took more time; it accounted for 7.6 (A), 8.1 (AS) and 9.7 (S) hours when working with a neural network of five layers holding 1000 neurons each, and for 19.1 (A), 20.4 (AS), and 21.6 (S) hours when working with a neural network of ten layers holding 1000 neurons each. These architectures did not improve the classification performance and, excepting the models based on the S modality, regularization terms set at 1 and 10 returned poorer classification performances.

3.2. Testing on Unseen Data

The main results characterizing the models’ performance on unseen data are included for the three modalities in Table 3. Classification performance on unseen data is important to understanding how these models will work in real operational conditions. The testing dataset (DS2) contained a number of 4592 observations covering three classes of events: “WORK” (3235, about 71% of the data), “MAN” (1243, about 27% of the data), and “OFF” (114, about 2% of the data). As shown, the order of classification performance changed, pointing out that the model based on the A modality would perform better.
The classification performance would not be particularly different when using models based on AS and S data, respectively. These returned a classification accuracy of 90% on unseen data, while the model based on modality A returned a classification accuracy of 92% (Table 3). However, the error levels were preserved at similar levels and in a similar order of magnitude among the AS, A, and S models, indicating a higher error in the case of the S model working on unseen data (LOSS = 0.352, Table 3). Overall, the classification accuracy of AS and A models on unseen data was close to that of the training and validation phase (Table 2 and Table 3), with differences that accounted for 2.4 and 0.6%, respectively.
Figure 5 shows the confusion matrices of the AS, A, and S models when working with unseen data. For instance, the AS model tested on unseen data confused “WORK” events with “MAN” events for a number of 73 observations, which may be due to similar movement speeds. In some other cases, it confused “MAN” with “OFF” events, which may also be due to a lower magnitude in acceleration and similar magnitudes in speed. As such, irrespective of the model used, there was some degree of inter-class similarity in the data, leading to confusion from the models.
In all cases, “WORK” events were not confused as “OFF” events (Figure 5a–c). Interestingly, the model based on S data was better at setting apart the data from “OFF” events since, in contrast to AS and A models, it did not confuse “MAN” with “OFF” events.

4. Discussion

Monitoring forest operations by advanced techniques is important and increasingly required in practice and science to remove the subjectivity of assessments and the errors that may occur due to data collection over long periods of time, as well as to control the costs incurred by collecting, processing, and analyzing the data. It may also serve as a better method to accommodate safety concerns when working nearby dangerous machines.
The results of this study indicate that the type of data sourcing modality is important in the performance of a neural network designed for event-based classification tasks. This is supported by the classification performance outcomes when using the vector magnitudes of triaxial acceleration coupled with GNSS speed as data inputs, for which the overall classification accuracy in the training and validation phase was of 93%. Using this combination of input signals also provides the context required for the localization of operations since the GNSS data are documented with features such as geographical coordinates, altitude, and heading. As such, the acceleration data are more likely to provide the distinctive features needed to identify the relevant events, whereas the GNSS speed helps in improving the accuracy of classification, particularly in events that involve movement. However, by comparing the classification accuracies of the best performing neural network models on DS1, it seems that the GNSS speed makes only a marginal contribution to the improvement in classification accuracy in the training and validation steps. For instance, the classification accuracies of the AS, A, and S modalities were 93, 92.8, and 84.8%, respectively, indicating that using the GNSS speed alone comes at a high expense in event-based classification accuracy. Also, using the vector magnitude of the acceleration signal alone provides a classification accuracy close to that provided when combined with the GNSS speed but, by using this approach, the spatial context of the operations is lost.
In terms of the neural network’s architecture and hyperparameter tuning, the AS and A modalities required a single hidden layer to achieve the best classification performance, whereas the S modality required a network with five hidden layers to show its best performance (Table 2, Table A1, Table A2 and Table A3). This hints at how the neural networks managed the complexity stored in the signals, indicating that accurately learning from the S modality with a single hidden layer was much more difficult to achieve, requiring more hidden layers but with fewer neurons. Also, finer regularization terms (α = 0.001 and 0.0001) were the best option to increase the classification accuracy based on the AS and A modalities, respectively, whereas a coarser regularization term (α = 1) was more suitable for the S modality. Nevertheless, very coarse regularization terms (α = 10) generally led to a dilution in classification accuracy, irrespective of the modality, depth, and width of the networks used (Table A1, Table A2 and Table A3). Neural networks using the AS and A modalities required a higher number of neurons per layer to give their best classification performance. In this regard, classification accuracy based on AS data worked the best with 1000 neurons, whereas classification accuracy based on A data worked the best with 100 neurons. In comparison, the neural network using S data required only 10 neurons to show its best performance. This may be another hint at how data complexity stored in the modalities’ signals was handled, since it seems that arriving at the best classification accuracy depended on a decreasing number of weights in the order AS, A, and S, respectively. It is worth mentioning that training and validating the neural networks using 5 and 10 hidden layers of 1000 neurons each did not provide additional benefits in terms of classification accuracy but used up important computational resources, irrespective of the modality used as data input.
The performance of the models on unseen data is essential to understanding how they will behave in real world applications. Although there was a minor dilution in classification accuracy, the tuned models of the AS, A, and S modalities showed a high classification accuracy in predicting the events on unseen data. While this confirms the generalization ability of the trained and validated models, the most frequently confused events were “WORK” and “MAN”, pointing out the inter-class similarity of the data. As a potential solution, these two classes could be merged to improve the classification performance. However, this approach would have led to missing an important work element for time study classification, and it would have inflated the time spent as main work time [18,19]. What is clear is that the “OFF” events were never misclassified as “WORK” or “MAN” events, which is promising considering that the time spent in such events stays outside the productive time [42],. From this point of view, separating the productive time from other time consumption categories would probably lead to an improved classification accuracy since working with binary classification problems based on triaxial acceleration data typically returns close to 100% in classification accuracy [26,43].
Poplar forests, on the other hand, are located on sites showing a wide variability in terms of soil condition [44], weed characteristics, and spacing of the trees [7,44], all of which may at least affect the distribution of events on classes, as well as the magnitude of the acceleration signals recorded. To these factors, one can add the variations in magnitude of acceleration due to the type of rotary tiller used, and particularly due to its size class. To some extent, these variations can be accommodated by the way in which the neural network models are built, since they use a standard data preprocessing flow that runs a standardization before training [45]. Data standardization provides input datasets that are scaled so as to have a mean value of zero and a standard deviation of 1, a procedure that is robust to outliers [46]. This is helpful to balance the effect of high magnitude observations, which adds to the benefits of using the vector magnitude of acceleration data to remove the effects of datalogger orientation [9,26]. Then, the GNSS signal in terms of the magnitude of the speed could be affected only by the operational conditions, for which sustaining much higher speeds than those observed in this study is less likely. As a fact, the moving speeds of machines used in short rotation forests are typically low and in the range of those observed in this study [47,48,49,50,51].
The main limitation of this study is related to the type of the dataset used. Ideally, robust models would consider variations that may be brought on by the type of machine used, the experience of the operators, and practice in various areas. For the time being, accounting for all these factors when developing the machine learning models is challenging. This comes mainly from the fact that there is no available data annotated for the operational monitoring of operations such as in this study, which is a common challenge in machine learning applications [17]. In addition to a resource-intensive process and the expertise required to annotate the data, large datasets may require an increased computational power to be able to train the models. In this study, training and testing took from a couple of minutes to a couple of hours, and there was a trend in the magnitude of computational resources depending on the architecture used for the neural networks. Future studies could accommodate this limitation and check whether there will be a significant difference in the classification accuracy of the models trained on newcoming data. Due to the way in which deep learning models are built, they are likely to perform with higher classification performances, an option that should be kept in mind in future studies. For instance, valorizing the available knowledge and models that shape the science of deep learning and computer vision holds a lot of potential for better understanding the events and making better and more contextualized predictions based on a combination of 1D and 2D channel data.

5. Conclusions

The classification performance of mechanized weed control operational events by conventional neural network machine learning was found to be high based on triaxial acceleration and speed data extracted from GNSS signals in both training–validation and testing phases. This option enables forest managers to obtain accurate time- and space-contextualized estimations of the machine performance and will help in removing many of the less efficient processes required to collect, process, and analyze time study data. Provided that an offline option will be sufficient, the methods described herein can solve the problem of operational monitoring. For more advanced tools, multi-modal online solutions could be researched, as well as how well deep learning and computer vision perform would in operational classification.

Author Contributions

Conceptualization, S.A.B.; methodology, S.A.B. and A.R.P.; validation, S.A.B.; formal analysis, S.A.B. and A.R.P.; investigation, S.A.B.; resources, S.A.B. and A.R.P.; data curation, S.A.B.; writing—original draft preparation, S.A.B. and A.R.P.; writing—review and editing, S.A.B. and A.R.P.; visualization, S.A.B.; supervision, S.A.B. and A.R.P.; project administration, S.A.B.; funding acquisition, A.R.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data supporting this study may be made available upon a reasonable request to the first author of the study.

Acknowledgments

The Authors acknowledge the support of the Department of Forest Engineering, Forest Management Planning, and Terrestrial Measurements for providing the equipment needed for data collection. Some activities in this study were funded by the inter-institutional agreement between Transilvania University of Brasov (Romania) and Mediterranean University of Reggio Calabria (Italy). Also, the Authors would like to thank Dr.eng. Marius Cheţa and to Dr.eng. Tiberiu Marogel-Popa for their support with data collection.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Classification performance of the AS modality in the training and validation phase.
Table A1. Classification performance of the AS modality in the training and validation phase.
ModalityNnαTTVTAUCCAF1PRECRECSPECLOSS
AS1100.0001113.0130.0180.9850.9260.9110.9010.9260.9660.198
1100.001108.2730.0150.9850.9250.9100.9000.9250.9660.199
1100.01108.5260.0170.9850.9200.8990.8950.9200.9640.204
1100.160.6540.0160.9800.9130.8850.8650.9130.9620.234
110132.5980.0180.9550.9070.8780.8560.9070.9380.292
1101016.6590.0160.9350.8590.8270.7970.8590.7940.455
AS5100.000157.9920.0210.9860.9280.9150.9040.9280.9650.191
5100.00164.7080.0320.9860.9280.9160.9050.9280.9650.191
5100.0161.3880.0230.9850.9280.9150.9050.9280.9650.193
5100.166.9640.0200.9850.9280.9150.9040.9280.9640.193
510180.3140.0330.9780.9140.8850.8640.9140.9570.239
5101045.3760.0220.9360.8640.8320.8030.8640.8100.442
AS10100.000158.5850.0290.9830.9180.9080.9000.9180.9670.208
10100.00159.4740.0300.9840.9240.9110.9000.9240.9630.201
10100.0157.1330.0280.9840.9250.9110.9000.9250.9630.198
10100.182.3390.0340.9840.9240.9100.8990.9240.9620.201
1010188.7140.0360.9810.9130.8840.8640.9130.9590.232
10101077.5270.0310.5000.7260.6110.5270.7260.2740.790
AS11000.0001114.7360.0280.9860.9290.9150.9040.9290.9640.192
11000.001112.8510.0400.9860.9270.9130.9020.9270.9640.193
11000.0182.0350.0290.9820.9130.8850.8650.9130.9610.223
11000.186.1500.0290.9820.9130.8850.8650.9130.9610.223
1100129.6450.0300.9560.9080.8800.8570.9080.9420.289
11001013.0960.0280.9350.8580.8250.7950.8580.7900.451
AS51000.0001110.9550.0910.9810.9140.9030.8930.9140.9640.216
51000.001101.3780.0720.9810.9170.9020.8910.9170.9540.215
51000.01100.5130.0870.9780.9100.8970.8850.9100.9570.235
51000.1151.3350.0800.9830.9200.9080.8970.9200.9660.207
51001222.8840.0870.9780.9120.8840.8630.9120.9540.241
510010158.2030.0840.9350.8660.8350.8060.8660.8200.429
AS101000.0001194.8140.1560.9800.9180.9080.8990.9180.9600.218
101000.001186.8880.1680.9820.9150.9050.8960.9150.9640.213
101000.01189.6640.1700.9790.9170.9000.8890.9170.9610.221
101000.1275.7600.1530.9810.9210.9060.8950.9210.9600.214
101001334.0310.1450.9760.9100.8820.8620.9100.9560.244
1010010434.3420.1580.5000.7260.6110.5270.7260.2740.791
AS110000.0001339.5000.1410.9860.9290.9170.9060.9290.9640.191
110000.001326.4980.1580.9860.9300.9170.9060.9300.9640.191
110000.01329.7880.1520.9820.9140.8860.8850.9140.9620.221
110000.1325.0970.1630.9820.9140.8860.8850.9140.9620.221
110001122.3860.1580.9560.9080.8790.8570.9080.9410.288
110001062.0330.1530.9350.8610.8290.8000.8610.8020.446
AS510000.00013444.3001.9770.9820.9200.9050.8950.9200.9640.212
510000.0013353.7292.0160.9830.9210.9110.9030.9210.9650.206
510000.015764.1622.0730.9810.9170.9050.8950.9170.9660.220
510000.15371.7651.9680.9810.9170.9040.8930.9170.9610.225
5100015936.6272.0290.9700.9060.8770.8540.9060.9350.265
51000105273.3511.9350.9180.7880.7310.7180.7880.5110.573
AS1010000.000111,822.4984.0640.9710.9160.8950.8870.9160.9550.265
1010000.00110,310.8225.0910.9730.9140.8920.8840.9140.9470.254
1010000.0111,545.2026.1150.9750.8970.8920.8900.8970.9650.255
1010000.111,281.5754.4230.9670.8920.8750.8610.8920.9260.282
101000118,172.0555.7810.9690.8760.8540.8490.8760.9550.311
1010001010,284.0115.2450.5000.7260.6110.5270.7260.2740.792
Note: N—number of hidden layers, n—number of neurons per hidden layer, α—regularization parameter, TT—training time (seconds), VT—validation time (seconds), AUC—area under the receiver operating curve, CA—classification accuracy, F1—F1 score, PREC—precision, REC—recall, SPEC—specificity, LOSS—cross entropy. The best performing model is highlighted in bold.
Table A2. Classification performance of the A modality in the training and validation phase.
Table A2. Classification performance of the A modality in the training and validation phase.
ModalityNnαTTVTAUCCAF1PRECRECSPECLOSS
A1100.0001131.3110.0120.9720.9240.9100.8980.9240.9510.232
1100.001131.9270.0150.9720.9240.9090.8970.9240.9510.232
1100.0184.3660.0140.9710.9150.8930.8860.9150.9470.249
1100.141.5720.0170.9710.9110.8830.8600.9110.9470.274
110133.9680.0140.9400.9110.8820.8610.9110.9490.322
1101020.7100.0160.9410.8200.7760.7500.8200.6300.530
A5100.000135.8990.0150.9690.9270.9150.9040.9270.9530.226
5100.00137.7530.0190.9700.9260.9140.9040.9260.9530.226
5100.0136.0950.0180.9700.9250.9130.9020.9250.9520.228
5100.150.6890.0190.9710.9270.9160.9050.9270.9520.225
510168.3170.0200.9640.9110.8830.8600.9110.9470.272
5101053.4250.0190.8790.8550.8200.7900.8550.7370.548
A10100.000142.2650.0280.9720.9260.9120.9000.9260.9500.224
10100.00138.7290.0210.9680.9260.9130.9000.9260.9490.228
10100.0143.5390.0270.9690.9270.9140.9020.9270.9490.226
10100.147.2660.0300.9670.9270.9150.9030.9270.9510.225
1010146.8890.0240.9670.9270.9150.9030.9270.9510.225
101010132.9330.0250.5000.7260.6110.5270.7260.2740.790
A11000.0001177.1140.0330.9720.9280.9140.9020.9280.9520.227
11000.001174.0690.0320.9720.9280.9140.9020.9280.9520.227
11000.01166.4680.0300.9720.9250.9110.8990.9250.9510.231
11000.1103.4470.0330.9690.9110.8830.8600.9110.9460.265
1100168.1590.0320.9390.9110.8820.8610.9110.9480.319
11001017.9310.0280.9430.7960.7440.7210.7960.5640.524
A51000.0001107.9250.0900.9690.9230.9130.9040.9230.9540.235
51000.001117.0540.0840.9690.9260.9140.9030.9260.9520.235
51000.01113.9970.0940.9700.9270.9150.9050.9270.9530.230
51000.1187.2610.0900.9720.9250.9140.9030.9250.9540.232
51001286.8170.0870.9660.9120.8830.8600.9120.9460.269
510010123.1660.0840.9380.8550.8200.7910.8550.7400.477
A101000.0001172.3700.1530.9690.9180.9040.8920.9180.9460.246
101000.001162.7440.1650.9680.9230.9110.8990.9230.9490.241
101000.01190.3240.1590.9700.9190.9020.8900.9190.9440.239
101000.1190.0540.0880.9720.9250.9140.9030.9250.9540.232
101001294.7040.0850.9660.9120.8830.8600.9120.9460.269
1010010132.1070.0990.9380.8550.8200.7910.8550.7400.477
A110000.0001376.8960.1530.9710.9270.9130.9010.9270.9520.227
110000.001377.8935.2420.9710.9260.9130.9010.9260.9520.227
110000.01381.6330.1650.9710.9270.9130.9010.9270.9520.228
110000.1393.5550.1530.9700.9120.8830.8610.9120.9480.253
110001172.6610.1420.9400.9110.8830.8610.9110.9490.319
110001080.5110.1500.9420.8290.7890.7600.8290.6730.523
A510000.00013523.7632.0050.9670.9080.8990.8910.9080.948060.256
510000.0014215.9932.0900.9680.9140.9010.8890.9140.945730.253
510000.014812.0802.0350.9670.9120.9000.8890.9120.946580.249
510000.16389.8622.0440.9670.9140.8960.8830.9140.945180.243
5100017175.1132.0490.9570.9110.8820.8600.9110.943950.280
51000104760.1212.1230.9250.7820.7230.7180.7820.478030.567
A1010000.00018451.1844.2510.9670.9130.8950.8840.9130.9450.260
1010000.0017971.3564.5040.9600.9060.8900.8760.9060.9150.289
1010000.0115,169.5794.9090.9600.9080.8950.8830.9080.9400.281
1010000.110,456.1994.9930.9600.8920.8850.8800.8920.9410.297
101000115,724.2934.4410.9610.9120.8830.8600.9120.9430.277
1010001010,937.5704.4790.5000.7260.6110.5270.7260.2740.794
Note: N—number of hidden layers, n—number of neurons per hidden layer, α—regularization parameter, TT—training time (seconds), VT—validation time (seconds), AUC—area under the receiver operating curve, CA—classification accuracy, F1—F1 score, PREC—precision, REC—recall, SPEC—specificity, LOSS—cross entropy. The best performing models are highlighted in bold and the model with the lowest computational requirements was retained for testing.
Table A3. Classification performance of the S modality in the training and validation phase.
Table A3. Classification performance of the S modality in the training and validation phase.
ModalityNnαTTVTAUCCAF1PRECRECSPECLOSS
S1100.000122.7300.0140.9000.8470.8180.7910.8470.8000.454
1100.00122.9280.0150.8990.8470.8180.7910.8470.8000.454
1100.0122.3760.0140.8990.8470.8180.7910.8470.8010.454
1100.124.3380.0120.9000.8460.8160.7880.8460.7930.455
110129.5790.0160.8840.8440.8120.7830.8440.7690.473
1101025.1960.0150.8700.8160.7750.7470.8160.6410.565
S5100.000131.9680.0210.8990.8480.8200.7940.8480.8140.450
5100.00137.1130.0180.8970.8480.8200.7940.8480.8150.451
5100.0134.4990.0230.8980.8480.8200.7940.8480.8160.451
5100.143.3780.0180.8980.8480.8200.7940.8480.8160.451
510161.5930.0200.8840.8480.8190.7920.8480.8050.462
5101059.8560.0230.5000.7260.6110.5270.7260.2740.790
S10100.000134.2800.0260.8990.8480.8200.7940.8480.8140.451
10100.00133.0860.0240.8980.8480.8190.7930.8480.8100.451
10100.0137.2790.0220.8980.8470.8200.7940.8470.8140.452
10100.162.7760.0300.8980.8480.8200.7950.8480.8160.451
10101113.6240.0300.8770.8480.8190.7920.8480.8040.462
101010123.9100.0230.5000.7260.6110.5270.7260.2740.790
S11000.000128.5830.0280.9000.8470.8200.7940.8470.8160.451
11000.00127.8980.0270.8990.8470.8200.7940.8470.8160.451
11000.0125.9510.0280.9000.8470.8200.7940.8470.8140.451
11000.131.8910.0260.8990.8480.8190.7930.8480.8060.452
1100149.7290.0310.8890.8440.8120.7830.8440.7710.470
11001025.9090.0250.8700.8160.7750.7470.8160.6410.561
S51000.000181.1340.0880.8980.8470.8200.7950.8470.8180.452
51000.00190.4760.0890.8980.8470.8190.7940.8470.8150.452
51000.01132.2770.0900.8990.8470.8190.7920.8470.8070.451
51000.1222.8790.0880.8980.8470.8180.7920.8470.8080.454
51001212.6960.0820.8960.8480.8190.7920.8480.8060.457
510010133.8320.0810.8630.7960.7480.7250.7960.5630.614
S101000.0001199.1420.1650.8990.8470.8180.7920.8470.8070.453
101000.001208.4650.1580.8990.8470.8190.7930.8470.8100.452
101000.01328.1090.1710.8960.8470.8200.7950.8470.8170.455
101000.1498.4910.1520.8960.8460.8180.7930.8460.8150.458
101001413.1170.1760.8880.8470.8190.7930.8470.8110.462
1010010604.8660.1660.5000.7260.6110.5270.7260.2740.791
S110000.0001123.8320.1680.8980.8480.8200.7930.8480.8100.453
110000.001117.6160.1440.8980.8480.8200.7930.8480.8100.453
110000.01122.7340.1620.8980.8480.8200.7930.8480.8100.453
110000.1135.3290.1550.8990.8480.8190.7920.8480.8050.453
110001125.7810.1590.8890.8440.8130.7850.8440.7760.471
1100010141.9330.1410.8710.8190.7790.7500.8190.6550.556
S510000.00013967.5782.0700.8970.8470.8190.7930.8470.8140.456
510000.0016176.5442.0290.8990.8480.8200.7940.8480.8140.452
510000.016545.2422.0770.9000.8480.8190.7930.8480.8100.452
510000.15265.8862.0640.8930.8470.8200.7950.8470.8180.461
5100018161.3032.1860.8860.8480.8190.7920.8480.8000.463
51000104835.6672.2740.8110.7590.6860.6880.7590.4090.681
S1010000.00019320.8164.9930.8970.8460.8180.7920.8460.8060.458
1010000.00117,128.9454.4420.8970.8480.8200.7940.8480.8140.455
1010000.0114,130.6913.8500.8940.8470.8190.7930.8470.8130.456
1010000.113,607.8064.4610.8980.8470.8190.7930.8470.8120.454
101000112,868.9693.8380.8840.8440.8170.7930.8440.8170.488
1010001011,687.6634.8180.5000.7260.6110.5270.7260.2740.794
Note: N—number of hidden layers, n—number of neurons per hidden layer, α—regularization parameter, TT—training time (seconds), VT—validation time (seconds), AUC—area under the receiver operating curve, CA—classification accuracy, F1—F1 score, PREC—precision, REC—recall, SPEC—specificity, LOSS—cross entropy. The best performing model is highlighted in bold.
Figure A1. Classification accuracy (CA) by the modality used (AS, A, S) and the architecture of the neural network. Legend: green stands for a single hidden layer, yellow stands for 5 hidden layers, red stands for 10 hidden layers, circles stand for 10 neurons, crosses stand for 100 neurons and triangles stand for 1000 neurons.
Figure A1. Classification accuracy (CA) by the modality used (AS, A, S) and the architecture of the neural network. Legend: green stands for a single hidden layer, yellow stands for 5 hidden layers, red stands for 10 hidden layers, circles stand for 10 neurons, crosses stand for 100 neurons and triangles stand for 1000 neurons.
Forests 15 02019 g0a1
Figure A2. Training time (seconds) by the modality used (AS, A, S) and the architecture of the neural network. Legend: green stands for a single hidden layer, yellow stands for 5 hidden layers, red stands for 10 hidden layers, circles stand for 10 neurons, crosses stand for 100 neurons and triangles stand for 1000 neurons.
Figure A2. Training time (seconds) by the modality used (AS, A, S) and the architecture of the neural network. Legend: green stands for a single hidden layer, yellow stands for 5 hidden layers, red stands for 10 hidden layers, circles stand for 10 neurons, crosses stand for 100 neurons and triangles stand for 1000 neurons.
Forests 15 02019 g0a2

References

  1. Mologni, O.; Dyson, P.; Amishev, D.; Proto, A.R.; Zimbalatti, G.; Cavalli, R.; Grigolato, S. Tensile Force Monitoring on Large Winch-Assist Forwarders Operating in British Columbia. Croat. J. For. Eng. 2018, 39, 193–204. [Google Scholar]
  2. Stoilov, S.; Proto, A.R.; Angelov, G.; Papandrea, S.F.; Borz, S.A. Evaluation of salvage logging productivity and costs in the sensitive forests of Bulgaria. Forests 2021, 12, 309. [Google Scholar] [CrossRef]
  3. Casado, M.; Acuña, L.; Basterra, L.A.; Ramón-Cueto, G.; Vecilla, D. Grading of structural timber of Populus×euramericana clone I-214. Holzforschung 2012, 66, 633–638. [Google Scholar] [CrossRef]
  4. Papandrea, S.F.; Cataldo, M.F.; Bernardi, B.; Zimbalatti, G.; Proto, A.R. The Predictive Accuracy of Modulus of Elasticity (MOE) in the Wood of Standing Trees and Logs. Forests 2022, 13, 1273. [Google Scholar] [CrossRef]
  5. Industrial Applications of Poplar Wood. Available online: https://propopulus.eu/en/industrial-applications-of-poplar-wood/ (accessed on 1 October 2024).
  6. Dickmann, D.I. Silviculture and biology of short-rotation woody crops in temperate regions: Then and now. Biomass Bioenerg. 2006, 30, 696–705. [Google Scholar] [CrossRef]
  7. Marogel-Popa, T.; Cheţa, M.; Marcu, M.V.; Duţă, I.C.; Ioraş, F.; Borz, S.A. Manual cultivation operations in poplar stands: A characterization of job difficulty and risks of health impairment. Int. J. Environ. Res. Public Health 2019, 16, 1911. [Google Scholar] [CrossRef]
  8. Zhai, S.; Shi, Y.; Zhou, J.; Liu, J.; Huang, D.; Zou, A.; Jiang, P. Simulation Optimization and Experimental Study of the Working Performance of a Vertical Rotary Tiller Based on the Discrete Element Method. Actuators 2022, 11, 342. [Google Scholar] [CrossRef]
  9. Keefe, R.F.; Zimbelman, E.G.; Wempe, A.M. Use of smartphone sensors to quantify the productive cycle elements of hand fallers on industrial cable logging operations. Int. J. For. Eng. 2019, 30, 132–143. [Google Scholar] [CrossRef]
  10. Becker, R.M.; Keefe, R.F. A novel smartphone-based activity recognition modelling method for tracked equipment in forest operations. PLoS ONE 2022, 17, e0266568. [Google Scholar] [CrossRef]
  11. Wang, G.; Zhang, W.; Chen, M.; Ji, M.; Diao, X.; Miao, H. Enhancing Sustainable Afforestation through Innovative Earth Auger Design: A Simulation Study in Hilly Regions. Sustainability 2024, 16, 5402. [Google Scholar] [CrossRef]
  12. Tampekis, S.; Kantartzis, A.; Arabatzis, G.; Sakellariou, S.; Kolkos, G.; Malesios, C. Conceptualizing Forest Operations Planning and Management Using Principles of Functional Complex Systems Science to Increase the Forest’s Ability to Withstand Climate Change. Land 2024, 13, 217. [Google Scholar] [CrossRef]
  13. Hiesl, P.; Benjamin, J.G. Applicability of international harvesting equipment productivity studies in Maine, U.S.A.: A literature review. Forests 2013, 4, 898–921. [Google Scholar] [CrossRef]
  14. Demsar, J.; Curk, T.; Erjavec, A.; Gorup, C.; Hocevar, T.; Milutinovic, M.; Mozina, M.; Polajnar, M.; Toplak, M.; Staric, A.; et al. Orange: Data Mining Toolbox in Python. J. Mach. Learn. Res. 2013, 14, 2349–2353. [Google Scholar]
  15. Building A Neural Net from Scratch Using R—Part 1. Available online: https://rviews.rstudio.com/2020/07/20/shallow-neural-net-from-scratch-using-r-part-1/ (accessed on 23 July 2024).
  16. Python AI: How to Build a Neural Network & Make Predictions. Available online: https://realpython.com/python-ai-neural-network/ (accessed on 23 July 2024).
  17. Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep learning for sensor-based human activity recognition: Overview, challenges and opportunities. J. ACM 2021, 77, 1–40. [Google Scholar] [CrossRef]
  18. Acuna, M.; Bigot, M.; Guerra, S.; Hartsough, B.; Kanzian, C.; Kärhä, K.; Lindroos, O.; Magagnotti, N.; Roux, S.; Spinelli, R.; et al. Good Practice Guidelines for Biomass Production Studies; CNR IVALSA Sesto Fiorentino (National Research Council of Italy—Trees and Timber Institute): Sesto Fiorentino, Italy, 2012; 52p. [Google Scholar]
  19. Björheden, R.; Apel, K.; Shiba, M.; Thompson, M. IUFRO Forest Work Study Nomenclature; Swedish University of Agricultural Science, Department of Operational Efficiency: Grapenberg, Sweden, 1995; 16p. [Google Scholar]
  20. Ruiz-Garcia, L.; Sanchez-Guerrero, P. A Decision Support Tool for Buying Farm Tractors, Based on Predictive Analytics. Agriculture 2022, 12, 331. [Google Scholar] [CrossRef]
  21. Lanças, K.P.; Marques Filho, A.C.; Santana, L.S.; Ferraz, G.A.e.S.; Faria, R.O.; Martins, M.B. Agricultural Tractor Test: A Bibliometric Review. AgriEngineering 2024, 6, 2229–2248. [Google Scholar] [CrossRef]
  22. Puška, A.; Nedeljković, M.; Šarkoćević, Ž.; Golubović, Z.; Ristić, V.; Stojanović, I. Evaluation of Agricultural Machinery Using Multi-Criteria Analysis Methods. Sustainability 2022, 14, 8675. [Google Scholar] [CrossRef]
  23. Bacescu, N.M.; Cadei, A.; Moskalik, T.; Wisniewski, M.; Talbot, B.; Grigolato, S. Efficiency assessment of fully mechanized harvesting system through the use of fleet management system. Sustainability 2022, 14, 16751. [Google Scholar] [CrossRef]
  24. Spinelli, R.; Laina-Relano, R.; Magagnotti, N.; Tolosana, E. Determining observer and method effects on the accuracy of elemental time studies in forest operations. Balt. For. 2013, 19, 301–306. [Google Scholar]
  25. Musat, E.C.; Apafaian, A.I.; Ignea, G.; Ciobanu, V.D.; Iordache, E.; Derczeni, R.A.; Spârchez, G.; Vasilescu, M.M.; Borz, S.A. Time expenditure in computer aided time studies implemented for highly mechanized forest equipment. Ann. For. Res. 2016, 59, 129–143. [Google Scholar]
  26. Borz, S.A. Development of a modality-invariant multi-layer perceptron to predict operational events in motor-manual willow felling operations. Forests 2021, 12, 406. [Google Scholar] [CrossRef]
  27. Borz, S.A.; Talagai, N.; Cheţa, M.; Chiriloiu, D.; Gavilanes Montoya, A.V.; Castillo Vizuete, D.D.; Marcu, M.V. Physical strain, exposure to noise and postural assessment in motor-manual felling of willow short rotation coppice: Results of a preliminary study. Croat. J. For. Eng. 2019, 40, 377–388. [Google Scholar] [CrossRef]
  28. Borz, S.A.; Cheta, M.; Birda, M.; Proto, A.R. Classifying operational events in cable yarding by a machine learning application to GNSS-collected data: A case study on gravity-assisted downhill yarding. Bull. Transilv. Univ. Bras. 2022, 15, 13–32. [Google Scholar] [CrossRef]
  29. Şofletea, N.; Curtu, L. Dendrologie, 2nd ed.; Pentru Viaţă Publishing House: Braşov, Romania, 2008; ISBN 978-973-85874-4-1. [Google Scholar]
  30. Abrudan, I.V. Împăduriri; Transilvania University Press: Braşov, Romania, 2006. [Google Scholar]
  31. Florescu, I.; Nicolescu, N.V. Silviculture; Lux Libris Publishing House: Brasov, Romania, 1996. [Google Scholar]
  32. Castro Pérez, S.N.; Borz, S.A. Improving the event-based classification accuracy in pit-drilling operations: An application by neural networks and median filtering of the acceleration input signal data. Sensors 2021, 21, 6288. [Google Scholar] [CrossRef]
  33. Ministry of Waters, Forests and Environment Protection; National Forest Administration. Unified Work Time and Production Rates for Forestry, Ministry of Waters; Forests and Environment Protection: Bucharest, Romania, 1997. [Google Scholar]
  34. Spinelli, R.; Visser, R.J.M. Analyzing and estimating delays in wood chipping operations. Biomass Bioenergy 2009, 33, 429–433. [Google Scholar] [CrossRef]
  35. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on Machine Learning (ICML 2013), Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  36. Nair, V.; Hinton, G.E. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML 2010), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  37. Kingma, D.P.; Ba, J.L. 2015: ADAM: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  38. Fawcett, T. An introduction to ROC analysis. Pattern Recogn. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  39. Kamilaris, A.; Prenafeta-Boldu, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef]
  40. Cross Validation in Machine Learning. Available online: https://www.geeksforgeeks.org/cross-validation-machine-learning/ (accessed on 23 July 2024).
  41. 5 Reasons Why You Should Use Cross-Validation in Your Data Science Projects. Available online: https://towardsdatascience.com/5-reasons-why-you-should-use-cross-validation-in-your-data-science-project-8163311a1e79 (accessed on 23 July 2024).
  42. Bulling, A.; Blanke, U.; Schiele, B. A tutorial of human activity recognition using body-worn inertial sensors. ACM Comput. Surv. 2014, 46, 1–33. [Google Scholar] [CrossRef]
  43. Borz, S.A.; Forkuo, G.O.; Oprea-Sorescu, O.; Proto, A.R. Development of a robust machine learning model to monitor the operational performance of fixed-post multi-blade vertical sawing machines. Forests 2022, 13, 1115. [Google Scholar] [CrossRef]
  44. Marogel-Popa, T. Performance of Planting and Cultivation Operations in Poplar Forests Located in the Danube Meadow. Ph.D. Thesis, Transilvania University of Brasov, 2020. Available online: https://www.unitbv.ro/documente/cercetare/doctorat-postdoctorat/sustinere-teza/2020/marogel-tiberiu/REZUMAT_TEZA_DOCTORAT_MAROGEL-POPA_TIBERIU.pdf (accessed on 6 August 2024).
  45. Neural Network. Available online: https://orangedatamining.com/widget-catalog/model/neuralnetwork/ (accessed on 6 August 2024).
  46. Normalization vs Standardization. Available online: https://www.geeksforgeeks.org/normalization-vs-standardization/ (accessed on 6 August 2024).
  47. Talagai, N.; Marcu, M.V.; Zimbalatti, G.; Proto, A.R.; Borz, S.A. Productivity in partly mechanized planting operations of willow short rotation coppice. Biomass Bioenergy 2020, 138, 105609. [Google Scholar] [CrossRef]
  48. Manzone, M.; Balsari, P. Planters performance during a very short rotation coppice planting. Biomass Bioenergy 2014, 67, 188–192. [Google Scholar] [CrossRef]
  49. Manzone, M.; Bergante, S.; Facciotto, G.; Balsari, P. A prototype for horizontal long cuttings planting in short rotation coppice. Biomass Bioenergy 2017, 107, 214–218. [Google Scholar] [CrossRef]
  50. Bush, C.; Volk, T.A.; Eisenbies, M.H. Planting rates and delays during the establishment of willow biomass crops. Biomass Bioenergy 2015, 83, 290–296. [Google Scholar] [CrossRef]
  51. Borz, S.A.; Nita, M.D.; Talagai, N.; Scriba, C.; Grigolato, S.; Proto, A.R. Performance of small-scale technology in planting and cutback operations of short-rotation willow crops. Trans. ASABE 2019, 62, 167–176. [Google Scholar] [CrossRef]
Figure 1. Study location: (a) the plot taken into study showing the GPS locations (yellow) collected during the field observation; (b) location of the study area at national level (yellow dot). Note: the maps were created in open source QGis software (QGis 3.4.12) by using freely available vector datasets of the country and the open access Bing ® aerial data as accessible at the study time.
Figure 1. Study location: (a) the plot taken into study showing the GPS locations (yellow) collected during the field observation; (b) location of the study area at national level (yellow dot). Note: the maps were created in open source QGis software (QGis 3.4.12) by using freely available vector datasets of the country and the open access Bing ® aerial data as accessible at the study time.
Forests 15 02019 g001
Figure 2. Configuration of the equipment taken into study: (a) the farm tractor equipped with the rotary tiller showing the location of the handheld GNSS receiver (1); (b) the rotary tiller observed in the study, (c) placement of the triaxial accelerometer (2) on the rotary tiller.
Figure 2. Configuration of the equipment taken into study: (a) the farm tractor equipped with the rotary tiller showing the location of the handheld GNSS receiver (1); (b) the rotary tiller observed in the study, (c) placement of the triaxial accelerometer (2) on the rotary tiller.
Forests 15 02019 g002
Figure 3. The workflow used to train, validate the models, and to save the best performing model in Orange Visual Programming. Legend: 1—the “Data” widget enables loading data and setting the feature and target variables, 2—the “Neural Network” widget enables the configuration of the network architecture and hyperparameter tuning each time a model is trained and validated, 3—the “Test and Score” widget enables the selection of the training and validation options as well as the extraction of classification performance metrics once the process is complete, and 4—the “Save Model” widget enables model saving. Note: the software enables the communication of various kinds of information between the widgets by connecting them.
Figure 3. The workflow used to train, validate the models, and to save the best performing model in Orange Visual Programming. Legend: 1—the “Data” widget enables loading data and setting the feature and target variables, 2—the “Neural Network” widget enables the configuration of the network architecture and hyperparameter tuning each time a model is trained and validated, 3—the “Test and Score” widget enables the selection of the training and validation options as well as the extraction of classification performance metrics once the process is complete, and 4—the “Save Model” widget enables model saving. Note: the software enables the communication of various kinds of information between the widgets by connecting them.
Forests 15 02019 g003
Figure 4. The workflow used to make predictions on unseen data by the saved models. Legend: 1—a “Data” widget used to load data and set the feature and target variables for unseen data, 2—the “Load Model” widget used to successively load the best performing models saved for AS, A and S modalities, 3—the “Predictions” widget used to make predictions on classification performance for unseen data, and 4—the “Confusion Matrix” widget used to show misclassifications. Note: the software enables the communication of various kinds of information between the widgets by connecting them.
Figure 4. The workflow used to make predictions on unseen data by the saved models. Legend: 1—a “Data” widget used to load data and set the feature and target variables for unseen data, 2—the “Load Model” widget used to successively load the best performing models saved for AS, A and S modalities, 3—the “Predictions” widget used to make predictions on classification performance for unseen data, and 4—the “Confusion Matrix” widget used to show misclassifications. Note: the software enables the communication of various kinds of information between the widgets by connecting them.
Forests 15 02019 g004
Figure 5. Confusion matrices of predictions made on unseen data: (a) predictions made by the best model for AS modality; (b) predictions made by the best model for A modality; (c) predictions made by the best model for S modality.
Figure 5. Confusion matrices of predictions made on unseen data: (a) predictions made by the best model for AS modality; (b) predictions made by the best model for A modality; (c) predictions made by the best model for S modality.
Forests 15 02019 g005
Table 1. Events observed and documented based on the analysis of video data.
Table 1. Events observed and documented based on the analysis of video data.
EventAbbreviation Used for AnnotationDescription
Active workWORKThe tractor moves and powers the tiller which is used to remove the weed and till the soil. Conventionally, tiller disengagement was included in this category.
ManeuveringMANIncludes the exit and entering maneuvers, maneuvers taken in the plot, and very short stops (couple of seconds) in the plot. Tiller was disengaged during these events.
No movement, engine onSTOPThe tractor is stopped with the engine running for longer periods of time.
No movement, engine offOFFThe tractor is stopped with the engine turned off, typically at the headland and at the beginning and end of the working day.
Table 2. Summary of the best models in terms of classification performance.
Table 2. Summary of the best models in terms of classification performance.
Modality and
Dataset
Optimal Architecture and
Hyperparameters
Event ClassAUCCAF1PRECRECSPECLOSS
AS, DS1N = 1, n = 1000, α = 0.001, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.9860.9300.9170.9060.9300.9640.191
WORK0.9860.9300.9170.9060.9300.9640.191
MAN0.9980.9820.9870.9870.9870.9660.048
STOP0.9840.9420.8680.8160.9270.9460.131
OFF0.9390.9700.0000.0000.0001.0000.084
A, DS1N = 1, n = 100, α = 0.001, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.9720.9280.9140.9020.9280.9520.227
WORK0.9850.9780.9850.9810.9890.9500.080
MAN0.9720.9420.8670.8190.9210.9480.154
STOP0.9240.9700.0000.0000.0001.0000.088
OFF0.9480.9650.5580.5660.5490.9830.086
S, DS1N = 5, n = 1000, α = 1, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.8860.8480.8190.7920.8480.8000.463
WORK0.9180.9110.9410.9130.9710.7550.268
MAN0.8530.8550.6650.6300.7030.8940.343
STOP0.8850.9700.0000.0000.0001.0000.099
OFF0.8780.9600.0000.0000.0001.0000.120
Note: N—number of hidden layers, n—number of neurons per hidden layer, α—regularization parameter, AUC—area under the receiver operating curve, CA—classification accuracy, F1—F1 score, PREC—precision, REC—recall, SPEC—specificity, LOSS—cross entropy.
Table 3. Summary of the models’ performance on unseen data.
Table 3. Summary of the models’ performance on unseen data.
Modality and
Dataset
Optimal Architecture and
Hyperparameters
Event ClassAUCCAF1PRECRECSPECLOSS
AS, DS2N = 1, n = 1000, α = 0.001, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.9840.9060.9220.9460.9060.9850.187
WORK0.9980.9060.9880.9980.9770.9950.187
MAN0.9760.9060.8140.8820.7550.9620.187
OFF0.9270.9060.2500.1650.5180.9330.187
A, DS2N = 1, n = 100, α = 0.001, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.9810.9220.9310.9430.9220.9760.188
WORK0.9950.9220.9840.9930.9760.9830.188
MAN0.9740.9220.8520.8810.8250.9580.188
OFF0.9210.9220.2830.2060.4470.9560.188
S, DS2N = 5, n = 1000, α = 1, Activation function = ReLU, solver = ADAM, number of iterations = 1,000,000, stratified cross-validation by 10 folds.OVERALL0.9020.9010.8870.8760.9010.8360.352
WORK0.9160.9010.9480.9170.9810.7870.352
MAN0.8920.9010.8110.8520.7740.9500.352
OFF0.8680.9010.0000.0000.0001.0000.352
Note: N—number of hidden layers, n—number of neurons per hidden layer, α—regularization parameter, AUC—area under the receiver operating curve, CA—classification accuracy, F1—F1 score, PREC—precision, REC—recall, SPEC—specificity, LOSS—cross entropy.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Borz, S.A.; Proto, A.R. Predicting Operational Events in Mechanized Weed Control Operations by Offline Multi-Modal Data and Machine Learning Provides Highly Accurate Classification in Time Domain. Forests 2024, 15, 2019. https://doi.org/10.3390/f15112019

AMA Style

Borz SA, Proto AR. Predicting Operational Events in Mechanized Weed Control Operations by Offline Multi-Modal Data and Machine Learning Provides Highly Accurate Classification in Time Domain. Forests. 2024; 15(11):2019. https://doi.org/10.3390/f15112019

Chicago/Turabian Style

Borz, Stelian Alexandru, and Andrea Rosario Proto. 2024. "Predicting Operational Events in Mechanized Weed Control Operations by Offline Multi-Modal Data and Machine Learning Provides Highly Accurate Classification in Time Domain" Forests 15, no. 11: 2019. https://doi.org/10.3390/f15112019

APA Style

Borz, S. A., & Proto, A. R. (2024). Predicting Operational Events in Mechanized Weed Control Operations by Offline Multi-Modal Data and Machine Learning Provides Highly Accurate Classification in Time Domain. Forests, 15(11), 2019. https://doi.org/10.3390/f15112019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop