Next Article in Journal
Identification and Classification of Aluminum Scrap Grades Based on the Resnet18 Model
Next Article in Special Issue
Dense Semantic Forecasting with Multi-Level Feature Warping
Previous Article in Journal
Real-Time Risk Assessment for Road Transportation of Hazardous Materials Based on GRU-DNN with Multimodal Feature Embedding
Previous Article in Special Issue
A Survey on Depth Ambiguity of 3D Human Pose Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Detection and Classification of Returnable Packaging Based on YOLOV4 Algorithm

1
University of Rijeka, Trg Braće Mažuranića 10, 51000 Rijeka, Croatia
2
Faculty of Engineering, University of Rijeka, Vukovarska 58, 51000 Rijeka, Croatia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(21), 11131; https://doi.org/10.3390/app122111131
Submission received: 19 September 2022 / Revised: 27 October 2022 / Accepted: 28 October 2022 / Published: 2 November 2022

Abstract

:
This article describes the implementation of the You Only Look Once (YOLO) detection algorithm for the detection of returnable packaging. The method of creating an original dataset and creating an augmented dataset is shown. The model was evaluated using mean Average Precision (mAP), F1 score , Precision, Recall, Average Intersection over Union (Average IoU) score, and Average Loss. The training was conducted in four cycles, i.e., 6000, 8000, 10,000, and 20,000 max batches with three different activation functions Mish, ReLU, and Linear (used in 6000 and 8000 max batches). The influence train/test dataset ratio was also investigated. The conducted investigation showed that variation of hyperparameters (activation function and max batch sizes) have a significant influence on detection and classification accuracy with the best results obtained in the case of YOLO version 4 (YOLOV4) with the Mish activation function and max batch size of 20,000 that achieved the highest mAP of 99.96% and lowest average error of 0.3643.

1. Introduction

Today, modern society produces a growing amount of waste which has a huge impact on the environment. Some of the disposed waste materials can last in the environment for long periods up to hundreds to even thousands of years (for example plastics) which can cause significant impacts on animals and plants. Waste management is the most important factor for reducing and preventing the possible impact of hazardous waste in nature. The main reason is that every waste management system consists of actions such as collection, transport, treatment, and disposal of waste, which results in a reduction of the waste materials disposed of in nature [1]. Additionally, to reduce disposed waste in nature, along with the waste management system, it is necessary to implement advanced computational methods that, in combination with artificial intelligence (AI), significantly contribute to the detection and classification of harmful waste.
Object detection has had a great impact in the field of computer vision tasks such as image annotation [2], vehicle counting [3], activity recognition [4], face detection [5], etc., in the past decade. It also has had an essential role in scene understanding, which gained a high level of popularity in security, transportation, medical and military applications. In addition to the above examples, it also has multiple applications: autonomous driving [6], people detection in security [7], vehicle detection with AI in transportation [8], medical feature detection in healthcare, and in waste management [9].
Over the last two decades, AI methods have gained momentum in offering alternative computational approaches to solve solid waste management problems. In the case of the detection and classification of recyclable waste, convolutional neural networks (CNN) were mostly utilized.
Mao et al. [10] used the genetic algorithm (GA) for optimization of a fully connected layer, DenseNet121 to improve the classification accuracy of the result ThrashNet [11] dataset which consists of 2527 images in six categories of waste (glass, paper, cardboard, plastic, metal, and trash). From the obtained results, it was seen that the optimized DenseNet121 achieved the highest accuracy of 99.6%. Adedeji and Wang [12] proposed the intelligent waste material classification, which is based on ResNet-50 and support vector machines (SVM) applied to TrashNet dataset [11]. ResNet-50 was used as a feature extractor while the SVM was used to classify the waste into the specified groups, the obtained model achieved an accuracy score of 87%. Bobulski and Kubanek [13] performed the plastic waste classification using CNN on the following classes: polyethylene terephthalate, high-density polyethylene, polypropylene, and polystyrene. Authors in the research created a unique dataset, they also investigated the different configurations of CNN (15 and 23-layer network), diverse image resolutions (120 × 120, and 227 × 227 pixels), and the influence of different train/test dataset ratios (60:40 up to 90:10). The highest classification accuracy was achieved in case of 23-layer CNN with the image resolution of 227 × 227 pixels of 99.23% at train/test ratio 90:10. Shi et al. [14] proposed the multi-layer hybrid CNN (similar to VggNet with fewer parameters and higher classification accuracy) and tested on TrashNet [11] dataset, for intelligent waste classification to overcome the low classification accuracy and long running times of the majority of CNN-s. In this investigation, the highest classification accuracy achieved was 92.6%. The SVM along with Histogram of Oriented Gradient (HOG) features CNN, and ResNet50 CNN [15] were applied to publicly available waste dataset from Kaggle [16]. The goal was to identify a single object in images and classify it into one of the given categories (cardboard, glass, metal, paper, plastic, and trash). The highest accuracy score was 95.35% and it was achieved using ResNet50 architecture. On the original dataset Altikat et al. [17] applied the four and five-layer deep CNN for classifying the paper, glass, plastic, and organic waste. The investigation showed that the highest classification accuracy of 83% and 76.7% was achieved in the case of organic waste with four and five-layered deep CNN, respectively. In the last few years, the YOLO algorithm has been increasingly used in the detection and classification of non-biodegradable waste. Aishwarya et al. [18] applied the YOLO algorithm to detect non-biodegradable waste from the bins. Classification accuracy was faulty and imprecise with the results of 75% for metal, 65% for plastic, and 60% for glass. Wahyutama and Hwang [19] reported the development of a smart trash bin that separates and collects trash recyclables using a webcam and YOLO algorithm where the classification accuracy of recyclables using YOLO was 91% in an optimal computing environment and 75% when implemented on Raspberry Pi. Lin [20] used the YOLO-Green object detection algorithm for detecting trash in real time. In this paper, the original dataset was developed from real-world trash divided into seven types of solid waste (batteries, clothes, electronic waste, glass, metal, paper, and plastic). The YOLO-Green achieved a classification accuracy of 78.04%. The YOLO-based neural network model with variational autoencoder was employed for automatic detection and classification of recycling waste [21]. The highest classification accuracy was 69.7%.
The superiority of the YOLO algorithm compared to other detection algorithms can be demonstrated by additional research and applications in practice. For example, Kim et al. [22] trained YOLOV4, Single Shot MultiBox Detector (SSD), and Faster-Region Based Convolutional Neural Networks (Faster-RCNN) for the detection and classification of vehicle models. The obtained results showed that the YOLOV4 algorithm has had the best performance (mAP in the amount of 98.19% and FPS in the amount of 82.1). SSD proved to be the fastest (FPS in the amount of 105.14), but the problem comes from the simplified light model (Mobilnet- lightweight model) which frequently failed to detect a vehicle. Faster-RCNN turned out to be the slowest model (FPS in the amount of 36.32), although it was slightly less accurate (mAP valued at 93.40%) than YOLOV4, due to its detection speed it was unable to perform the given task. According to [21], YOLO can be applied as an algorithm for an automatic waste recycling system. The authors trained the modified YOLO (YOLO with variational Autoencoder) and Fast R-CNN model on a custom-made dataset consisting of cans, batteries, and plastic bottles. The obtained results prove that the modified YOLO model with a prediction accuracy rate of 69.70% and a localization rate of 22.10% again outperforms the competitive Fast R-CNN with an accuracy rate of 71.60% and a localization rate of 8.60%. In the paper, ref. [23] AI-controlled Outdoor Autonomous Trash-Collecting Robot was used for underwater garbage collection. Several object detection algorithms (Mask-RCNN, YOLOV4, and YOLOV4-tiny) were compared and tested for the outdoor waste detection and classification task. The obtained results were evaluated with the best map and the detection time where YOLOV4 and YOLOV4-tiny outperformed the Mask-RCNN in the detection and localization of outdoor trash. YOLOV4 and YOLOV4-Tiny achieved the best mAP of 99.32% and 95.25% with a detection speed of 32.76 and 5.21 ms, while the Mask-RCNN barely achieved the best mAP of 89.10% and had a very slow detection time in the amount of 3973.29 ms. One of the important factors when choosing the optimal algorithm for implementation in an automatic system is the memory of the model itself. Tian et al. [24] trained Faster R-CNN, 4S-YOLOv4, YOLOv4, YOLOv3, SSD, and 4SP-YOLOv4, on a dataset consisting of 6600 images of underwater garbage. The obtained results show that 4SP YOLO requires significantly less memory for obtaining weight factors than Faster R-CNN or SSD. In this particular case, 4SP YOLO requires 10 2 less memory than SSD and up to 10 6 times less memory than Faster R-CNN. This fact gives the possibility of implementing the YOLO detection algorithm to a printed circuit board-based microcontrollers such as Raspberry Pi, CaffeLatte, etc.
The goal of this paper is to examine the YOLOV4 algorithm and achieve the optimal performance by varying activation function and max batch size hyperparameters on a custom-made, three-class (plastic, glass, and aluminum) dataset that contains 3788 images (2838 original images and 950 downloaded from Kaggle [25]). Based on the idea of this investigation and the extensive literature overview, the following hypotheses questions arise:
  • Is it possible to achieve high classification accuracy with the YOLOV4 algorithm?
  • Is it possible to improve classification accuracy by tweaking the specific hyperparameters (max batch size and activation function) of the YOLOV4 algorithm?
  • Does the train/test dataset ratio has any influence on YOLOV4 classification accuracy?
The structure of this paper can be divided into the following sections, i.e., Materials and Methods, Results and Discussion, and Conclusions. In Materials and Methods, the dataset description is given as well as the YOLOV4 algorithm and evaluation methods. In Results and Discussion, the results of the trained YOLOV4 algorithm with different hyperparameters are presented as well as discussed. The conclusions section contains a list of conclusions obtained during this investigation that is based on the results and discussion section and provides direct answers to the hypotheses defined in the Introduction section.

2. Materials and Methods

In this section, the procedure of dataset development is described, as well as the utilized YOLOV4 algorithm. The first step for training the YOLOV4 algorithm was to create a diverse and equally distributed dataset. With this fact, the description of creating a dataset can be summarized in the following steps:
  • gathering of original images of returnable packaging,
  • combination of original images to those obtained from Kaggle,
  • performing the dataset augmentation to enlarge the number of images to get as many combinations as possible combinations for training, testing, and validation set,
  • labeling the bounding boxes for determination of objects inside of selected image in YOLOV4 compatible format, and
  • testing the video camera detection in real-time to check the operation of the designed detection algorithm.
The procedure of collecting original images of returnable packaging such as plastic, glass, and aluminum bottles/cans is described in the following subsection.

2.1. Dataset Description and Preparation

The data collection strategy over the life of the service/product is essential for the development of a successful project [26] and it cannot be a series of one-off exercises. Every time a user comes in contact with a desired product or service, it would be desirable to collect data from the interaction [27] to gain a constant flow of data to improve products/programs. Upon reaching that level of data collection, each new customer contributes to the size and diversity of the newly created dataset while the advantage of building this strategy and collecting data is that it becomes difficult for competitors to practically replicate the dataset used in these projects [28]. With larger and more up-to-date data the AI becomes better and much more valuable and because of that the first step was to collect, an original unprocessed packaging is shown in Figure 1. As already stated the part of the dataset was photographed in its natural form, without any processing. These images were created in three steps: Initially, the packaging was photographed without modifications to its original forms. After that, thermal and mechanical processing was applied to obtain different images. Finally, multiple packaging objects, in both processed and unprocessed form were grouped and photographed. It should be noted that the packaging was empty, i.e., did not contain any liquid. The packaging was photographed in several positions: upright, lying down, rotated, or shifted concerning the original position, which contributed to possible conditions during the detection.
To expand the dataset, the mechanical and thermal loading on plastic and aluminum packaging was applied to change their shape. The glass packaging could not undergo such modification since it is a hard and brittle material and would shatter under mechanical loading or would crack at higher temperatures. The application of such external loading on packaging greatly increased the dataset size as well as its quality. Additional improvement of dataset quality was conducted by photographing a group of packaging. The examples of packaging deformed by the application of thermal loading, mechanical loading, and a group of all three are shown in Figure 2.
All the photographs in the dataset are grey-scale because they have small memory size images that have fewer characteristics that can affect the training of YOLOV4 CNN [29,30]. When creating the original dataset, it is important to pay attention that it should be as diverse as possible [31] so that the ML algorithm has plenty different learning examples as possible [32]. The initial dataset consists of 2838 photographs with an additional 950 which are downloaded from Kaggle [25].
The dataset is not sufficient to develop a good AI model [33] and to enlarge it the data augmentation was performed [34,35]. The methods of geometric augmentation [36] that were used are rotation, blurring, flipping, translating, and zooming (scaling) shown in Figure 3.
The whole dataset collection was performed using OpenCV [37] library in Python programming language. The number of original images, augmented images, and total images in the dataset is presented in Table 1.
In Table 1 the number of original, augmented and total images for each class (plastic, glass, and aluminum) is shown. From the number of original images per class, it can be seen that the dataset was not perfectly balanced [38]. The augmentation was used to compensate for the dataset imbalance and improve its quality [39]. After the original dataset was obtained and augmentation performed the next step was labeling the bounding boxes for the YOLOV4 algorithm using Visual Object Tagging Tool (VoTT) [40]. The marking of masks, i.e., border frames for a single image is shown in Figure 4.
The dataset must be well-prepared [41] for the algorithm to have the best examples available for training, i.e., learning, so the important thing while labeling the dataset was being careful to gather as much vital information as possible and to avoid labeling unnecessary elements [42]. Correct and incorrect labeling of bounding boxes is shown in Figure 5.
In Figure 5, the correct (Figure 5a) and incorrect (Figure 5b) labeling of bounding boxes is shown. Incorrect labeling of bounding boxes (Figure 5b) shows that additional elements such as an empty workspace surface were taken into account during labeling or an unnecessa ry additional element which of course, affects the training of the YOLOV4 algorithm and overall detection. After exporting the VoTT file, the corresponding TXT file is obtained which is shown in Table 2. Each row consists of columns in which each row column denotes a particular property: the first row of the column indicates the image class defined in the VoTT software, and the second and third denote the center of the object along the x and y axes, fourth and fifth denote the width and height of the element [43]. With the successful completion of these actions, the dataset is prepared for training the machine learning algorithm.

2.2. Description of Methods

YOLOV4 is the most advanced detection algorithm with a real-time object recognition system that can recognize multiple objects in a single frame [44]. It uses a completely different approach from previous detection systems where it applies one neural network to the whole frame, i.e., the image, and then the neural network divides the image into regions and predicts the boundary frames which then calculates the probability for each possible region, i.e., class of objects. According to [44], the basic idea of the YOLOV4 algorithm is the division of input data into a matrix of a certain dimension by which each matrix address is responsible for predicting objects within the center of that cell where then for every object that algorithm can detect, the probability coefficient is calculated visible from Figure 6.
As seen in Figure 6, each matrix provides boundary frameworks and calculates a reliability estimate for the corresponding framework. After the reliability estimate is calculated the reliability score is determined which shows how secure the model is and that the frame contains the object [45]. This type of detection algorithm has several advantages: it can recognize several objects in one frame [46], during testing the entire image frame is reviewed so that predictions are based on the global image context and it also provides predictions with a single evaluation of the network as unlikely similar methods such as R-CNN which require thousands of examples of a single image [47,48]. Due to the aforementioned properties, the YOLOV4 algorithm is extremely fast even up to a thousand times faster than R-CNN, and up to a hundred times faster, than Fast R-CNN [49]. The YOLOV4 design allows end-to-end training and real-time speed while maintaining high average accuracy [50] and its flow chart can be seen in Figure 7.
The process starts with taking an image from the COCO dataset or custom-labeled dataset where the trained weights are pre-trained so it is necessary to update and adjust, i.e., fine-tune the weights then the process of predicting bounding boxes begins, which results in large boxes and small boxes where the Non-Max Suppression (NMS) procedure obtains the main box that at the end of detection is the result itself [51,52]. NMS is defined as a local maximum where the local maximum is greater than all of its neighbors except himself. For some n neighborhood, any pixel consists of 1-D case, i.e., n pixel left and right and in the case of the square region, it must be centered around the pixel under consideration. NMS consists of several computer vision algorithms [53], some of them are the measure of the selected point in the whole image, and even the entire scale space, and then the local maximum is selected. Efficiency is a major feature in its applications and the only competitor to NMS is the Maximum filter [54] which calculates the value of each pixel of each neighbor, where the process is much slower than NMS. Given the maximum filter response in each pixel, NMS reduces additional comparative values of each pixel with the largest neighbor. After training the YOLOV4, it is necessary to evaluate the obtained results. There are several suitable metrics, but the most important and generally accepted metrics are mAP and IoU. Using these metrics, the algorithm calculates a reliability coefficient ranging from 0 to 1, i.e., from 0% to 100% confidence that the predicted class from final detection is accurate and true. Depending on the results, it is necessary to either additionally train the algorithm or adjust training parameters to obtain better and more accurate results.

2.3. YOLOV4 Limitations and Accuracy Measures

Although YOLOV4 seems to be the best algorithm for detecting application objects, it is essential to pay attention to a few problems and limitations that can occur along the way. The detection algorithm has problems separating small objects in images that appear in groups, although the fourth version of the YOLOV4 has been improved a lot compared to its predecessors [55]. The reason for this is that each neural network is limited to detecting an individual object [22] therefore, the above detection algorithm makes it difficult to detect and localize small objects that are naturally found in group images. IoU is one of the significant metrics needed to measure the accuracy and localization of objects in a frame which can be calculated by finding the area that intersects between the limit frame for a particular prediction and the assumed accuracy limits of that area first taken. The total area covered by these two boundary frames (Union), is then calculated. The intersection divided by the union gives the ratio of the overlap to the total area, which gives a relatively good estimate of how close the boundary frame is to the source prediction, and then after the defined terms, IoU [56,57,58] can be calculated as Equation (1), where A is the area of prediction box and B is the area of target box:
IoU = | A B | | A B |
In Equation (1), the numerator represents the calculation of the intersection/overlap area between the predicted boundary box [59] and the ground-truth bounding box [60] while the Union represents the total union between them. The IoU is in a range from 0 to 100 and the closer the value is to 100 the better it is, as shown in Figure 8.
Accuracy is measured with an evaluation metric mAP, which measures in what percentage the algorithm predicted the object from all individual classes correctly and it is expressed in percentages that go from 0% to 100%. The mathematical description of mAP can be represented by Equation (2). Where the A P k is the average Precision of the classes and the n is the number of classes [61].
mAP = 1 n k = 1 k = n A P k .
The Precision determines how accurate your model is and it can be calculated with the expression [62]:
Precision = TP TP + FP .
The recall is a measure that shows how many times the model finds all the positive cases while training Recall [62] and can be calculated with:
Recall = TP TP + FN .
Finally, the F1 score is an evaluation metric that measures models accuracy in combination with Recall and Precision [63] and can be calculated with the expression:
F 1 score = 2 × Recall × Precision Recall + Precision ,
where:
  • TP—True Positive predictions,
  • FP—False Positive predictions, and
  • FN—False Negative predictions.
YOLOV4 has intrusive spatial constraints in terms of bounding box predictions hence each grid cell can only predict two possible boxes and can only have one unique class. This constraint can limit the number of objects that the model can predict. A model can struggle with small objects that appear in groups such as a flock of birds or a group of ants [64].

Configuration, Setup, and Hyper-Parameters Used for YOLOV4 Project

Before training the YOLOV4 algorithm, it was necessary to define a few configuration files within the YOLOV4 folder. Data folder obj.data contains some vital information where in several lines is described the number of classes that the algorithm must train, the location of the train.txt file that is separated from the total dataset for neural network training, the location of obj.names that containing class names, and test.txt which is used to check or test the neural network. Folders train.txt and test.txt can vary depending on the division of the total dataset into a train set and a test set in the process.py python script. Given that a relatively small set of data was used in this research it was necessary to change a train/test ratio in python script proces.py to get better results. After configuring the above files, the next step was to configure the yolov4-custom.cfg file, which defines max batch size, and set this parameter the total number of training examples present in one series, batch, and subdivision which determines how many images can be processed in parallel, learning rate, etc. All changed hyperparameters (max batch size, activation functions) and train/test dataset ratios are listed in Table 3. By changing, hyperparameters and train/test ratios the results of the evaluation metrics were changed.
As seen from Table 3 there are three activation functions used, i.e., ReLU, Mish, and Linear. Several max batch sizes—6000, 8000, and 10,000—were used for each activation function. However, an additional max batch size of 20,000 was used with the Mish and ReLU activation functions. Due to the relatively small dataset train/test, the ratio has been modified with the most common used train/test ratios that are: 70/30, 75/25, and 80/20 [65]. These ratios were chosen due to the structure, more specifically, the size, of the dataset. Given that it is a small dataset, by choosing a training set ratio greater than 80% or less than 70%, there is a very high probability of overfitting, which is most recommended to be avoided. Another reason the authors used this range of the training dataset (70–80%) is that it is the most commonly used range [66].
After the training with specific parameters that have been changed in the configuration file a certain loss [67] occurred as in the expression:
ε = λ coord i = 0 S 2 j = 0 B i j obj x i x ^ i 2 + y i y ^ i 2 + λ coord i = 0 S 2 j = 0 B i j obj w i w ^ i 2 + h i h ^ i 2 + i = 0 S 2 j = 0 B i j obj C i C ^ i 2 + λ noobj i = 0 S 2 j = 0 B i j noobj C i C ^ i 2 + i = 0 S 2 i obj c classes p i ( c ) p ^ i ( c ) 2 ,
where:
  • x and y-coordinates of a particular grid in offset in a range from 0 to 1,
  • λ coord and λ noobj indicates parameters for increasing the loss value of bounding box coordinates while predicting and decreasing the loss value of confidence predictions bounding boxes that does not appear to contain any objects,
  • i obj indicates whether the object appears in cell i and
  • i j obj indicates that the j-th predictor of the boundary frame in the cell is “responsible” for that prediction.
Notably, the loss function can only penalize the classification error if the represented object is in that specific network cell. Unnecessary problems can possibly occur, so it is recommended to use the value of λ coord = 5 and λ noobj = 0.5 [67]. Sum squared error can be also equal to errors in big boxes and small boxes; however, given error metrics should directly reflect on small deviations in big boxes and they matter less than in small boxes. To avoid this problem and confusion bounding boxes for predicting square root error are partially calculated instead of measuring the width and height of boxes directly. YOLOV4 can predict more than one bounding box per fully calculated grid cell. While training, it is necessary to assign a predictor tool that has the task mainly responsible for predicting an object based on its highest IoU with its ground truth. Each possible predictor gets better and it is improving in predicting certain aspects such as size, aspect ratio, or even class of the object shown with expression (6) [67].

3. Results and Discussion

In this section, the results of the investigation on the influence of YOLOV4 hyperparameters (max batch size, and activation functions) and train/test ratios on classification accuracy are presented using mAP, Average IoU, F1 score , Precision, and Recall metrics. Combinations of hyperparameters and train/test ratios are shown in Table 3 and the key points during the training of the YOLOV4 algorithm are described. In addition, the average loss charts during training and the metric evaluation of the trained model are available in the Appendix A, Appendix B, Appendix C and Appendix D section. The Appendix subsection consists of four categories and three subcategories, while the categories represent each training cycle of 6000, 8000, 10,000, and 20,000 max batch size, and the subcategories represent the Linear ReLU and Mish activation functions. The final table of evaluation metrics can be seen in Table 4 and Table 5.

3.1. Results

As previously stated, the obtained results are classified into four cycles. Based on the obtained results, it is evident that during the first 1200 iterations the loss curve has a higher value and it is greater than 2000. When reaching the 1200th iteration, there is a rapid reduction of the average loss curve until the 1800th iteration. From that moment on, the loss curve drops linearly until the final iteration. In the last defined iteration of training, the final value of the average loss is obtained, shown in Table 4 and Table 5. The above analysis refers to all trained models of this research. Appendix A represents the average loss charts for 6000, Appendix B for 8000, Appendix C for 10,000, and Appendix D for 20,000 max batches. Appendix A.1 and Appendix B.1 represents Linear activation function. Appendix A.2, Appendix B.2, Appendix C.1 and Appendix D.1 represents ReLU. The final Mish activation function represents Appendix A.3, Appendix B.3, Appendix C.2 and Appendix D.2 appendix subsections.
The results of the first and second training cycles, i.e., are given in Table 4. Linear, ReLU, and Mish activation functions were used in 70/30,75/25, and 80/20 train/test ratios. All models were evaluated with mAP, F1 score , Average IoU, Precision, Recall, and Average loss metrics.
In the first part of the first training cycle for the Linear activation function with 70/30, 75/25, and 80/20 train test ratios, the following results were obtained: mAP was between 59.11% and 60.18%, F1 score was in a range between 0.47 and 0.52, Average IoU varied between 28.1% and 29.98%, the Precision score was between 0.40 and 0.44 while the Recall and Average loss in between 0.58 and 0.64 and 2.4285 and 2.6178. Progress in obtaining the final amount of Average loss can be seen in the Loss charts given in Appendix A.1.
The second part of the first training cycle for ReLU activation function evaluation metrics is changed for all tree train/test ratios. The mAP was between 94.16% and 95.12%, the F1 score was between 0.81 and 0.87, while the Precision and Recall had similar values where the minimum value was 0.75 and the maximum was 0.93, the best Average IoU is 62.18% and the worst 58.21%. The Average loss did not change a lot and it was between 1.18 and 1.26. Charts of Average loss are presented in Appendix A.2.
In the third part of the first cycle for the Mish activation function, the following results peak values are the following: mAP value of 94.25%, F1 score of 0.86, Average IoU 61.52%, Precision 0.81, Recall 0.93 and Average loss in the amount of 1.2489. As in the previous two parts, average loss charts can be seen in Appendix A.3.
For the fourth and fifth, parts of the second training cycle, the results of the obtained models are slightly better than in the first three parts. The maximum mAP value is 99.77% in favor of the ReLU activation function, while the minimum value is 91.27% in favor of the Linear activation function. The maximum value of the F1 score was on several occasions 0.97 for the ReLU and Mish activation functions, which was also the peak value, while the minimum value was 0.59 for the Linear activation function. Charts of the training process can be seen in Appendix C in Appendix C.1 and Appendix C.2.
The results of the third and fourth training cycles, i.e., 10,000 and 20,000 max batch size are given in Table 5. The Linear activation function is eliminated because of poor performance from the previous investigation. As in the last case, each training cycle of this research was conducted using the previous three train/test ratios. In the last training cycle, it is evident that the lowest average loss results were obtained compared to other training cycles. The results of mAP vary between 88.33% and 99.96%, the highest F1 score was 1.00 while the lowest was 0.86. Both Precision and Recall ranged between 0.79 and 1. The highest score achieved for the Average IoU was 91.13%. The whole training process and average loss curve for the last two training cycles can be seen in Appendix D. The training process for the Relu activation function is shown in Appendix D.1 and for Mish in Appendix D.2.

3.2. Discussion

As seen from Table 4 and Table 5 the entire investigation conducted in this paper can be divided into four different cycles based on max batch sizes, i.e., 6000, 8000, 10,000, and 20,000. The first cycle had 6000 batches with different activation functions and different train/test ratios. The ReLU activation function with a 75/25 ratio had the best performance in this cycle as well it had a low average loss and the best mAP percentage score.
In the second cycle of training again ReLU activation function was the best but Mish was close with its performance and metrics while in the third cycle only Mish and ReLU activation functions were evaluated. The main reason for that is because they outperformed the Linear activation function and there is no need to waste time on training the YOLOV4 algorithm with under-performed hyperparameters. In the third cycle, ReLU is not an absolute choice because the only thing that was better while using the Mish activation function is average loss. Referring to Table 5 it is visible in row Average loss that Mish is slightly better than ReLU but still, these metrics are not good enough for implementation in real situations.
In the last training cycle, the ReLu and Mish activation functions were re-evaluated for all three train test ratios. All of the metrics are higher, which contributes to the very detection of returnable packaging. A possible explanation for this is the Mish activation function has a low operating cost with a variety of properties and smooth, non-monotonic nature. It is unbounded above and bounded below that means it can avoid saturation which can cause the training to slow down. That results in more flexibility while training the algorithm and improving it [68]. The training of the algorithm was stopped at the moment when the average losses were no longer decreasing.
Looking at the fourth cycle it can be noticed that with the Mish activation function in the 75/25 train/test ratio gets outstanding results. The mAP is nearly 100%, the F1 score is 1.00 and the Average IoU is 91.13%. The only thing that can be improved is to minimize the Average loss which is at this moment 0.3643, but there is a risk to Overfit detection algorithm model. At this moment while testing the trained detector if the thresh is put below 0.85%, the algorithm could detect wrong which is not preferable. A Linear activation function is the worst by all means while training this detection algorithm, it is not flexible as Mish and it has poor quality as shown in Table 4. By statistics, the ReLU activation function is the best but further analysis is needed to explore all possibilities while training the YOLOV4 algorithm.
Based on the values shown in Table 4 and Table 5, it can be seen that the train/test ratio has slight or no influence on the classification accuracy when compared to two hyperparameters (max batch size and activation functions). Obtained weights from the YOLOV4 model were tested with new input images that previously did not belong to the initial dataset used in this investigation. The testing was conducted using a web camera, where the percentage of reliability varied depending on the object location or angle of the camera visible from Figure 9, Figure 10, Figure 11 and Figure 12.
Detection results for glass packaging are shown on Figure 9. It is visible that the YOLOV4 algorithm perfectly detects and classifies the given object. It is also visible a high coefficient of reliability for every object in the image. Plastic packaging detection results are shown in Figure 10 where it can be seen that the YOLOV4 algorithm perfectly detects and classifies plastic packaging with a high coefficient of reliability on every object on the image.
Aluminum packaging detection can be seen on Figure 11. As previous 2 classes, it is again visible that YOLOV4 perfectly detects and classifies the aluminum class of packaging.
The last given example of detection is a mix of all 3 classes together on the same spot as shown in Figure 12. The trained model of the YOLOV4 algorithm detects all returnable packaging with a high coefficient of reliability.

4. Conclusions

This paper presents a detection algorithm for returnable packaging recycling. The method of creating a dataset, marking bounding boxes, and what to pay attention to while compiling the dataset is shown. In addition, the method of training the desired algorithm is shown, as settings that are used with its corresponding files, and the overall configuration is shown. Toward the end of the paper, the results are presented and the differences between each training cycle are explained. Quality results can be achieved through the use of AI. For example, by using CNN, it is possible to achieve high-quality detection algorithms. In this case, YOLOV4 gives almost perfect results. So the answers to the hypotheses defined in the introduction section are:
  • The YOLOV4 algorithm can achieve high classification accuracy in process of detection and classification of returnable packaging; however, the tuning process of YOLOV4 algorithm hyperparameters must be performed.
  • The classification accuracy was improved by tuning just two hyperparameters: max batch size, and type of activation function used in the YOLOV4 algorithm. The investigation showed that max batch size has a greater influence than the type of activation function.
  • The train/test dataset ratio did not have a notable influence on classification accuracy for a couple of reasons. The first reason is that the values of train/test ratios were the same as those ratios that are commonly used in other research papers. The lower size of the training dataset size would cause a weakly trained YOLOV4 model, while the higher size of the training dataset could cause potential overfitting of the YOLOV4 model.
This paper shows different properties of the neural networks, in this case, it can be seen that by changing the hyperparameters such as max batch size, and activation function, different results can be achieved. The quality indicators of the trained algorithm were mAP, Average loss, and Average IoU, whereby YOLOV4 at 20,000 batches in train/test ratio of 75/25 with Mish activation function shows the flexibility of data processing and gives great results. The worst example of this work is the use of the Linear activation function which shows the worst results in all training cycles. ReLU and Mish show outstanding results, but in this case, Mish prevails because of the computational properties, training speed, and data processing flexibility which has a greater application than the ReLU or Leaky ReLU activation function. These results are shown in Table 4 and Table 5, in which in the last training cycle YOLOV4 achieves results of mAP 99.96%, F1 score 1.00 Average IoU 91.13%, and the Precision and Recall of 1.00, which is almost perfect considering the possible threat of the overfitting problem. The main contribution of this paper concerns the selection of different activation functions. The YOLOV4 detection algorithm uses the basic Mish activation function, which achieves superior properties however, the ReLU function has different computational properties than Mish. For this reason, it is not recommended to reject the ReLU activation function, because, for a different type of dataset, it is most likely that ReLU can surpass the results that are normally obtained using the default Mish activation function. Also, the importance of this paper is in devising innovative ideas that could accelerate the recycling process of recyclable materials/packaging. The problems of global climate change are visible in several aspects and future life must prevent natural disasters. Although the results are practically perfect, it is necessary to devote further research related to changing the YOLOV4 hyperparameters or YOLOV4 architecture to achieve even better and more accurate results.

Author Contributions

Conceptualization, Z.C.; methodology, M.G., S.B.Š. and N.A.; software, M.G.; validation, S.B.Š. and N.A.; formal analysis, N.A. and Z.C.; investigation, M.G. and S.B.Š.; resources, S.B.Š. and N.A.; data curation, M.G. and Z.C.; writing—original draft preparation, M.G. and S.B.Š.; writing—review and editing, N.A. and Z.C.; visualization, M.G.; supervision Z.C.; project administration Z.C.; funding acquisition, S.B.Š. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

A partial set of the used dataset can be found at the link: https://www.kaggle.com/datasets/arkadiyhacks/drinking-waste-classification, (accessed on 11 October 2022).

Acknowledgments

This research has been (partly) supported by the CEEPUS network CIII-HR-0108, European Regional Development Fund under the grant KK.01.1.1.01.0009 (DATACROSS), project CEKOM under the grant KK.01.2.2.03.0004, Erasmus+ project WICT under the grant 2021-1-HR01-KA220-HED-000031177 and University of Rijeka scientific grant uniri-tehnic-18-275-1447.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In this part of the Appendix section, training charts of the YOLOV4 algorithm model for 6000 max batches are attached, each sub-section represents a single activation function for all three train/test ratios.

Appendix A.1

Figure A1. Average loss curve for 6000 max batches 70/30 train/test ratio for Linear activation function.
Figure A1. Average loss curve for 6000 max batches 70/30 train/test ratio for Linear activation function.
Applsci 12 11131 g0a1
Figure A2. Average loss curve for 6000 max batches 75/25 train/test ratio for Linear activation function.
Figure A2. Average loss curve for 6000 max batches 75/25 train/test ratio for Linear activation function.
Applsci 12 11131 g0a2
Figure A3. Average loss curve for 6000 max batches 80/20 train/test ratio for Linear activation function.
Figure A3. Average loss curve for 6000 max batches 80/20 train/test ratio for Linear activation function.
Applsci 12 11131 g0a3

Appendix A.2

Figure A4. Average loss curve for 6000 max batches 70/30 train/test ratio for ReLU activation function.
Figure A4. Average loss curve for 6000 max batches 70/30 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a4
Figure A5. Average loss curve for 6000 max batches 75/25 train/test ratio for ReLU activation function.
Figure A5. Average loss curve for 6000 max batches 75/25 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a5
Figure A6. Average loss curve for 6000 max batches 80/20 train/test ratio for ReLU activation function.
Figure A6. Average loss curve for 6000 max batches 80/20 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a6

Appendix A.3

Figure A7. Average loss curve for 6000 max batches 70/30 train/test ratio for Mish activation function.
Figure A7. Average loss curve for 6000 max batches 70/30 train/test ratio for Mish activation function.
Applsci 12 11131 g0a7
Figure A8. Average loss curve for 6000 max batches 75/25 train/test ratio for Mish activation function.
Figure A8. Average loss curve for 6000 max batches 75/25 train/test ratio for Mish activation function.
Applsci 12 11131 g0a8
Figure A9. Average loss curve for 6000 max batches 80/20 train/test ratio for Mish activation function.
Figure A9. Average loss curve for 6000 max batches 80/20 train/test ratio for Mish activation function.
Applsci 12 11131 g0a9

Appendix B

In this part of the Appendix section, training charts of the YOLOV4 algorithm model for 8000 max batches are attached, each sub-section represents a single activation function for all three train/test ratios.

Appendix B.1

Figure A10. Average loss curve for 8000 max batches 70/30 train/test ratio for Linear activation function.
Figure A10. Average loss curve for 8000 max batches 70/30 train/test ratio for Linear activation function.
Applsci 12 11131 g0a10
Figure A11. Average loss curve for 8000 max batches 75/25 train/test ratio for Linear activation function.
Figure A11. Average loss curve for 8000 max batches 75/25 train/test ratio for Linear activation function.
Applsci 12 11131 g0a11
Figure A12. Average loss curve for 8000 max batches 80/20 train/test ratio for Linear activation function.
Figure A12. Average loss curve for 8000 max batches 80/20 train/test ratio for Linear activation function.
Applsci 12 11131 g0a12

Appendix B.2

Figure A13. Average loss curve for 8000 max batches 70/30 train/test ratio for ReLU activation function.
Figure A13. Average loss curve for 8000 max batches 70/30 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a13
Figure A14. Average loss curve for 8000 max batches 75/25 train/test ratio for ReLU activation function.
Figure A14. Average loss curve for 8000 max batches 75/25 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a14
Figure A15. Average loss curve for 8000 max batches 80/20 train/test ratio for ReLU activation function.
Figure A15. Average loss curve for 8000 max batches 80/20 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a15

Appendix B.3

Figure A16. Average loss curve for 8000 max batches 70/30 train/test ratio for Mish activation function.
Figure A16. Average loss curve for 8000 max batches 70/30 train/test ratio for Mish activation function.
Applsci 12 11131 g0a16
Figure A17. Average loss curve for 8000 max batches 75/25 train/test ratio for Mish activation function.
Figure A17. Average loss curve for 8000 max batches 75/25 train/test ratio for Mish activation function.
Applsci 12 11131 g0a17
Figure A18. Average loss curve for 8000 max batches 80/20 train/test ratio for Mish activation function.
Figure A18. Average loss curve for 8000 max batches 80/20 train/test ratio for Mish activation function.
Applsci 12 11131 g0a18

Appendix C

In this part of the Appendix section, training charts of the YOLOV4 algorithm model for 10,000 max batches are attached, each sub-section represents a single activation function for all three train/test ratios.

Appendix C.1

Figure A19. Average loss curve for 10,000 max batches 70/30 train/test ratio for ReLU activation function.
Figure A19. Average loss curve for 10,000 max batches 70/30 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a19
Figure A20. Average loss curve for 10,000 max batches 75/25 train/test ratio for ReLU activation function.
Figure A20. Average loss curve for 10,000 max batches 75/25 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a20
Figure A21. Average loss curve for 10,000 max batches 80/20 train/test ratio for ReLU activation function.
Figure A21. Average loss curve for 10,000 max batches 80/20 train/test ratio for ReLU activation function.
Applsci 12 11131 g0a21

Appendix C.2

Figure A22. Average loss curve for 10,000 max batches 70/30 train/test ratio for Mish activation function.
Figure A22. Average loss curve for 10,000 max batches 70/30 train/test ratio for Mish activation function.
Applsci 12 11131 g0a22
Figure A23. Average loss curve for 10,000 max batches 75/25 train/test ratio for Mish activation function.
Figure A23. Average loss curve for 10,000 max batches 75/25 train/test ratio for Mish activation function.
Applsci 12 11131 g0a23
Figure A24. Average loss curve for 10,000 max batches 80/20 train/test ratio for Mish activation function.
Figure A24. Average loss curve for 10,000 max batches 80/20 train/test ratio for Mish activation function.
Applsci 12 11131 g0a24

Appendix D

In this part of the Appendix section, training charts of the YOLOV4 algorithm model for 20,000 max batches are attached, each sub-section represents a single activation function for all three train/test ratios.

Appendix D.1

Figure A25. Average loss curve for 20,000 max batches 70/30 train/test ratio for ReLu activation function.
Figure A25. Average loss curve for 20,000 max batches 70/30 train/test ratio for ReLu activation function.
Applsci 12 11131 g0a25
Figure A26. Average loss curve for 20,000 max batches 75/25 train/test ratio for ReLu activation function.
Figure A26. Average loss curve for 20,000 max batches 75/25 train/test ratio for ReLu activation function.
Applsci 12 11131 g0a26
Figure A27. Average loss curve for 20,000 max batches 80/20 train/test ratio for ReLu activation function.
Figure A27. Average loss curve for 20,000 max batches 80/20 train/test ratio for ReLu activation function.
Applsci 12 11131 g0a27

Appendix D.2

Figure A28. Average loss curve for 20,000 max batches 70/30 train/test ratio for Mish activation function.
Figure A28. Average loss curve for 20,000 max batches 70/30 train/test ratio for Mish activation function.
Applsci 12 11131 g0a28
Figure A29. Average loss curve for 20,000 max batches 75/25 train/test ratio for Mish activation function.
Figure A29. Average loss curve for 20,000 max batches 75/25 train/test ratio for Mish activation function.
Applsci 12 11131 g0a29
Figure A30. Average loss curve for 20,000 max batches 80/20 train/test ratio for Mish activation function.
Figure A30. Average loss curve for 20,000 max batches 80/20 train/test ratio for Mish activation function.
Applsci 12 11131 g0a30

References

  1. Reno, J. Waste and waste management. Annu. Rev. Anthropol. 2015, 44, 557–572. [Google Scholar] [CrossRef] [Green Version]
  2. Guan, L. Multimedia Image and Video Processing; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  3. Alsanabani, A.A.; Ahmed, M.A.; Al Smadi, A.A. Vehicle counting using detecting-tracking combinations: A comparative analysis. In Proceedings of the 4th International Conference on Video and Image Processing, Shanghai, China, 25–27 November 2020; pp. 48–54. [Google Scholar]
  4. Wu, J.; Osuntogun, A.; Choudhury, T.; Philipose, M.; Rehg, J.M. A scalable approach to activity recognition based on object use. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  5. Kumar, A.; Kaur, A.; Kumar, M. Face detection techniques: A review. Artif. Intell. Rev. 2019, 52, 927–948. [Google Scholar] [CrossRef]
  6. Feng, D.; Harakeh, A.; Waslander, S.L.; Dietmayer, K. A review and comparative study on probabilistic object detection in autonomous driving. IEEE Trans. Intell. Transp. Syst. 2021, 23, 9961–9980. [Google Scholar] [CrossRef]
  7. Aydin, I.; Othman, N.A. A new IoT combined face detection of people by using computer vision for security application. In Proceedings of the 2017 International Artificial Intelligence and Data Processing Symposium (IDAP), Malatya, Turkey, 16–17 September 2017; pp. 1–6. [Google Scholar]
  8. Husain, A.A.; Maity, T.; Yadav, R.K. Vehicle detection in intelligent transport system under a hazy environment: A survey. IET Image Process. 2020, 14, 1–10. [Google Scholar] [CrossRef]
  9. Alsubaei, F.S.; Al-Wesabi, F.N.; Hilal, A.M. Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment. Appl. Sci. 2022, 12, 2281. [Google Scholar] [CrossRef]
  10. Mao, W.L.; Chen, W.C.; Wang, C.T.; Lin, Y.H. Recycling waste classification using optimized convolutional neural network. Resour. Conserv. Recycl. 2021, 164, 105132. [Google Scholar] [CrossRef]
  11. Yang, M.; Thung, G. Classification of trash for recyclability status. CS229 Proj. Rep. 2016, 2016, 3. [Google Scholar]
  12. Adedeji, O.; Wang, Z. Intelligent waste classification system using deep learning convolutional neural network. Procedia Manuf. 2019, 35, 607–612. [Google Scholar] [CrossRef]
  13. Bobulski, J.; Kubanek, M. Waste classification system using image processing and convolutional neural networks. In International Work-Conference on Artificial Neural Networks; Springer: Cham, Switzerland, 2019; pp. 350–361. [Google Scholar]
  14. Shi, C.; Tan, C.; Wang, T.; Wang, L. A waste classification method based on a multilayer hybrid convolution neural network. Appl. Sci. 2021, 11, 8572. [Google Scholar] [CrossRef]
  15. Meng, S.; Chu, W.T. A study of garbage classification with convolutional neural networks. In Proceedings of the 2020 Indo–Taiwan 2nd International Conference on Computing, Analytics and Networks (Indo-Taiwan ICAN), Rajpura, India, 7–15 February 2020; pp. 152–157. [Google Scholar]
  16. Cchangcs. Garbage Classification. 2018. Available online: https://www.kaggle.com/datasets/asdasdasasdas/garbage-classification (accessed on 11 October 2022).
  17. Altikat, A.; Gulbe, A.; Altikat, S. Intelligent solid waste classification using deep convolutional neural networks. Int. J. Environ. Sci. Technol. 2022, 19, 1285–1292. [Google Scholar] [CrossRef]
  18. Aishwarya, A.; Wadhwa, P.; Owais, O.; Vashisht, V. A waste management technique to detect and separate non-biodegradable waste using machine learning and YOLO algorithm. In Proceedings of the 2021 11th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 28–29 January 2021; pp. 443–447. [Google Scholar]
  19. Wahyutama, A.B.; Hwang, M. YOLO-Based Object Detection for Separate Collection of Recyclables and Capacity Monitoring of Trash Bins. Electronics 2022, 11, 1323. [Google Scholar] [CrossRef]
  20. Lin, W. YOLO-Green: A Real-Time Classification and Object Detection Model Optimized for Waste Management. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 51–57. [Google Scholar]
  21. Ye, A.; Pang, B.; Jin, Y.; Cui, J. A YOLO-based neural network with VAE for intelligent garbage detection and classification. In Proceedings of the 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence, Sanya, China, 24–26 December 2020; pp. 1–7. [Google Scholar]
  22. Kim, J.A.; Sung, J.Y.; Park, S.H. Comparison of Faster-RCNN, YOLO, and SSD for real-time vehicle type recognition. In Proceedings of the 2020 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Seoul, Korea, 1–3 November 2020; pp. 1–4. [Google Scholar]
  23. Kulshreshtha, M.; Chandra, S.S.; Randhawa, P.; Tsaramirsis, G.; Khadidos, A.; Khadidos, A.O. OATCR: Outdoor autonomous trash-collecting robot design using YOLOv4-tiny. Electronics 2021, 10, 2292. [Google Scholar] [CrossRef]
  24. Tian, M.; Li, X.; Kong, S.; Wu, L.; Yu, J. A modified YOLOv4 detection method for a vision-based underwater garbage cleaning robot. Front. Inf. Technol. Electron. Eng. 2022, 23, 1217–1228. [Google Scholar] [CrossRef]
  25. Drinking Waste Classification. Available online: https://www.kaggle.com/arkadiyhacks/drinking-waste-classification (accessed on 24 November 2021).
  26. Liu, Z.J.; Chen, L.; Wu, D.; Ding, W.; Zhang, H.; Zhou, W.; Fu, Z.Q.; Wang, B.C. A multi-dataset data-collection strategy produces better diffraction data. Acta Crystallogr. Sect. A Found. Crystallogr. 2011, 67, 544–549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Herrera, Y.M.; Kapur, D. Improving data quality: Actors, incentives, and capabilities. Political Anal. 2007, 15, 365–386. [Google Scholar] [CrossRef] [Green Version]
  28. Han, K.; Lee, S.; Lee, W.; Lee, J.; Lee, D.h. An Evaluation Dataset and Strategy for Building Robust Multi-turn Response Selection Model. arXiv 2021, arXiv:2109.04834. [Google Scholar]
  29. Howse, J. OpenCV Computer Vision with Python; Packt Publishing: Birmingham, UK, 2013; Volume 27. [Google Scholar]
  30. Bui, H.M.; Lech, M.; Cheng, E.; Neville, K.; Burnett, I.S. Using grayscale images for object recognition with convolutional-recursive neural network. In Proceedings of the 2016 IEEE Sixth International Conference on Communications and Electronics (ICCE), Ha-Long, Vietnam, 27–29 July 2016; pp. 321–325. [Google Scholar]
  31. Grum, E.; Vasseure, B. How to Select the Best Dataset for a Task? 2004. Available online: https://pdfhall.com/how-to-select-the-best-dataset-for-a-task_5b6ecd57097c47a9568b4687.html (accessed on 11 October 2022).
  32. Soekhoe, D.; Van Der Putten, P.; Plaat, A. On the impact of data set size in transfer learning using deep neural networks. In International Symposium on Intelligent Data Analysis; Springer: Cham, Switzerland, 2016; pp. 50–60. [Google Scholar]
  33. Feng, S.; Zhou, H.; Dong, H. Using deep neural network with small dataset to predict material defects. Mater. Des. 2019, 162, 300–310. [Google Scholar] [CrossRef]
  34. Lippi, M.; Bonucci, N.; Carpio, R.F.; Contarini, M.; Speranza, S.; Gasparri, A. A yolo-based pest detection system for precision agriculture. In Proceedings of the 2021 29th Mediterranean Conference on Control and Automation (MED), Puglia, Italy, 22–25 June 2021; pp. 342–347. [Google Scholar]
  35. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
  36. Taylor, L.; Nitschke, G. Improving deep learning with generic data augmentation. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 1542–1547. [Google Scholar]
  37. Beyeler, M. Machine Learning for OpenCV; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  38. Ramyachitra, D.; Manikandan, P. Imbalanced dataset classification and solutions: A review. Int. J. Comput. Bus. Res. (IJCBR) 2014, 5, 1–29. [Google Scholar]
  39. Picard, S.; Chapdelaine, C.; Cappi, C.; Gardes, L.; Jenn, E.; Lefèvre, B.; Soumarmon, T. Ensuring Dataset Quality for Machine Learning Certification. In Proceedings of the 2020 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW), Coimbra, Portugal, 12–15 October 2020; pp. 275–282. [Google Scholar]
  40. Yadav, S.; Ekbal, A.; Saha, S.; Bhattacharyya, P. Deep learning architecture for patient data de-identification in clinical records. In Proceedings of the Clinical Natural Language Processing Workshop (ClinicalNLP), Osaka, Japan, 11 December 2016; pp. 32–41. [Google Scholar]
  41. Brownlee, J. Data Preparation for Machine Learning: Data Cleaning, Feature Selection, and Data Transforms in Python; Machine Learning Mastery: Melbourne, Australia, 2020. [Google Scholar]
  42. Lin, J.P.; Sun, M.T. A YOLO-based traffic counting system. In Proceedings of the 2018 Conference on Technologies and Applications of Artificial Intelligence (TAAI), Taichung, Taiwan, 30 November–2 December 2018; pp. 82–85. [Google Scholar]
  43. Dharneeshkar, J.; Aniruthan, S.; Karthika, R.; Parameswaran, L. Deep Learning based Detection of potholes in Indian roads using YOLO. In Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020; pp. 381–385. [Google Scholar]
  44. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  45. Yu, J.; Zhang, W. Face mask wearing detection algorithm based on improved YOLO-v4. Sensors 2021, 21, 3263. [Google Scholar] [CrossRef]
  46. Ma, J.; Chen, L.; Gao, Z. Hardware implementation and optimization of tiny-yolo network. In International Forum on Digital TV and Wireless Multimedia Communications; Springer: Singapore, 2017; pp. 224–234. [Google Scholar]
  47. Fast, P.D.U.A.B. R-CNN. In Digital TV and Wireless Multimedia Communication: Proceedings of the 14th International Forum, IFTC 2017, Shanghai, China, 8–9 November 2017; Revised Selected Papers; Springer: Berlin/Heidelberg, Germany, 2018; Volume 815, p. 172. [Google Scholar]
  48. Roh, M.C.; Lee, J.y. Refining faster-RCNN for accurate object detection. In Proceedings of the 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, Japan, 8–12 May 2017; pp. 514–517. [Google Scholar]
  49. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 October 2015; pp. 1440–1448. [Google Scholar]
  50. Chen, S.; Lin, W. Embedded system real-time vehicle detection based on improved YOLO network. In Proceedings of the 2019 IEEE 3rd Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 11–13 October 2019; pp. 1400–1403. [Google Scholar]
  51. Shinde, S.; Kothari, A.; Gupta, V. YOLO based human action recognition and localization. Procedia Comput. Sci. 2018, 133, 831–838. [Google Scholar] [CrossRef]
  52. Neubeck, A.; Van Gool, L. Efficient non-maximum suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; Volume 3, pp. 850–855. [Google Scholar]
  53. Hosang, J.; Benenson, R.; Schiele, B. Learning non-maximum suppression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4507–4515. [Google Scholar]
  54. Wang, A.; Hu, C.; Liu, X.; Iwahori, Y.; Kang, R. A modified non-maximum suppression algorithm. In Information Science and Electronic Engineering; CRC Press: Boca Raton, FL, USA, 2016; pp. 83–86. [Google Scholar]
  55. Shi, P.; Jiang, Q.; Shi, C.; Xi, J.; Tao, G.; Zhang, S.; Zhang, Z.; Liu, B.; Gao, X.; Wu, Q. Oil Well Detection via Large-Scale and High-Resolution Remote Sensing Images Based on Improved YOLO v4. Remote Sens. 2021, 13, 3243. [Google Scholar] [CrossRef]
  56. Baressi Šegota, S.; Lorencin, I.; Smolić, K.; Anđelić, N.; Markić, D.; Mrzljak, V.; Štifanić, D.; Musulin, J.; Španjol, J.; Car, Z. Semantic Segmentation of Urinary Bladder Cancer Masses from CT Images: A Transfer Learning Approach. Biology 2021, 10, 1134. [Google Scholar] [CrossRef] [PubMed]
  57. Štifanić, D.; Musulin, J.; Jurilj, Z.; Šegota, S.; Lorencin, I.; Anđelić, N.; Vlahinić, S.; Šušteršič, T.; Blagojević, A.; Filipović, N.; et al. Semantic segmentation of chest X-ray images based on the severity of COVID-19 infected patients. EAI Endorsed Trans. Bioeng. Bioinform. 2021, 1, e3. [Google Scholar] [CrossRef]
  58. Wu, T.H.; Wang, T.W.; Liu, Y.Q. Real-time vehicle and distance detection based on improved yolo v5 network. In Proceedings of the 2021 3rd World Symposium on Artificial Intelligence (WSAI), Guangzhou, China, 18–20 June 2021; pp. 24–28. [Google Scholar]
  59. Yang, W.; Jiachun, Z. Real-time face detection based on YOLO. In Proceedings of the 2018 1st IEEE International Conference on Knowledge Innovation and Invention (ICKII), Jeju Island, Korea, 23–27 July 2018; pp. 221–224. [Google Scholar]
  60. Seo, J.; Sa, J.; Choi, Y.; Chung, Y.; Park, D.; Kim, H. A yolo-based separation of touching-pigs for smart pig farm applications. In Proceedings of the 2019 21st International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 17–20 February 2019; pp. 395–401. [Google Scholar]
  61. Francies, M.L.; Ata, M.M.; Mohamed, M.A. A robust multiclass 3D object recognition based on modern YOLO deep learning algorithms. Concurr. Comput. Pract. Exp. 2022, 34, e6517. [Google Scholar] [CrossRef]
  62. Zhang, P.; Su, W. Statistical inference on recall, precision and average precision under random selection. In Proceedings of the 2012 9th International Conference on Fuzzy Systems and Knowledge Discovery, Chongqing, China, 29–31 May 2012; pp. 1348–1352. [Google Scholar]
  63. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 1–13. [Google Scholar] [CrossRef] [Green Version]
  64. Charuchinda, P.; Kasetkasem, T. Land Cover Mapping Using the Class Activation Map. Ph.D. Thesis, Kasetsart University, Bangkok, Thailand, 2019. [Google Scholar]
  65. Carpenter, G.A.; Grossberg, S. A self-organizing neural network for supervised learning, recognition, and prediction. IEEE Commun. Mag. 1992, 30, 38–49. [Google Scholar] [CrossRef]
  66. Nguyen, Q.H.; Ly, H.B.; Ho, L.S.; Al-Ansari, N.; Le, H.V.; Tran, V.Q.; Prakash, I.; Pham, B.T. Influence of data splitting on performance of machine learning models in prediction of shear strength of soil. Math. Probl. Eng. 2021, 2021, 4832864. [Google Scholar] [CrossRef]
  67. Yin, Y.; Li, H.; Fu, W. Faster-YOLO: An accurate and faster object detection method. Digit. Signal Process. 2020, 102, 102756. [Google Scholar] [CrossRef]
  68. Chai, E.; Ta, L.; Ma, Z.; Zhi, M. ERF-YOLO: A YOLO algorithm compatible with fewer parameters and higher accuracy. Image Vis. Comput. 2021, 116, 104317. [Google Scholar] [CrossRef]
Figure 1. An example of the unprocessed dataset.
Figure 1. An example of the unprocessed dataset.
Applsci 12 11131 g001
Figure 2. Physical transformation actions performed on the dataset.
Figure 2. Physical transformation actions performed on the dataset.
Applsci 12 11131 g002
Figure 3. Geometric image augmentation for the expansion of the dataset.
Figure 3. Geometric image augmentation for the expansion of the dataset.
Applsci 12 11131 g003
Figure 4. Labeling boundary boxes of a single image using VoTT software.
Figure 4. Labeling boundary boxes of a single image using VoTT software.
Applsci 12 11131 g004
Figure 5. Labeling the dataset (a) correct way, (b) the incorrect way.
Figure 5. Labeling the dataset (a) correct way, (b) the incorrect way.
Applsci 12 11131 g005
Figure 6. An illustration of YOLOV4-based object detection.
Figure 6. An illustration of YOLOV4-based object detection.
Applsci 12 11131 g006
Figure 7. Diagram of the YOLOV4 algorithm applied for the detection of recyclable packaging.
Figure 7. Diagram of the YOLOV4 algorithm applied for the detection of recyclable packaging.
Applsci 12 11131 g007
Figure 8. Different types of IoU values.
Figure 8. Different types of IoU values.
Applsci 12 11131 g008
Figure 9. Results of YOLOV4 detection, for glass packaging using web cam.
Figure 9. Results of YOLOV4 detection, for glass packaging using web cam.
Applsci 12 11131 g009
Figure 10. Results of YOLOV4 detection, for plastic packaging using web cam.
Figure 10. Results of YOLOV4 detection, for plastic packaging using web cam.
Applsci 12 11131 g010
Figure 11. Results of YOLOV4 detection, for aluminum packaging using web cam.
Figure 11. Results of YOLOV4 detection, for aluminum packaging using web cam.
Applsci 12 11131 g011
Figure 12. Results of YOLOV4 detection, for mixed packaging using web cam.
Figure 12. Results of YOLOV4 detection, for mixed packaging using web cam.
Applsci 12 11131 g012
Table 1. Dataset used in training YOLOV4 algorithm.
Table 1. Dataset used in training YOLOV4 algorithm.
ClassType of PackagingNumber of Original ImagesNumber of Augmented ImagesNumber of Total Images
1Plastic10038691872
2Glass12246001824
3Aluminum6116941205
Table 2. Example of YOLOV4 annotations for Figure 5a TXT file.
Table 2. Example of YOLOV4 annotations for Figure 5a TXT file.
ClassX CenterY CenterWidthHeight
10.28550.83670.25930.2282
10.58440.64220.27650.2217
10.60080.36440.19490.2619
10.29770.30140.24950.2511
Table 3. Variation of YOLOV4 algorithm hyperparameters and the size of train/test dataset ratio.
Table 3. Variation of YOLOV4 algorithm hyperparameters and the size of train/test dataset ratio.
No.Max BatchesActivation FunctionTrain/Test RatioNo.Max BatchesActivation FunctionTrain/Test Ratio
16000Linear70/301910,000ReLU70/30
2ReLU20Mish
3Mish21ReLU75/25
4Linear75/2522Mish
5ReLU23ReLU80/20
6Mish24Mish
7Linear80/202520,000ReLU70/30
8ReLU26Mish
9Mish27ReLU75/25
108000Linear70/3028Mish
11ReLU29ReLU80/20
12Mish30Mish
13Linear75/25
14ReLU
15Mish
16Linear80/20
17ReLU
18Mish
Table 4. Overall overview of the results obtained through this research for first two cycles.
Table 4. Overall overview of the results obtained through this research for first two cycles.
Max Batches6000
Activation FunctionLinearReLuMish
Train/test ratio70/3075/2580/2070/3075/2580/2070/3075/2580/20
mAP60.18%59.85%59.11%93.12%94.17%94.16%94.25%93.25%92.26%
F1 score 0.520.480.470.870.850.810.860.840.83
Average IoU29.98%28.82%28.14%62.18%60.49%58.21%61.52%60.77%59.98%
Precision0.440.410.400.840.790.750.810.780.76
Recall0.640.590.580.920.920.900.930.920.92
Average Loss2.61782.48732.42851.18881.18971.26891.17701.22031.2406
Max Batches8000
Activation FunctionLinearReLuMish
Train/test ratio70/3075/2580/2070/3075/2580/2070/3075/2580/20
mAP94.51%93.17%91.27%99.77%96.17%95.93%99.71%99.70%99.65%
F1 score 0.590.700.760.980.970.970.980.970.98
Average IoU60.11%57.15%52.27%80.75%78.12%77.63%79.80%81.24%81.25%
Precision0.530.640.720.970.970.960.980.970.97
Recall0.690.790.800.990.980.990.990.990.99
Average Loss1.84111.78282.03890.90960.86600.92150.94430.90910.9145
Table 5. Overall overview of the results obtained through this research for the third and fourth training cycle.
Table 5. Overall overview of the results obtained through this research for the third and fourth training cycle.
Max Batches10,000
Activation FunctionLinearReLuMish
Train/test ratio70/3075/2580/2070/3075/2580/2070/3075/2580/20
mAP///99.90%99.90%99.77%99.84%99.83%99.81%
F1 score ///0.990.990.980.990.990.99
Average IoU///84.14%84.59%83.41%83.44%83.59%83.33%
Precision///0.980.990.980.990.990.98
Recall///1.001.001.001.000.991.00
Average Loss///0.63200.69510.68620.69450.68090.6500
Max Batches20,000
Activation FunctionLinearReLuMish
Train/test ratio70/3075/2580/2070/3075/2580/2070/3075/2580/20
mAP///88.33%88.60%89.08%89.92%99.96%99.94%
F1 score ///0.860.860.870.871.001.00
Average IoU///84.41%86.68%83.90%82.73%91.13%91.06%
Precision///0.950.960.940.931.001.00
Recall///0.750.790.800.811.001.00
Average Loss///0.92190.84960.79140.85980.36430.3400
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Glučina, M.; Baressi Šegota, S.; Anđelić, N.; Car, Z. Automated Detection and Classification of Returnable Packaging Based on YOLOV4 Algorithm. Appl. Sci. 2022, 12, 11131. https://doi.org/10.3390/app122111131

AMA Style

Glučina M, Baressi Šegota S, Anđelić N, Car Z. Automated Detection and Classification of Returnable Packaging Based on YOLOV4 Algorithm. Applied Sciences. 2022; 12(21):11131. https://doi.org/10.3390/app122111131

Chicago/Turabian Style

Glučina, Matko, Sandi Baressi Šegota, Nikola Anđelić, and Zlatan Car. 2022. "Automated Detection and Classification of Returnable Packaging Based on YOLOV4 Algorithm" Applied Sciences 12, no. 21: 11131. https://doi.org/10.3390/app122111131

APA Style

Glučina, M., Baressi Šegota, S., Anđelić, N., & Car, Z. (2022). Automated Detection and Classification of Returnable Packaging Based on YOLOV4 Algorithm. Applied Sciences, 12(21), 11131. https://doi.org/10.3390/app122111131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop