Next Article in Journal
Cost-Effective Vibration Analysis through Data-Backed Pipeline Optimisation
Next Article in Special Issue
Human–Machine Differentiation in Speed and Separation Monitoring for Improved Efficiency in Human–Robot Collaboration
Previous Article in Journal
Multi-Domain Convolutional Neural Networks for Lower-Limb Motor Imagery Using Dry vs. Wet Electrodes
Previous Article in Special Issue
Improved Mutual Understanding for Human-Robot Collaboration: Combining Human-Aware Motion Planning with Haptic Feedback Devices for Communicating Planned Trajectory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots

1
Daegu-Gyeongbuk Research Center, Electronics and Telecommunications Research Institute (ETRI), Daegu 42994, Korea
2
Division of Electronic Engineering, Jeonbuk National University, 567 Baekje-daero, Deokjin-gu, Jeonju 54896, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(19), 6674; https://doi.org/10.3390/s21196674
Submission received: 22 August 2021 / Revised: 3 October 2021 / Accepted: 5 October 2021 / Published: 8 October 2021
(This article belongs to the Special Issue Human-Robot Collaborations in Industrial Automation)

Abstract

:
Human-robot interaction has received a lot of attention as collaborative robots became widely utilized in many industrial fields. Among techniques for human-robot interaction, collision identification is an indispensable element in collaborative robots to prevent fatal accidents. This paper proposes a deep learning method for identifying external collisions in 6-DoF articulated robots. The proposed method expands the idea of CollisionNet, which was previously proposed for collision detection, to identify the locations of external forces. The key contribution of this paper is uncertainty-aware knowledge distillation for improving the accuracy of a deep neural network. Sample-level uncertainties are estimated from a teacher network, and larger penalties are imposed for uncertain samples during the training of a student network. Experiments demonstrate that the proposed method is effective for improving the performance of collision identification.

1. Introduction

With the increasing demands of collaborative tasks between humans and robots, the research on human–robot interaction has received great attention from researchers and engineers in the field of robotics [1]. Robots that can collaborate with humans are called collaborative robots (cobots), and cobots differ from conventional industrial robots in that they do not require a fence to prevent access. Previously, the application of robots is limited to performing simple and repetitive tasks in well-structured and standardized environments such as factories and warehouses. However, the development of sensing and control technologies has significantly expanded the area of application of cobots [2], and they are beginning to be applied to several tasks around us. More specifically, their applications have been diversified from traditional automated manufacturing and logistics industries to more general tasks such as medical [3], service [4,5], food and beverage industries [6], and these tasks require more elaborate sensing and complicated control techniques. Furthermore, with the development of intelligent algorithms including intention estimation [7] and gesture recognition [8], cobots can be utilized in wider application areas.
In general, robots have advantages over humans in repetitive tasks, and humans are better at making comprehensive decisions and judgments. Therefore, human–robot collaboration possibly increases the efficiency of intelligent systems through complementary synergies. As the scope of robotics applications gradually expands through collaborative work, interaction with humans or unstructured environments has become an important technical issue, which requires the implementation of advanced perception and control algorithms. Especially, collision detection and identification techniques are indispensable elements to improve the safety and reliability of collaborative robots [9,10].
To perform cooperative tasks with the aid of human–robot interactions, several studies have been carried out to detect and identify robot collisions for the safety of workers [11]. Previous work can be categorized into two approaches: the first category is the study on the control of collaborative robots by predicting possible collisions and the other is the study of responses after impacts. While collision avoidance is more advantageous in terms of safety [12], this approach inevitably requires additional camera sensors for action recognition of coworkers or 3D reconstruction of surrounding environments [13]. Furthermore, it is difficult to completely avoid abrupt and unpredictable collisions. Therefore, techniques for collision identification are essential to improve the safety and reliability of collaborative robots.
Collision detection algorithms investigate external forces [14] or currents [15] to determine whether a true collision has occurred on an articulated robot. A key element in the procedure of collision detection is the estimatation of external torques. A major approach to estimating external torques is utilizing torque sensor signals to compute internal joint torques based on the physical dynamics of robots, and several other methods to construct momentum observers to estimate external torques without the use of torque sensors. The method that does not use torque sensors is called sensorless external force estimation, and an elaborate modeling of the observer and filter is essential for the precise estimation of external forces [16,17,18,19]. External forces are further processed by a thresholding method [20] or classification algorithm [21], to determine whether a collision has occurred. Recently, deep-learning-based methods have outperformed traditional model-based methods in detecting collisions [22]. Beyond collision detection, the identification of collision locations is beneficial for the construction of more reliable collaborate robots, by making them react appropriately in collision situations.
To ensure the proper responses of collaborative robots in cases of collisions, it is necessary to identify collision locations. The collision identification technique can be defined as a multiclass classification of time series sensor data according to collision locations. In early studies, collision identification was mainly based on the elaborate modeling of filters [23] and observers [24], and a frequency domain analysis was conducted to improve the accuracy of collision identification [25]. To address the classification problem, machine learning techniques, which were employed to analyze time series data, have also been applied to collision identification [26]. Recently, support vector machines [27] and probabilistic methods [28] were applied to improve the reliability of collision identification systems. In [29], the collision identification performance was improved by utilizing additional, sensors such as inertial measurement units, and analyzing their vibration features.
In this paper, we propose a method that can identify collisions on articulated robots by utilizing deep neural networks for joint sensor signals. Collision identification refers to a technique that not only detects the occurrence of a collision, but also determines its location. Recently, a collision detection method was proposed by Heo et al. [22]; we extend this existing method for collision identification and improve the robustness of the deep neural network. To improve the performance of the collision identification system, we construct a deeper network, which is called a teacher network, to distill its probabilistic knowledge to a student network. In the process of distilling knowledge, we employ the uncertainties of the teacher network to focus on learning difficult examples, mostly collision samples. This paper is organized as follows. Section 2 presents related work, Section 3 explains collision modeling and data collection, and Section 4 presents the proposed method. Section 5 and Section 6 presents the experimental results and conclusion, respectively.

2. Related Work

2.1. Deep Learning Methods for Collision Identification of Collaborative Robots

Collision detection is a key technology to ensure the safety and reliability of collaborative robots. Although most previous methods were based on the mathematical modeling of robots [30,31,32], recently, deep learning methods have shown promising results for this goal. Min et al. [33] estimated vibration features based on the physical modeling of robots and utilized neural networks for collision identification. Xu et al. [34] combined neural networks and nonlinear disturbance observer for collision detection. Park et al. [35] combined a convolutional neural network and support vector machine to detect collisions, and Heo et al. [22] employed causal convolutions, which were previously utilized for auto-regressive models in WaveNet [36] to detect collisions based on joint sensor signals including torque, position, and velocity. Maceira et al. [37] employed recurrent neural networks to infer the intentions of external forces in collaborative tasks, and Czubenko et al. [38] proposed an MC-LSTM, which combines convolutions and recurrent layers for collision detection. Mohammadi et al. [13] utilized external vision sensors to further recognize human actions and collisions.

2.2. Knowledge Distillation

Knowledge distillation was proposed by Hinton et al. [39] to train a student network with the aid of a deeper network, which is called a teacher network. Probabilistic responses of the teacher network are beneficial to improve the accuracy of the student network because the probabilities of false categories were also utilized during knowledge distillation. Although most early methods directly distill the logits of a teacher network, Park et al. [40] utilized the logits’ relations, and Meng et al. [41] proposed a conditional teacher–student learning framework. Furthermore, knowledge from intermediate feature maps was distilled for network minimization [42] and performance improvement [43,44]. Knowledge distillation has been employed in various applications such as object detection [45], semantic segmentation [46], domain adaptation [47], and defense for adversarial examples [48]. Recently, the teacher–student learning framework has been applied with other advanced learning methodologies such as adversarial learning [49] and semi-supervised learning [50].

2.3. Uncertainty Estimation

Uncertainty plays an important role in interpreting the reliability of machine learning models and their predictions. Probabilistic approaches and Bayesian methods have been regarded as useful mathematical tools to quantify predictive uncertainties [51]. Recently, Gal and Ghahramani proposed Monte Carlo dropout (MC-dropout) [52], which can be interpreted as an approximate Bayesian inference of deep Gaussian processes, by utilizing dropout [53] at test time. Lakshminarayanan et al. [54] proposed deep ensembles for the better quantification of uncertainties, and Amersfoort et al. [55] proposed deterministic uncertainty quantification, which is based on a single model to address the problem of computational cost of MC-dropout and deep ensembles. Uncertainties have been utilized to quantify network confidences [56], selecting out-of-distribution samples [57], and improving the performance of deep neural networks [58,59], in various application areas such as medical image analysis [60] and autonomous driving [61].

3. Collision Modeling and Data Collection

3.1. Mathematical Modelling of Collisions

This section explains the mathematical modeling of dynamic equations for 6 Degrees of Freedom (DoF) articulated robots. In order to operate a robot through a desired trajectory and move it to a target position, precise control torque is required for each joint motor, and the control torque can be represented as the following dynamic equation:
τ = M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + g ( q ) ,
where τ R n is the control torque, M ( q ) R n × n is the inertia matrix of the articulated robot, C ( q , q ˙ ) R n × n is the matrix of Coriolis and Centrifugal torques, g ( q ) R n is the vector of gravitational torques, and q, q ˙ , q ¨ are the angular position, velocity, and acceleration of each joint, respectively. The dynamic equation can be obtained through the Newton–Euler method or the Euler–Lagrange equation using the mechanical and physical information of the robot. Since the dynamic equation of the robot is given as (1), in the absence of external force, external torques can be computed by subtracting the control torques from measured torques.
When a joint torque sensor is installed onto each joint, the torque generated on each joint due to external force is given as follows:
τ e x t = τ s τ ,
where τ e x t is the external torques generated onto each joint due to a collision, and τ s is torque values measured by joint torque sensors. The external torque can be precisely estimated under an accurate estimation of robot dynamics and physical parameters of the articulated robot such as the mass and center of mass of each link.
In robots that are not equipped with a joint torque sensor, sensorless methods are utilized to estimate external torques. Sensorless methods are basically based on the current signal of each joint motor, and an additional state variable p = M ( q ) q ˙ is defined to reformulate the dynamic equation as follows:
p ˙ = C ( q , q ˙ ) q ˙ g ( q ) f ( q , q ˙ ) + τ m ,
where f ( q , q ˙ ) is the friction matrix, and τ m is the motor torque. In the case of the sensorless method, it is necessary to obtain the transmitted torque from the motor to the link to estimate the collision torque. Therefore, the friction must additionally be considered in the existing robot dynamics equation. A main issue in sensorless external torque estimation is the elaborate design of observer and filter under the dynamic Equation (3), and the effect of disturbance can be reduced using momentum state variables. Due to the effect of noise and nonlinear frictional force, sensorless methods are generally less precise in the estimation of external torques compared to methods that utilize joint torque sensors. Through the methods mentioned above, it is possible to obtain the torques generated in each joint due to the collision of the robot. Then, the collision identification algorithm can determine collision locations from joint torques obtained through sensor or sensorless methods.

3.2. Data Collection and Labeling

Figure 1a presents the 6-DoF articulated robot to collect sensor data, which include the information of joint torque, current, angular position, and angular velocity. The Denavit–Hartenberg parameters of the articulated robot are presented in [62]. From the 6-DoF articulated robot, joint sensor signals were obtained with the sampling rate of 1 kHz, and a data sample collected at time t can be expressed as
x t = [ τ t , i t , θ t , w t ] R 24 ,
where τ t , i t , θ t , w t are six-dimensional vectors corresponding to torque, current, angular position, and angular velocity, respectively; the i-th components of these vectors indicate the sensor signals obtained at the i-th joint. Figure 1b shows the definition of collision categories according to collision locations. Collisions were generated at six locations, and in the case of no collision, which refers to the normal state, a label of 0 was assigned. In the case of a collision, a categorical label corresponding to the location was assigned to generate ground truth data.
Joint sensor data were collected, along with collision time and category, by applying intentional collisions at different locations. The collision time and category were converted into ground truth data which have an identical length to the corresponding sensor signals, as shown in Figure 2. For a collision occurrence, the corresponding category was assigned to 0.2 s of data samples from the collision time; each collision is represented as 200 collision samples in the ground truth data. We collected joint sensor signals for 5586 intentional collisions along with their ground truth data; the number of collisions, which were applied to different locations, is equal. This dataset was divided into training, validation, and test sets with the ratio of 70%, 10%, and 20%, as presented in Table 1.

4. Proposed Method

This section presents the proposed method for the collision identification of articulated robots. Firstly, two neural network architectures are presented; one of them is a student network and the other architecture is a teacher network for knowledge distillation. The second part explains the proposed knowledge distillation method, which considers the predictive uncertainties of the teacher network. Lastly, a post-processing is utilized to improve the robustness of the proposed algorithm by reducing noise in network predictions.

4.1. Network Architectures

This paper employs the network architecture presented by Heo et al. [22] as a base network model. Heo et al. [22] proposed a deep neural network, called CollisionNet, to detect collisions in articulated robots. Its architecture is composed of causal convolutions to reduce detection delay and dilated convolutions to achieve large receptive fields. We modeled the base network by modifying the last fully connected layer in CollisionNet to conduct multiclass classification and identify collision locations. The base network is composed of seven convolution layers and three fully connected layers, and its details are identical to CollisionNet except the last layer; convolution filters with the size of 3 are utilized for both regular and dilated convolutions, the depth of the intermediate features is increased from 128 to 512, and the dilation ratio is increased by a factor of two. The architecture of the base network is identically utilized as a student network in the process of knowledge distillation.
Figure 3 shows the architecture of the teacher network. To construct the teacher network, three regular convolutions in the base network are replaced into convolution blocks. A convolution block contains four convolution layers with a skip connection, and therefore, the number of parametric layers in the teacher network increases to 19. The number of channels in the second and third convolution layers in a convolution block are identical to the number of output channels of the corresponding regular convolution layers. The number of trainable parameters in the teacher network is 6.63 million; therefore, it has more capacity to learn the training data compared to the base network, which has 2.79 million parameters. Dropout layers with a dropout ratio of 0.5 are added to the fully connected layers in the teacher network, and Monte Carlo samples from the teacher network are acquired by applying dropout at the test time.

4.2. Uncertainty-Aware Knowledge Distillation

The teacher network is trained with the cross-entropy loss between the softmax prediction y ^ T and its one hot encoded label y . The i-th component of y ^ T indicates the predicted probability that the input sample belongs to the i-th category. In our case, seven categories exist, which contain the normal state and six possible collision locations. The loss function for the training of the teacher network is defined as
l c e ( y , y ^ T ) = i y i log ( y ^ T , i ) ,
where y i and y ^ T , i are the i-th components of y and y ^ T , respectively.
After training the teacher network, K logits, z ^ T 1 , ⋯, z ^ T K are obtained from an input sample by utilizing MC-dropout [52]. These logits are computed by randomly ignoring 50% of neurons in the fully connected layers in the teacher network. Based on the K logits of the teacher network, the i-th component of the uncertainty vector is computed by
u i = 1 K k ( z ^ T , i k z ¯ T , i ) 2 ,
where z ¯ T , i is the i-th component of the averaged logit z ¯ T , which is computed by
z ¯ T = 1 K k z ^ T k .
The uncertainty u i is the variance of logits; therefore, the value of the uncertainty increases as distances between the logits increase.
The total loss L for the training of the student network is composed of two loss functions, as follows:
L = l c e ( y , y ^ S ) + l k d ( z ¯ T , z ^ S , u ) ,
where l c e ( y , y ^ S ) is the cross-entropy loss between the softmax prediction of the student network and its corresponding label, u is the uncertainty vector whose i-th component is u i , and l k d ( z ¯ T , z ^ S , u ) is the uncertainty-aware knowledge distillation loss. The knowledge distillation loss os obtained by computing uncertainty-weighted Kullback–Leibler divergence (KL divergence) between σ ( z ^ S , T ) and σ ( z ¯ T , T ) , as follows:
l k d ( z ¯ T , z ^ S , u ) = i u i σ ( z ¯ T , T ) i { log ( σ ( z ^ S , T ) i ) log ( σ ( z ¯ T , T ) i ) } ,
where σ ( z , T ) is the softmax function with the temperature T, and σ ( z , T ) i is the i-th component of σ ( z , T ) . In (9), σ ( z , T ) i can be computed as
σ ( z , T ) i = exp ( z i / T ) j exp ( z j / T ) .
The overall procedure for the training of the student network is presented in Figure 4.

4.3. Post-Processing

The post-processing to reduce errors in network predictions is inspired by a connected component analysis in image-processing techniques. In the labeled data, a collision is represented by connected samples, with a non-zero number corresponding to its location. However, a few predictions may differ from their adjacent predictions, because a neural network independently infers predictions for different data samples. Based on the collision properties in the labeled data, incorrect predictions are reduced by the post-processing presented in Figure 5.
The post-processing is composed of two steps; in Figure 5, (a) shows predictions from the student network, and (b) and (c) present the results after the first and second post-processing steps, respectively. In the first step, non-zero connected samples are grouped, and the number of samples for each category are counted. Predictions in a group are replaced into the category which corresponds to the maximum frequency, as presented in Figure 5b. In the second step, if the number of non-zero connected samples is less than a threshold value, then these samples are regarded as the normal state. The threshold value of 10 samples is utilized in experiments, and Figure 6 presents examples of the results of the post-processing.

5. Experiments

5.1. Experimental Environment and Evaluation Measures

The proposed algorithm is developed within a hardware environment including Intel core i7-10700 CPU, 32GB DDR4 RAM, and RTX 3080 GPU. In experiments, Python and Pytorch are mainly utilized to implement the proposed algorithm and to conduct an ablation study. To demonstrate the proposed method, the dataset is gathered from a collaborative robot, which consists of six rotating joints. The cobot weighs 47 kg, has a maximum payload of 10 kg, and reaches up to 1300 mm. The actuator consists of motors manufactured from Parker, motor drivers from Welcon, and embedded joint torque sensors in each joint. The hardware of the cobot contains a custom embedded controller, based on real-time linux kernel, and it communicates with drivers through EtherCAT with a cycle time of 1 ms.
To demonstrate the effectiveness of the proposed method, we evaluate the algorithm in three ways: (1) sample-level accuracy, (2) collision-level accuracy, and (3) time delay. In the process of collision identification, deep neural networks perform sample-level multiclass classification, which classifies each sample, composed of a 24-dimensional sequence of sensor data, into the normal state or one of six collision locations. To evaluate the sample-level accuracy of deep neural networks, we measure R e c a l l , P r e c i s i o n , and F 1 - s c o r e for each sample, which are defined as follows:
R e c a l l = T P / ( T P + F N ) , P r e c i s i o n = T P / ( T P + F P ) , F 1 - s c o r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l ,
where T P , F P , F N are the numbers of true positives, false positives, and false negatives, respectively. True positive is a correctly identified collision sample, false positive is an incorrect prediction, which is classified into a collision, and false negative is an incorrect prediction which is classified into the normal state.
Collision-level accuracy is another important measure for evaluating a collision identification system. Because collaborative robots respond to each collision, reducing the number of false positive collisions is an important issue. R e c a l l , P r e c i s i o n , and F 1 - s c o r e are computed as (11) with different definitions of T P , F P , and F N to measure the collision-level accuracy. A group of connected samples that are classified into a collision is regarded as a true positive if the intersection over union (IoU) between the connected predictions and its corresponding true collision samples is greater than 0.5. A group of predictions that are classified into a false category of collisions is regarded as a false positive, and a false negative is a missed collision. Figure 7 shows several cases of T P , F P , and F N for measuring the collision-level accuracy.
Finally, the time delay is measured to evaluate the processing time of the collision identification system. For the safe and reliable collaborations of human and robots, the processing time is required to be reduced as possible. The total processing time is composed of the inference time of a neural network, detection delay for collisions, and post-processing time. Based on these three types of evaluation measure, the effectiveness of the proposed method is demonstrated in experiments.

5.2. Training of Neural Networks

To train the neural networks, Adam optimizer [63] is utilized with a learning rate of 10 4 . The learning rate is decreased to 10 5 after training 200 epochs. Figure 8 presents f1-scores for the training and validation datasets during the training of 500 epochs. As shown in Figure 8, after training a sufficientl number of epochs, the validation accuracy was not further decreased. Therefore, in the following experiments, the accuracies of deep neural networks are evaluated for the test set after training 300 epochs.
To train the student network, the temperature of the softmax function is set to 5 during the process of knowledge distillation. The temperature value has to be greater than 1 to soften probabilistic predictions of neural network, and temperature values between 2 and 5 are usually used for knowledge distillation in the previous literature [39]. In our experiments, modifications to the temperature value glead to insignificant changes in the experimental results. In Figure 9, (a) shows the first dimension of 24-dimensional sensor data, which corresponds to the torque signal at the first joint, and (b) presents uncertainties measured by MC-dropout with the value of K = 4 . As shown in Figure 9, the uncertainties of collision samples are high compared to normal state samples. By weighting the uncertainties on the KL-divergence between probabilistic predictions of the student and teacher network, the student network is able to focus on learning difficult data samples.

5.3. Sample-Level Accuracy

The first measure to evaluate the performance of deep neural networks is the sample-level accuracy. As explained in Section 4.1, the architecture of the deep neural network proposed in [22] is employed to construct the base model. To demonstrate the effectiveness of uncertainty-aware knowledge distillation for the problem of collision identification, we compare the accuracies of the proposed method with those of the base model and a student network. The student network has an identical architecture to the base model, and is trained by distilling knowledge in the teacher network without employing uncertainty information. Table 2 presents the sample-level recall, precision, and f1-score of four neural network models; the proposed method means another student network, which is trained by uncertainty-aware knowledge distillation. The last row of Table 2 presents the sample-level accuracies of the teacher network. As presented in Table 2, the f1-scores of the proposed method are comparable to those of the teacher network; it is worth noting that the proposed method employs a lightweight network compared to the teacher network.

5.4. Collision-Level Accuracy

This section presents the collision-level accuracies. As collaborative robots react to each collision, reducing the number of false-positive collisions is a very important issue in reliable collision identification systems. In the labeled data, a collision is represented by 200 non-zero samples; therefore, false-positive collisions, which are composed of a few fals- positive samples, are not effectively reflected in the sample-level accuracies. Although the sample-level accuracies of the four neural network models are above 98%, there are a considerable number of false-positive collisions. To compute the collision-level accuracies, a group of non-zero predictions is regarded as a collision, and Table 3 presents the numbers of true-positive, false-positive, and false-negative collisions of the four neural network models. In Table 3, the base model, student network, and proposed method have an identical network architecture to CollisionNet [22]; the student network is trained by regular knowledge distillation, and the proposed method employs uncertainties during knowledge distillation. As shown in Table 3, the number of false positives is significantly reduced after the post-processing. Table 4 presents the collision-level recall, precision, and f1-score of the four neural networks. By utilizing probabilistic labels and uncertainties from the teacher network, the proposed method produces better accuracies, despite its lightweight network architecture compared to the teacher network.

5.5. Analysis for the Processing Time

The processing time is another important factor for responding to external forces within an acceptable timeframe. In the collision identification system, the total processing time is composed of the inference time of a neural network, time delay for detecting a collision, and post-processing time. Table 5 presents the averaged processing time for each step. The teacher network requires an 83% longer inference time compared to the base model, student network, and proposed method. The detection delay is measured by averaging the time intervals between collision occurrences and their corresponding first true-positive samples. As presented in Table 5, the proposed method requires 2.6350 ms to identify a collision occurrence, and this satisfies the requirement for the safety of collaborative robots.

6. Conclusions

This paper proposes a collision identification method for collaborative robots. To identify the locations of external forces, the propose method employs a deep neural network, which is composed of causal convolutions and dilated convolutions. The key contribution is the method of capturing sample-level uncertainties and distilling the knowledge of a teacher network to train a student network, with consideration of predictive uncertainties. In the knowledge distillation, KL-divergence between the predictions of the student and teacher networks are weighted by the predictive uncertainties to focus on data samples that are difficult to train. Furthermore, we also propose a post-processing to reduce the number of false-positive collisions. Experiments were conducted with a 6-DoF-articulated robot, and we demonstrated that the uncertainty is beneficial to improving the accuracy of the collision identification method.

Author Contributions

Conceptualization, W.K. and S.J.L.; methodology, S.J.L.; software, S.J.L.; validation, S.J.L.; formal analysis, S.J.L.; data curation, W.K.; writing—original draft preparation, W.K. and S.J.L.; writing—review and editing, W.K., Y.J. and S.J.L.; visualization, S.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1G1A1009792). This research was supported by “Research Base Construction Fund Support Program” funded by Jeonbuk National University in 2021, and partially supported by Electronics and Telecommunications Research Institute (ETRI) grant funded by the Korean government. (21ZD1130, Development of ICT Convergence Technology for Daegu-Gyeongbuk Regional Industry).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goodrich, M.A.; Schultz, A.C. Human-Robot Interaction: A Survey; Now Publishers Inc.: Boston, MA, USA, 2008. [Google Scholar]
  2. Ajoudani, A.; Zanchettin, A.M.; Ivaldi, S.; Albu-Schäffer, A.; Kosuge, K.; Khatib, O. Progress and prospects of the human–robot collaboration. Auton. Rob. 2018, 42, 957–975. [Google Scholar] [CrossRef] [Green Version]
  3. Haidegger, T. Autonomy for surgical robots: Concepts and paradigms. IEEE Trans. Med. Rob. Bionics 2019, 1, 65–76. [Google Scholar] [CrossRef]
  4. Berezina, K.; Ciftci, O.; Cobanoglu, C. Robots, Artificial Intelligence, and Service Automation in Restaurants; Emerald Group Publishing: Bingley, UK, 2020. [Google Scholar]
  5. Wilson, G.; Pereyda, C.; Raghunath, N.; de la Cruz, G.; Goel, S.; Nesaei, S.; Minor, B.; Schmitter-Edgecombe, M.; Taylor, M.E.; Cook, D.J. Robot-enabled support of daily activities in smart home environments. Cognitive Syst. Res. 2019, 54, 258–272. [Google Scholar] [CrossRef] [PubMed]
  6. Iqbal, J.; Khan, Z.H.; Khalid, A. Prospects of robotics in food industry. Food Sci. Technol. 2017, 37, 159–165. [Google Scholar] [CrossRef] [Green Version]
  7. Petković, T.; Puljiz, D.; Marković, I.; Hein, B. Human intention estimation based on hidden Markov model motion validation for safe flexible robotized warehouses. Rob. Comput. Integr. Manuf. 2019, 57, 182–196. [Google Scholar] [CrossRef] [Green Version]
  8. Oudah, M.; Al-Naji, A.; Chahl, J. Hand gesture recognition based on computer vision: A review of techniques. J. Imaging 2020, 6, 73. [Google Scholar] [CrossRef] [PubMed]
  9. Vicentini, F. Collaborative robotics: A survey. J. Mech. Des. 2021, 143, 040802. [Google Scholar] [CrossRef]
  10. Zhang, S.; Wang, S.; Jing, F.; Tan, M. A sensorless hand guiding scheme based on model identification and control for industrial robot. IEEE Trans. Ind. Inf. 2019, 15, 5204–5213. [Google Scholar] [CrossRef]
  11. Haddadin, S.; De Luca, A.; Albu-Schäffer, A. Robot collisions: A survey on detection, isolation, and identification. IEEE Trans. Rob. 2017, 33, 1292–1312. [Google Scholar] [CrossRef] [Green Version]
  12. Morikawa, S.; Senoo, T.; Namiki, A.; Ishikawa, M. Realtime collision avoidance using a robot manipulator with light-weight small high-speed vision systems. In Proceedings of the 2007 IEEE International Conference on Robotics and Automation (ICRA), Roma, Italy, 10–14 April 2007; pp. 794–799. [Google Scholar]
  13. Mohammadi Amin, F.; Rezayati, M.; van de Venn, H.W.; Karimpour, H. A mixed-perception approach for safe human–robot collaboration in industrial automation. Sensors 2020, 20, 6347. [Google Scholar] [CrossRef]
  14. Lu, S.; Chung, J.H.; Velinsky, S.A. Human-robot collision detection and identification based on wrist and base force/torque sensors. In Proceedings of the 2005 IEEE international Conference on Robotics and Automation (ICRA), Barcelona, Spain, 18–22 April 2005; pp. 3796–3801. [Google Scholar]
  15. Lee, S.D.; Kim, M.C.; Song, J.B. Sensorless collision detection for safe human-robot collaboration. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 2392–2397. [Google Scholar]
  16. Han, L.; Xu, W.; Li, B.; Kang, P. Collision detection and coordinated compliance control for a dual-arm robot without force/torque sensing based on momentum observer. IEEE/ASME Trans. Mechatron. 2019, 24, 2261–2272. [Google Scholar] [CrossRef]
  17. De Luca, A.; Albu-Schaffer, A.; Haddadin, S.; Hirzinger, G. Collision detection and safe reaction with the DLR-III lightweight manipulator arm. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Beijing, China, 9–15 October 2006; pp. 1623–1630. [Google Scholar]
  18. Briquet-Kerestedjian, N.; Makarov, M.; Grossard, M.; Rodriguez-Ayerbe, P. Generalized momentum based-observer for robot impact detection—Insights and guidelines under characterized uncertainties. In Proceedings of the 2017 IEEE Conference on Control Technology and Applications (CCTA), Maui, HI, USA, 27–30 August 2017; pp. 1282–1287. [Google Scholar]
  19. Mamedov, S.; Mikhel, S. Practical aspects of model-based collision detection. Front. Rob. AI 2020, 7, 162. [Google Scholar]
  20. Geravand, M.; Flacco, F.; De Luca, A. Human-robot physical interaction and collaboration using an industrial robot with a closed control architecture. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013; pp. 4000–4007. [Google Scholar]
  21. Sharkawy, A.N.; Koustoumpardis, P.N.; Aspragathos, N. Neural network design for manipulator collision detection based only on the joint position sensors. Robotica 2020, 38, 1737–1755. [Google Scholar] [CrossRef]
  22. Heo, Y.J.; Kim, D.; Lee, W.; Kim, H.; Park, J.; Chung, W.K. Collision detection for industrial collaborative robots: A deep learning approach. IEEE Rob. Autom. Lett. 2019, 4, 740–746. [Google Scholar] [CrossRef]
  23. Hu, J.; Xiong, R. Contact force estimation for robot manipulator using semiparametric model and disturbance Kalman filter. IEEE Trans. Ind. Electron. 2017, 65, 3365–3375. [Google Scholar] [CrossRef]
  24. Ren, T.; Dong, Y.; Wu, D.; Chen, K. Collision detection and identification for robot manipulators based on extended state observer. Control Eng. Pract. 2018, 79, 144–153. [Google Scholar] [CrossRef]
  25. Kouris, A.; Dimeas, F.; Aspragathos, N. A frequency domain approach for contact type distinction in human–robot collaboration. IEEE Rob. Autom. Lett. 2018, 3, 720–727. [Google Scholar] [CrossRef]
  26. Jo, S.; Kwon, W. A Comparative Study on Collision Detection Algorithms based on Joint Torque Sensor using Machine Learning. J. Korea Robot. Soc. 2020, 15, 169–176. [Google Scholar] [CrossRef]
  27. Pan, J.; Manocha, D. Efficient configuration space construction and optimization for motion planning. Engineering 2015, 1, 46–57. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, Z.; Qian, K.; Schuller, B.W.; Wollherr, D. An online robot collision detection and identification scheme by supervised learning and bayesian decision theory. IEEE Trans. Autom. Sci. Eng. 2020, 18, 1144–1156. [Google Scholar] [CrossRef]
  29. Birjandi, S.A.B.; Kühn, J.; Haddadin, S. Observer-extended direct method for collision monitoring in robot manipulators using proprioception and imu sensing. IEEE Rob. Autom. Lett. 2020, 5, 954–961. [Google Scholar] [CrossRef]
  30. Caldas, A.; Makarov, M.; Grossard, M.; Rodriguez-Ayerbe, P.; Dumur, D. Adaptive residual filtering for safe human-robot collision detection under modeling uncertainties. In Proceedings of the 2013 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, Wollongong, Australia, 9–12 July 2013; pp. 722–727. [Google Scholar]
  31. Makarov, M.; Caldas, A.; Grossard, M.; Rodriguez-Ayerbe, P.; Dumur, D. Adaptive filtering for robust proprioceptive robot impact detection under model uncertainties. IEEE/ASME Trans. Mechatron. 2014, 19, 1917–1928. [Google Scholar] [CrossRef]
  32. Birjandi, S.A.B.; Haddadin, S. Model-Adaptive High-Speed Collision Detection for Serial-Chain Robot Manipulators. IEEE Rob. Autom. Lett. 2020, 5, 6544–6551. [Google Scholar] [CrossRef]
  33. Min, F.; Wang, G.; Liu, N. Collision detection and identification on robot manipulators based on vibration analysis. Sensors 2019, 19, 1080. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Xu, T.; Fan, J.; Fang, Q.; Zhu, Y.; Zhao, J. A new robot collision detection method: A modified nonlinear disturbance observer based-on neural networks. J. Intell. Fuzzy Syst. 2020, 38, 175–186. [Google Scholar] [CrossRef]
  35. Park, K.M.; Kim, J.; Park, J.; Park, F.C. Learning-based real-time detection of robot collisions without joint torque sensors. IEEE Rob. Autom. Lett. 2020, 6, 103–110. [Google Scholar] [CrossRef]
  36. Oord, A.V.D.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  37. Maceira, M.; Olivares-Alarcos, A.; Alenya, G. Recurrent neural networks for inferring intentions in shared tasks for industrial collaborative robots. In Proceedings of the 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 665–670. [Google Scholar]
  38. Czubenko, M.; Kowalczuk, Z. A Simple Neural Network for Collision Detection of Collaborative Robots. Sensors 2021, 21, 4235. [Google Scholar] [CrossRef]
  39. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  40. Park, W.; Kim, D.; Lu, Y.; Cho, M. Relational knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3967–3976. [Google Scholar]
  41. Meng, Z.; Li, J.; Zhao, Y.; Gong, Y. Conditional teacher-student learning. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6445–6449. [Google Scholar]
  42. Yim, J.; Joo, D.; Bae, J.; Kim, J. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4133–4141. [Google Scholar]
  43. Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. Fitnets: Hints for thin deep nets. arXiv 2014, arXiv:1412.6550. [Google Scholar]
  44. Zagoruyko, S.; Komodakis, N. Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv 2016, arXiv:1612.03928. [Google Scholar]
  45. Yim, J.; Joo, D.; Bae, J.; Kim, J. Learning efficient object detection models with knowledge distillation. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–7 December 2017; pp. 742–751. [Google Scholar]
  46. Hou, Y.; Ma, Z.; Liu, C.; Hui, T.W.; Loy, C.C. Inter-region affinity distillation for road marking segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12486–12495. [Google Scholar]
  47. Gupta, S.; Hoffman, J.; Malik, J. Cross modal distillation for supervision transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2827–2836. [Google Scholar]
  48. Papernot, N.; McDaniel, P.; Wu, X.; Jha, S.; Swami, A. Distillation as a defense to adversarial perturbations against deep neural networks. In Proceedings of the 2016 IEEE symposium on security and privacy (SP), San Jose, CA, USA, 22–26 May 2016; pp. 582–597. [Google Scholar]
  49. Liu, Y.; Chen, K.; Liu, C.; Qin, Z.; Luo, Z.; Wang, J. Structured knowledge distillation for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2604–2613. [Google Scholar]
  50. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. arXiv 2017, arXiv:1703.01780. [Google Scholar]
  51. Ghahramani, Z. Probabilistic machine learning and artificial intelligence. Nature 2015, 521, 452–459. [Google Scholar] [CrossRef] [PubMed]
  52. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the 33rd international conference on machine learning, New York, NY, USA, 19–24 June 2016; pp. 1050–1059. [Google Scholar]
  53. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  54. Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. arXiv 2016, arXiv:1612.01474. [Google Scholar]
  55. Van Amersfoort, J.; Smith, L.; Teh, Y.W.; Gal, Y. Uncertainty estimation using a single deep deterministic neural network. In Proceedings of the 37rd International Conference on Machine Learning, Online, 12–18 July 2020; pp. 9690–9700. [Google Scholar]
  56. Zhang, Z.; Dalca, A.V.; Sabuncu, M.R. Confidence calibration for convolutional neural networks using structured dropout. arXiv 2019, arXiv:1906.09551. [Google Scholar]
  57. Tagasovska, N.; Lopez-Paz, D. Single-model uncertainties for deep learning. arXiv 2018, arXiv:1811.00908. [Google Scholar]
  58. Shen, Y.; Zhang, Z.; Sabuncu, M.R.; Sun, L. Real-time uncertainty estimation in computer vision via uncertainty-aware distribution distillation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Online, 5–9 January 2021; pp. 707–716. [Google Scholar]
  59. Jin, X.; Lan, C.; Zeng, W.; Chen, Z. Uncertainty-aware multi-shot knowledge distillation for image-based object re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11165–11172. [Google Scholar]
  60. Mehrtash, A.; Wells, W.M.; Tempany, C.M.; Abolmaesumi, P.; Kapur, T. Confidence calibration and predictive uncertainty estimation for deep medical image segmentation. IEEE Trans. Med. Imaging 2020, 39, 3868–3878. [Google Scholar] [CrossRef]
  61. Oh, D.; Ji, D.; Jang, C.; Hyunv, Y.; Bae, H.S.; Hwang, S. Segmenting 2k-videos at 36.5 fps with 24.3 gflops: Accurate and lightweight realtime semantic segmentation network. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Online, 31 May–31 August 2020; pp. 3153–3160. [Google Scholar]
  62. Kwon, W.; Jin, Y.; Lee, S.J. Collision Identification of Collaborative Robots Using a Deep Neural Network. IEMEK J. Embed. Syst. Appl. 2021, 16, 35–41. [Google Scholar]
  63. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. The definition of labels. (a) presents 6-DoF articulated robot, and (b) presents the definition of categories; yellow arrows in (b) indicate categorical labels according to collision locations.
Figure 1. The definition of labels. (a) presents 6-DoF articulated robot, and (b) presents the definition of categories; yellow arrows in (b) indicate categorical labels according to collision locations.
Sensors 21 06674 g001
Figure 2. Examples of sensor signals and ground truth data. (a) shows a part of the acquired sensor signals, and (b) presents examples of generated ground truth data around collision occurrences. Green lines with numbers in (b) indicate labeled categories in the ground truth data.
Figure 2. Examples of sensor signals and ground truth data. (a) shows a part of the acquired sensor signals, and (b) presents examples of generated ground truth data around collision occurrences. Green lines with numbers in (b) indicate labeled categories in the ground truth data.
Sensors 21 06674 g002
Figure 3. The architecture of the teacher network.
Figure 3. The architecture of the teacher network.
Sensors 21 06674 g003
Figure 4. The procedure of uncertainty-aware knowledge distillation for the training of the student network; SN and TN indicate the student and teacher networks, respectively, and σ ( z , T ) is the softmax function with the temperature T.
Figure 4. The procedure of uncertainty-aware knowledge distillation for the training of the student network; SN and TN indicate the student and teacher networks, respectively, and σ ( z , T ) is the softmax function with the temperature T.
Sensors 21 06674 g004
Figure 5. The procedure for the post-processing. (a) presents the predictions from the student network, and (b) presents the result of grouping non-zero connected samples and assigning an identical category of the maximum frequent. (c) presents the result of a thresholding method.
Figure 5. The procedure for the post-processing. (a) presents the predictions from the student network, and (b) presents the result of grouping non-zero connected samples and assigning an identical category of the maximum frequent. (c) presents the result of a thresholding method.
Sensors 21 06674 g005
Figure 6. Examples of predictions before and after the post-processing. (a) presents predictions for the collision categories of 4 and 5, and (b) presents predictions for the collision categories of 2 and 3.
Figure 6. Examples of predictions before and after the post-processing. (a) presents predictions for the collision categories of 4 and 5, and (b) presents predictions for the collision categories of 2 and 3.
Sensors 21 06674 g006
Figure 7. Examples of true-positive, false-positive, and false-negative collisions for computing collision-level accuracies. (a) presents a T P collision, (b,d) present F P and F N cases, (c) presents T P and F P cases, and (e) presents a F P collision.
Figure 7. Examples of true-positive, false-positive, and false-negative collisions for computing collision-level accuracies. (a) presents a T P collision, (b,d) present F P and F N cases, (c) presents T P and F P cases, and (e) presents a F P collision.
Sensors 21 06674 g007
Figure 8. F1-scores for the training and validation datasets.
Figure 8. F1-scores for the training and validation datasets.
Sensors 21 06674 g008
Figure 9. Uncertainties measured by MC-dropout of the teacher network. (a) shows the first dimension of 24-dimensional sensor data, and (b) presents uncertainties measured by MC-dropout. In (a), red × marks indicate collision moments, and green lines represent labels for the normal state and locations of collisions.
Figure 9. Uncertainties measured by MC-dropout of the teacher network. (a) shows the first dimension of 24-dimensional sensor data, and (b) presents uncertainties measured by MC-dropout. In (a), red × marks indicate collision moments, and green lines represent labels for the normal state and locations of collisions.
Sensors 21 06674 g009
Table 1. The number of collisions and data samples. Total indicates the number of data samples, which were collected with a sampling rate of 1 kHz, and Collision indicates the number of collision samples.
Table 1. The number of collisions and data samples. Total indicates the number of data samples, which were collected with a sampling rate of 1 kHz, and Collision indicates the number of collision samples.
Training SetValidation SetTest Set
Collisions39065581122
SamplesTotalCollisionTotalCollisionTotalCollision
19,563,048781,2002,778,777111,6005,798,685224,400
Table 2. Sample-level accuracies of the four different neural network models before and after the post-processing.
Table 2. Sample-level accuracies of the four different neural network models before and after the post-processing.
Before Post-ProcessingAfter Post-Processing
Recall Precision F 1 - Score Recall Precision F 1 - Score
Base model98.161198.398598.279698.547399.061798.8038
Student network98.201598.345898.273698.599299.019898.8091
Proposed method98.311098.451698.381298.711999.046598.8789
Teacher network98.272998.533798.403198.562999.101198.8313
Table 3. The numbers of true-positive ( T P ), false-positive ( F P ), and false-negative ( F N ) collisions of the four neural network models before and after post-processing.
Table 3. The numbers of true-positive ( T P ), false-positive ( F P ), and false-negative ( F N ) collisions of the four neural network models before and after post-processing.
Before Post-ProcessingAfter Post-Processing
TP FP FN TP FP FN
Base model1119229311191213
Student network1118295411181094
Proposed method112020521120762
Teacher network111926731119773
Table 4. Collision-level accuracies of the four different neural network models before and after the post-processing.
Table 4. Collision-level accuracies of the four different neural network models before and after the post-processing.
Before Post-ProcessingAfter Post-Processing
Recall Precision F 1 - Score Recall Precision F 1 - Score
Base model99.732678.913988.110299.732690.241994.7502
Student network99.643679.122488.205299.643591.116595.1894
Proposed method99.821784.528391.540699.821793.645496.6350
Teacher network99.732680.735989.234499.732693.561996.5487
Table 5. The averaged processing time in ms for the collision identification.
Table 5. The averaged processing time in ms for the collision identification.
Inference TimeDetection DelayPost-ProcessingTotal
Base model1.76410.82390.20572.7938
Student network1.76410.61980.20572.5897
Proposed method1.76410.66510.20572.6350
Teacher network3.23480.70060.20574.1412
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kwon, W.; Jin, Y.; Lee, S.J. Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots. Sensors 2021, 21, 6674. https://doi.org/10.3390/s21196674

AMA Style

Kwon W, Jin Y, Lee SJ. Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots. Sensors. 2021; 21(19):6674. https://doi.org/10.3390/s21196674

Chicago/Turabian Style

Kwon, Wookyong, Yongsik Jin, and Sang Jun Lee. 2021. "Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots" Sensors 21, no. 19: 6674. https://doi.org/10.3390/s21196674

APA Style

Kwon, W., Jin, Y., & Lee, S. J. (2021). Uncertainty-Aware Knowledge Distillation for Collision Identification of Collaborative Robots. Sensors, 21(19), 6674. https://doi.org/10.3390/s21196674

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop