Next Article in Journal
Close-Range Coordination to Enhance Constant Distance Spacing Policies in Oversaturated Traffic Systems
Next Article in Special Issue
Human-to-Robot Handover Based on Reinforcement Learning
Previous Article in Journal
Self-Assembled TiN-Metal Nanocomposites Integrated on Flexible Mica Substrates towards Flexible Devices
Previous Article in Special Issue
NEP+: A Human-Centered Framework for Inclusive Human-Machine Interaction Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Social Type-Aware Navigation Framework for Mobile Robots in Human-Shared Environments

1
AgeTech-Service Convergence Major, Department of Electronic Engineering, Kyung Hee University, Yongin 17104, Republic of Korea
2
Department of Artificial Intelligence, Kyung Hee University, Yongin 17104, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(15), 4862; https://doi.org/10.3390/s24154862
Submission received: 25 June 2024 / Revised: 22 July 2024 / Accepted: 24 July 2024 / Published: 26 July 2024
(This article belongs to the Special Issue Intelligent Social Robotic Systems)

Abstract

:
As robots become increasingly common in human-populated environments, they must be perceived as social beings and behave socially. People try to preserve their own space during social interactions with others, and this space depends on a variety of factors, such as individual characteristics or their age. In real-world social spaces, there are many different types of people, and robots need to be more sensitive, especially when interacting with vulnerable subjects such as children. However, the current navigation methods do not consider these differences and apply the same avoidance strategies to everyone. Thus, we propose a new navigation framework that considers different social types and defines appropriate personal spaces for each, allowing robots to respect them. To this end, the robot needs to classify people in a real environment into social types and define the personal space for each type as a Gaussian asymmetric function to respect them. The proposed framework is validated through simulations and real-world experiments, demonstrating that the robot can improve the quality of interactions with people by providing each individual with an adaptive personal space. The proposed costmap layer is available on GitHub.

1. Introduction

Recent advances in robotics technology have enabled autonomous mobile robots to be used in various places. Many robots already perform in social spaces where they coexist with humans, such as museums, airports, and hotels [1,2]. When navigating social spaces, robots must avoid obstacles and humans safely. In this context, conventional mobile robot navigation systems consider humans as dynamic obstacles and focus on avoiding physical collisions [3,4,5]. However, humans still feel discomfort or lack trust in such robot behavior. Therefore, it is crucial for robots to avoid physical collisions with humans and ensure that humans feel safe and comfortable with the robot’s behavior [6]. Proxemics is a field of psychological research that studies how people use the physical space around them and how they position themselves concerning each other [7]. When people navigate an environment and engage in social interactions, they subconsciously consider the personal space of others. They may feel uncomfortable if the distance between two people is closer than expected. Robots should also respect these social norms in social spaces and consider people’s personal space [8]. In this context, many researchers are applying proxemics concepts to robot navigation, and this approach has been studied in Refs. [9,10,11].
Most studies treat all people the same and do not consider different age groups and situations. However, in the real world, children and adults have different social and physical characteristics, and people move in different patterns, such as moving alone or in groups. In particular, children have less ability to react to accidents than adults, which can lead to safety issues when interacting with robots [12,13]. Ref. [14] investigated human–robot interaction and found that children tend to keep more distance than adults when interacting with a robot. In this context, applying a one-size-fits-all avoidance strategy to everyone without considering diversity can make people uncomfortable with the robot’s behavior or lead to unexpected situations. Robots should adopt different avoidance strategies for different people to feel more comfortable and safer in their interactions. We propose a novel Social Type-Aware Navigation Framework for robots to directly detect and track people in real-world environments and adapt avoidance strategies for each individual.
Thus, we propose a framework that enables robots to move safely in social spaces between different ages and types of people, including socially vulnerable people such as children and the elderly, while understanding their personal space and social context. To this end, the framework includes a ‘Social Type Classification Pipeline’ that enables robots to detect people in real-world environments directly, classify them into different social types, and a ‘Social Type-Aware Costmap (STAC)’ that enables them to navigate while respecting the appropriate social distance according to each classified social type. Our framework is illustrated in Figure 1 and is designed as a plugin that can be easily integrated into an ROS-based navigation system, allowing it to be used with various path planning methods. Applying our framework enables robots to move among people more naturally and in a socially contextualized manner, as shown in Figure 2. Our main contributions to this work can be summarized as follows:
  • We present a pipeline using robot sensors to classify humans into diverse social types, moving beyond the conventional approach of treating all humans the same.
  • We develop a new layered costmap that adjusts personal space for each social type, enabling robots to navigate more sensitively and safely around humans.
  • We validate our framework through qualitative and quantitative analysis in simulated and real-world human–robot interaction scenarios and demonstrate its performance.
This paper is organized as follows: Section 2 describes the background and necessity of this research by reviewing the related works. Next, Section 3 details the Social Type-Aware Navigation Framework proposed in this paper. In Section 4, we qualitatively and quantitatively analyze the results of the experiments conducted with human–robot interaction scenarios in simulated and real-world environments to validate the performance of the proposed framework. Finally, Section 5 summarizes the contributions of this work and concludes the paper with suggestions for future works.
Figure 1. An overview of the Social Type-Aware Navigation Framework.
Figure 1. An overview of the Social Type-Aware Navigation Framework.
Sensors 24 04862 g001
Figure 2. Social Type-Aware Navigation Framework for comfort in social spaces.
Figure 2. Social Type-Aware Navigation Framework for comfort in social spaces.
Sensors 24 04862 g002

2. Related Work

2.1. Human Detection and Tracking

In dynamic and unknown environments, detecting nearby obstacles and people through sensors is essential for robots to navigate safely to their goal. LiDAR and cameras are commonly used for mobile robots, and LiDAR is useful for detecting humans by tracking the geometric features of their legs [15]. However, there is a limitation in identifying human features, such as age group, using only LiDAR.
On the other hand, with the improvement of computer vision systems, algorithms such as R-CNN [16], Faster R-CNN [17], and YOLO [18] have made it possible to detect people in camera images. For example, Ref. [19] presented a method to estimate the age and gender of a human in an image using YOLO. This approach was used in Ref. [20] to recognize the gender of a customer so that a robot could interact with them in a shopping center. Ref. [21] proposed a face detection and tracking framework using convolutional neural networks and the SORT [22] for integration into various real-world HRI systems. However, these methods mainly rely on frontal faces, which are less accurate for the side faces or small-sized faces due to the field of view (FoV) of the camera sensor on the mobile robot. To address this challenge, Refs. [23,24] proposed a novel approach to distinguish between adults and children by calculating the ratio of the torso to the face.
In public spaces and crowded environments, robots need to recognize when people are in groups to plan safer and more efficient paths and to adjust their interactions with people properly. In this context, many researchers have proposed many different methods to detect groups in images. In Ref. [25], the sensor was placed overhead to detect group interactions between people, i.e., from an exterior perspective, but it did not represent the robot vision and was challenging to implement in real-world navigation. Therefore, Refs. [26,27] detected groups from the robot perspective in real time by object detection and clustering using cameras installed on the robot.
Mobile robots should be able to track humans in real time and predict their trajectories to reasonably plan their path so that they can complete tasks more efficiently. To achieve this, Ref. [28] proposed a multi-modal tracking framework that integrates various tracking methods such as Nearest-Neighbor [29], Extended Nearest-Neighbor [30], Multi-Hypothesis [31], and Vision-Based MDL [32] to improve human tracking in crowded environments. Also, in Ref. [33], they used a real-time tracking algorithm, SORT [22], based on a Kalman filter to identify the bounding boxes captured by a camera to track a human continuously. These studies contribute to predicting human movements when a mobile robot operates in a social space.

2.2. Proxemics in Human–Robot Interaction

When a robot needs to navigate around humans, it is important to consider proxemic rules to understand and respond appropriately during interactions with humans. For example, Ref. [34] evaluated subjects’ reactions to the robot’s avoidance behavior with real-world experiments in which a robot and a human encountered each other in a hallway. They found that the subjects felt more comfortable when the robot kept away from them. Also, Ref. [35] defined the shape of personal space by evaluating the comfort humans feel with the movement of the humanoid and non-humanoid robots when passing around the person. These results suggest that robots should respect their personal space and behave predictably to ensure comfort and safety while interacting with humans. Ref. [11] developed this concept in this context and modeled a personal space as an asymmetric Gaussian function. Refs. [36,37] integrated the Gaussian function from Ref. [11] into a costmap by adding a cost around a person to make the robot respect their personal space. The research in Ref. [38] developed a dynamic social force model, which determines the parameters of personal space with a fuzzy inference system and continuously adjusts them through reinforcement learning. In Refs. [39,40], the robot detected human emotions and incorporated emotion-based proxemics constraints into the costmap. This approach makes human–robot interactions smoother by allowing the robot to recognize human emotions and react appropriately.
Meanwhile, human interaction in crowded environments is not limited to interactions between individuals. For example, humans tend to avoid groups they are not involved with. In this context, it is crucial for robots to understand the interactions with human groups and react accordingly. That is, robots should recognize the spaces generated by groups of humans and navigate among them effectively. Kendon’s F-formation [41] is a good tool for understanding the interactions between such groups, including the o-space shared by group members, the p-space where group members stand, and the R-space that divides the group from outer space. Refs. [42,43,44,45,46] took this idea and proposed an approach in which the robot maintains an appropriate distance from the group so that the individual as well as the group members feel comfortable, and Refs. [47,48] addressed how the robot detects groups and approaches them with an appropriate pose when joining the group.
Although many studies have applied proxemics to social navigation, a lack of research defines and considers personal space for children. By categorizing people into different age groups or situations instead of a single category and respecting their different personal spaces, a robot can navigate safely and be human-aware in real-world environments. This differentiated approach is an important step forward. It will make robots interact with all kinds of humans, including children.

3. Methods

3.1. Social Type Classification Pipeline

Many studies in the area of socially aware navigation use ceiling-mounted cameras to detect people [49]. In this work, we developed a pipeline to detect, track, and classify people using sensors installed on a robot for applications in various real-world environments.
First, we chose Ultralytics’ open-source YOLOv7 model [50] to detect people in RGB images. It provides high accuracy and fast processing speed and improves performance compared to existing models. We trained YOLOv7 on the CrowdHuman dataset [51], which provides three annotated datasets: human visible area, head, and full body. Applying the trained model, we can see that the bounding boxes for the head and the whole body are detected, as shown in Figure 3.
For the tracking system, we use SORT (Simple Online and Realtime Tracking) [22]. This algorithm tracks each uniquely identified object through the scene and can re-identify them even if they leave and re-enter the scene. The output of this system includes an ID for each object, along with its start and end positions in terms of x and y coordinates. These coordinates can be used to analyze the movement and path of each person. By analyzing the history of these positions, the system can estimate a person’s speed and trajectory, providing valuable insight into their movement patterns within the scene.
Figure 3 shows the results of detecting and tracking people using YOLOv7 and SORT in a Gazebo environment. The left image shows the Gazebo and the right image shows the detection and tracking results. The blue bounding boxes indicate detections of heads and the orange ones full bodies. The top of each orange bounding box also shows the ID of the detected person, which can be used to effectively distinguish between multiple people in the image.
To estimate the age of the detected person, we use the human head-to-body ratio [52]. Since there is a difference in this ratio for each age group, we can identify adults and children by comparing the size of the head and body bounding boxes of each detected person. If multiple humans are detected, classification is performed by ensuring that the center of the head bounding box is within the full body bounding box before calculating the ratio of the bounding boxes. The threshold for distinguishing between adults and children based on their head-to-body ratio was determined through prior experiments in various real-world settings. A ratio of 7.0 or less was classified as a child, while a ratio greater than 7.0 was classified as an adult. This classification ensures that the robot moves carefully when interacting with children.
After detecting and classifying humans, we use the optical principle of RGB-D cameras to determine their actual location. From the depth image extracted from the RGB-D image, the bounding box and the internal parameters of the camera can be used to convert the position of the human from the 2D image to 3D coordinates in the camera coordinate system. This conversion is calculated as follows:
X C = u p x f x Z C , Y C = v p y f y Z C , Z C = Z
In Equation (1), ( u , v ) is the center coordinate of the 2D bounding box, ( p x , p y ) is the principal point coordinate of the camera, ( f x , f y ) are the focal lengths of the camera, Z is the value of the corresponding pixel in the depth image, and  ( X C , Y C , Z C ) represents the 3D coordinate in the camera coordinate system. Finally, the actual human position can be obtained through a coordinate transformation between the camera and robot coordinate systems.
Figure 4 shows the result of the human position detected by the RGB-D camera in the robot coordinate system. It shows the process of detecting a person in the RGB image from the camera on the robot, combining the depth information of the corresponding bounding box with the internal parameters of the camera and converting the position in the 2D image to 3D spatial coordinates in the robot coordinate system. When a robot is navigating around people, the robot must know the actual position and movement of the person, which allows the robot to predict the future trajectories of the person and thus avoid collisions with the person in advance. In this paper, to estimate the human’s motion and obtain the human’s exact position, we use the Kalman filter [53].
The robot we use has a 90-degree field of view for the camera and a 270-degree field of view for the LiDAR. Since the field of view of the RGB-D sensor covers only 1/3 of the laser scanner’s field of view, we use Ref. [15], which estimates the position and speed of people with the LiDAR sensor, to recognize all the people around the robot.
There are not only individuals but also human groups in social space. Many researchers proposed a group identification method using clustering techniques in this context. However, these methods are limited in real-time applications due to data processing. Therefore, we propose a simple group detection algorithm that can recognize groups in a dynamic environment. According to proxemics, people tend to be distant from strangers more than familiar people. The algorithm calculates the differences in distance and direction of movement between people. It classifies them into a group if they move in the same direction and the distance is within a threshold range based on proxemics.
Algorithm 1 shows the proposed group recognition algorithm for recognizing groups based on individuals’ spatial and orientation data. It utilizes two pieces of information about each individual: their location, denoted by ( x i , y i ) , and their orientation, represented by the quaternion q i , to detect whether they are in a group.
The process starts by converting the quaternion into a rotation matrix (using the RotMat function) so that the orientation of each individual can be easily used for spatial analysis. This transformation is essential for determining each individual’s heading vector, which enables us to know where each individual is facing. Next, it calculates the distance and angular difference between individuals. The distance is determined using the Euclidean distance formula, and the angle is a simple calculation obtained by comparing the orientation vectors derived from the rotation matrix (using the VecAngle function). The final step is to apply a predefined threshold to these calculated distances and angles to classify objects as part of a group or as individual objects. If the distance and angle differences are within this threshold, they are considered to belong to the same group.
Algorithm 1 Group Classification
1:
Input: For each human p i , position as p i = ( x i , y i ) , orientation as quaternion q i
2:
Output: Classification of each pair as a group or individuals
3:
function RotMat(q)
4:
  Compute rotation matrix from quaternion q
5:
  return rotation matrix
6:
end function
7:
function VecAngle(v1, v2)
8:
    u 1 v 1 / v 1
9:
    u 2 v 2 / v 2
10:
   d p dot product of u 1 and u 2
11:
   a n g l e arccos ( max ( 1 , min ( 1 , d p ) ) )
12:
  return angle
13:
end function
14:
for  i = 1 to n 1 do                ▹ For each p i in people
15:
  for  j = i + 1 to ndo           ▹ For each p j in people after p i
16:
    d i s t ( x j x i ) 2 + ( y j y i ) 2
17:
    r i R o t M a t ( q i )
18:
    r i R o t M a t ( q j )
19:
    d i r i · [ 1 , 0 , 0 ]                 ▹ Direction vector of p i
20:
    d j r j · [ 1 , 0 , 0 ]                 ▹ Direction vector of p j
21:
    a n g l e V e c A n g l e ( d i , d j )
22:
   if ( d i s t < d i s t T h r e s h ) and ( a n g l e < a n g l e T h r e s h ) then
23:
    Classify p i and p j as a group
24:
  end if
25:
end for
26:
end for
In conclusion, our pipeline becomes an essential basis for the Social Type-Aware Costmap (STAC).

3.2. Social Type-Aware Costmap (STAC)

Costmap [54] is a data structure that represents the environments around the robot in the form of a 2D grid map, where each grid cell represents the ’cost’, or ’risk’, of the position. The cost represents the expected risk of the robot navigating through the area. The robot can choose a safe and efficient path by avoiding high-cost areas and preferring low-cost areas. Costmaps are widely used for socially aware motion planning and navigation by introducing non-lethal costs to represent social space [55]. Ref. [36] proposed a layered costmap, each containing semantic information about specific attributes such as obstacles, inflation, or proxemics. They represent proxemics by utilizing the detected position and velocity of humans to generate Gaussian distributions of costs around them [11].
While these approaches effectively represent human personal space, they must consider that it can vary from person to person. Hence, in this work, we propose a Social Type-Aware Costmap (STAC), a multi-layer costmap that includes the personal space for each type of person, such as adults, children, and groups. Our costmap allows the robot to respect children at a greater distance than adults. The STAC consists of two layers: the age-based layer, which represents personal space for different age groups, and the group-based layer, which represents group space.

3.2.1. Age-Based Layer

We use an asymmetric Gaussian function to represent the personal space in mathematical form [11]. It is defined in Equation (2) and is visually illustrated in Figure 5. This modeling method helps the robot to move while ensuring an appropriate distance from the humans around it.
f ( x , y ) = A exp d c o s ( θ θ i ) 2 σ x 2 + d s i n ( θ θ i ) 2 σ y 2
σ x = σ i ( 1 + k v v ) , σ y = σ i
Using the human position from our Social Type Classification Pipeline, we define a danger zone centered on the human’s location ( x 0 , y 0 ) . Given the moving direction of the person as θ i = atan 2 ( v y , v x ) , we calculate the cost value as the difference between the distance and direction of the person and any cell ( x , y ) on the costmap, where d is determined by d = ( x x 0 ) 2 + ( y y 0 ) 2 and θ is by θ = atan 2 ( ( y y 0 ) , ( x x 0 ) ) . The value decreases as it goes further from the person and adjusts based on the orientation difference between the person and the cell. The parameters of the Gaussian function, A , σ x , σ y , represent the amplitude of the function and the standard deviation in the horizontal and vertical directions, respectively, which determine the size and shape of the personal space around the person. This zone of danger scales with the direction of the person’s movement, and, the faster the person moves, i.e., the greater the speed, the wider this zone of danger becomes. This is regulated through the scaling factor k v , which adjusts the shape of the Gaussian function depending on the human’s speed and is shown in Equation (3). Through all this, the robot can accurately determine the personal space around people and move carefully between them accordingly. This is essential for the robot to respect social manners and personal space during interactions with people. Ref. [14] studied proxemics with a robot regarding participants’ age and found that children tended to stand further away from the robot than adults. This study suggests differences in proxemics between children and adults. Therefore, in this paper, we propose a new approach to make children’s personal space more extensive than that of adults. The size and shape of the personal space are based on the basic parameter σ 0 , and Equation (4) describes how to use this parameter to differentiate the size of the personal space for children and adults. This takes into account the different proxemic needs of children and adults.
σ i = σ 0 , if adult 1.4 σ 0 , if child
The personal space parameter σ 0 for adults is set based on the proxemics. This theory takes the average distance an adult maintains in their general social interactions with strangers as a reference to their personal space. For example, if an adult’s average distance in a typical interaction with a stranger is 1.2 m, this distance is set to σ 0 . On the other hand, research in Ref. [14] shows that children keep a distance of about 1.7 m on average when interacting with a robot. It is relatively longer than adults’ distance with a robot, so we should consider children’s personal space to be about 1.4 times larger than that of adults. Therefore, the personal space parameter for children is set to 1.4 times that of adults. This approach respects the difference in proxemics between children and adults and contributes to a more accurate interaction design.
Figure 6 shows the result of applying the age-based layer. Notice the difference in the size of the Gaussian distributions applied to adults and children, particularly that the distribution for a moving child is extended in the moving direction. Using this information, the robot can adapt its path to avoid getting too close to the child.

3.2.2. Group-Based Layer

f G ( x , y ) = A G exp d c o s ( θ θ i ¯ ) 2 σ x G 2 + d s i n ( θ θ i ¯ ) 2 σ y G 2
σ x G = σ G ( 1 + k v v ¯ ) , σ y G = σ G
To ensure human comfort in the context of human standing and moving in groups, the robot should also respect the group space shared by the group members. Compared to the age-based layer, represented based on the position and heading of individuals, shown in Figure 7a, the group-based layer captures the dynamics of the group of people shown in Figure 7b. It represents the group space based on the center position of the group ( x 0 ¯ , y 0 ¯ ) and the average direction θ i ¯ in which the group is moving together. Group space is defined similarly to individual space and is shown in Figure 7b and Equation (5). The group space is formed at the center of two individuals and, like the individual space described in Section 3.2.1, computes the cost values around the group members by comparing a random cell in the cost map with the group’s center position and moving direction. Here, d is determined by d = ( x x 0 ¯ ) 2 + ( y y 0 ¯ ) 2 , and θ is by θ = atan 2 ( ( y y 0 ¯ ) , ( x x 0 ¯ ) ) . The parameter σ G , which determines the size of the space, is defined by the distance d G from the group’s center to the farthest group member and the scaling factor k G . That is, it is calculated with σ G = k G d G . The wider the group members are separated, i.e., the larger d is, the larger σ G is, and the wider the danger zone around the group. It helps the robot keep a safe distance from all group members as it plans its path, considering the group.
Figure 8 shows a comparison of the results with and without the STAC, simulating a situation where two people are talking in a virtual environment, Gazebo. Figure 8a shows without STAC, which does not capture the interaction between the two people, and the area between them is shown as free space for the robot to pass through, while Figure 8b is the result of with STAC, which detects both the individual and group space of the two people and shows the space where the robot cannot pass through, i.e., the lethal space. The proposed costmap layer is released as an open-source repository on GitHub [56].

4. Experiments

We demonstrated on both a simulator and an actual mobile robot that the proposed method enabled the robot to arrive at its destination in a dynamic environment where humans move around without causing discomfort. By experimenting with the robot’s interactions with different types of humans, including children, adults, and human groups, we verified that our method provides more comfort for humans than the traditional method. It presents a significant challenge to existing navigation strategies that consider humans only as single objects.
We used the Navfn [57] as the global planner and the Dynamic Window Approach (DWA) [58] as the local planner, and the navigation settings were kept consistent throughout the experiments. The main comparison is between the traditional costmap (obstacles and inflation layer) with and without the Social Type-Aware Costmap (STAC). Also, due to the child’s small size, the LiDAR cannot detect the child’s legs, so the camera holds the last detection position for 4 seconds in case the child is out of range. A video with our experimental results is available at https://youtu.be/LywBJ9KIxSE, accessed on 24 June 2024.

4.1. Robot Specifications

As shown in Figure 9a,b, the mobile robot used in both simulations and real-world experiments is Jackal [59], which is equipped with a 2D LiDAR and RGB-D sensor. The LiDAR sensor can measure objects between 0.05 m and 25 m and has a field of view of 270 degrees. We used Intel RealSense D435 as the RGB-D sensor, which has a minimum depth capability of 10.5 cm and supports depth/RGB frame rates up to 90/30 fps. The camera provides sensor resolutions up to 1280 × 720 for depth and 1920 × 1080 for RGB, with the field of view (FoV) being 87° × 58° × 94° for depth and 69° × 42° × 77° for RGB. Figure 9c shows a comparison of the field of view of the LiDAR and camera installed on the robot. The LiDAR sensor has a field of view of 270 degrees and the RGB-D camera has a field of view of 87 degrees for depth.

4.2. Experimental Setup and Scenarios

Our study includes simulation experiments to evaluate the framework with the Gazebo and the Robot Operating System (ROS) and real-world experiments to validate the performance of the proposed framework in social space. The simulation test environment was a corridor-like environment with child and adult models. To simulate the behavior of social agents, we used pedsimROS [60], an ROS package that models walking pedestrians using the Social Force Model [61]. They were set to recognize the robot and avoid it if necessary, and the child model was moved slower than the adult model.
We designed three test scenarios in which the robot had to move from the initial position to the goal while passing an adult, a child, and a human group. Each of these scenarios was conducted five times to ensure the consistency and reliability of the results. These scenarios were designed to evaluate how socially appropriate the robot interacts with humans with and without the Social Type-Aware Costmap (STAC). With this comparison, we can see how the use of the STAC affects the robot’s social type-aware navigation.
The real-world experiments consisted of 8 scenarios, each involving the robot interacting with an adult and a child and group interactions with two participants. The group interaction experiments included cases with two adults and cases where one participant was a child and the other an adult. Each scenario was designed for two different situations: when the participants were stationary and moving. The more detailed description of the scenario and its illustrations is presented in Section 4.4. It allowed us to observe how the robot reacted in different social situations. All experiments were compared between traditional navigation methods and using the STAC to evaluate how the robot navigated. The STAC parameter settings for simulation and real-world environments are shown in Table 1.

4.3. Simulation Experiment Results

Figure 10 shows the trajectories of the robot and human model in the simulation experiment. Here, the robot’s destination is indicated by the red arrow. In Figure 10a, the red line shows the trajectory of the adult model and the green color shows the trajectory of the robot, where the light green is the trajectory of the robot when using the traditional costmap alone and the dark green is the trajectory when applying our proposed STAC method. Figure 10b shows the trajectory when the robot interacts with the child model, where the gray line shows the movement trajectory of the child model, and the light and dark orange show the robot trajectory when the traditional and STAC methods are applied, respectively. Figure 10c shows the trajectory of the robot avoiding an adult model moving in a group, comparing the results of the traditional and STAC methods. Here, the gray and cyan colors indicate the trajectories of the group members, the green indicates the trajectory of the robot, the light color indicates the traditional method, and the dark color indicates the trajectory when using the STAC method.
From these comparisons, we can see that, when using the STAC method, the robot moves with a clearer avoidance of the humans. In particular, when comparing Figure 10a,b, we can see that the robot tends to avoid the child model with a wider trajectory compared to the adult model. In addition, Figure 10c shows that the robot avoids the group more effectively when applying STAC than when using the traditional method, and, notably, we can see that the human nearest to the robot (light gray trajectory) tries to avoid the robot. The comparative analysis of these trajectories demonstrates the effectiveness of our approach on the robot’s avoidance strategy with respect to humans. More detailed numerical results can be found in Table 2. We used the following indices to evaluate and compare the methods.
  • Navigation Time (NT): The traveling time until the robot arrives at the goal position [s];
  • Path Length (PL): The total distance of the robot until it arrives at the goal position [m];
  • Minimum Distance (MD): The minimum distance to human in all cases [m];
  • Invasion Distance (ID): The total distance the robot moved while invading the personal space (1.2 m for an adult; 1.7 m for a child) [m].
Note that all the metrics except for minimum distance are average values.
Table 2. Simulation results.
Table 2. Simulation results.
Social TypeMethodNT [s]PL [m]MD [m]ID [m]
AdultTraditional22.399.510.760.59
STAC (Ours)22.699.591.110.15
ChildTraditional22.239.510.861.20
STAC-A22.779.651.230.99
STAC (Ours)23.629.751.460.68
GroupTraditional31.6912.680.410.19
STAC (Ours)30.2812.911.000.33

4.3.1. Adult

When using only the traditional costmap, the robot considered the adult as an obstacle and followed the shortest path to the goal, as shown by the light green line in Figure 10a. It tended to approach without any specific avoidance behavior that causes discomfort to the adult. However, MD and ID, indicators of human discomfort were found to be 0.76 m and 0.59 m, respectively, where MD means that the robot was closer to the person by 0.76 m. ID means that the robot invaded about 0.59 m, the minimum distance of 1.2 m that a human tries to maintain with a stranger, which means that, with the traditional costmap, the robot could get too close to the personal space of an adult and cause discomfort.
On the other hand, when applying our proposed STAC system, the robot approached the human at a minimum distance of 1.11 m, which is slightly closer than the human personal space threshold of 1.2 m. Moreover, the minimum distance the robot approached into the personal space was reduced by 0.44 m compared to the traditional method. Although STAC did not fully respect the human’s personal space, it still represents a significant improvement in ensuring that the human does not feel uncomfortable with the robot’s behavior compared to the traditional costmap method, as shown by the green line in Figure 10a. In addition, the navigation time and path length increased by 0.3 s and 0.08 m, respectively, compared to the conventional costmap results, but the difference was insignificant. These results show that STAC can be applied to navigate without significantly reducing navigation efficiency while respecting personal space and ensuring that humans are not uncomfortable with the robot’s behavior in human–robot interactions.

4.3.2. Child

In the child–robot interaction scenario, we compared the traditional costmap, STAC, and STAC-A, which used the same settings as adults. The results show that the traditional costmap performed best regarding NT and PL. However, STAC performed the best in terms of MD and ID, which are indicators of ensuring a safe distance to a child. When the child’s personal space was set to 1.7 m, the ID value was 0.68 m despite using STAC. This is because the costmap generated around the child was defined as a Gaussian with a radius of about 1.7 m, but the center of the costmap was set to high risk, and the edges were set to relatively low risk. This is because, if the entire costmap was set to high risk, the robot would not be able to generate a path and would stop. While STAC-A also performed well regarding personal space consideration, the STAC method is more appropriate as it provides a larger personal space for children to protect them from unexpected situations during interaction with the robot. Furthermore, despite the NT and PL of 23.62 s and 9.75 m, respectively, increases of about 1.4 s and 0.24 m compared to the traditional method, the difference is insignificant throughout the navigation. The important thing is that the results show that our method reliably detects children and fully understands the extended personal space of children compared to the experiment with adults, applying different avoidance strategies than adults. This result demonstrates that our proposed Social Type-Aware Navigation Framework effectively distinguishes between children and adults and applies appropriate navigation strategies for each. As shown in Figure 10a,b, it is clear that the robot takes different trajectories when interacting with an adult and a child.

4.3.3. Group

In a group interaction with a robot, we set the goal to the center of the group members. Using the traditional method, the robot planned the shortest path through the group’s center, which caused people to move to avoid the robot, as shown by the light green line in Figure 10c. While this approach may appear to reduce PL, increase NT, and improve ID values, this is actually because the robot had to stop while driving between people. The MD of 0.41 m may be a sign that the robot caused discomfort rather than providing comfort.
In contrast, when STAC is applied, the robot plans its path in a way that recognizes and respects personal and group space. As shown by the green line in Figure 10c, this allows people to move while maintaining their original trajectory, providing the robot a way of behaving that considers social etiquette. In particular, the traditional approach of trying to drive through the middle of a group is contrary to social etiquette, which is an important consideration for robot operation in public spaces. Using STAC, the robot can plan its avoidance path, reducing NT while achieving good performance in MD and ID, making human–robot interactions more natural and comfortable.

4.4. Real-World Experiments

4.4.1. Individuals

The real-world individual experimental scenarios were designed for two social types, an adult and a child, as illustrated in Figure 11. The experiments consisted of four different scenarios, including situations where the human is standing or moving, with a distance of about 0.4 m between the robot and the human. The robot navigates from a starting to a ’Robot goal’. The human standing is located 3.5 m vertically and 0.4 m horizontally away from the robot’s starting point, which can be seen in Figure 11a,b. On the other hand, in the scenario where the human is moving, the human and the robot start from the same point. The human moves about 6 m vertically and 0.4 m horizontally to the ’Human goal’, as shown in Figure 11c,d. By experimenting and comparing the robot’s response in these different situations using only the traditional costmap and the proposed STAC technique, we evaluated the performance of our framework in detecting the human and moving to the goal while respecting the personal space of different social types.
In the real-world experiments, we found that, across all scenarios, NT and PL, the navigation efficiency metrics as shown in Table 3 were best when using only the traditional costmap. However, STAC achieved better performance on MD and ID, indicators of human comfort. In the scenario where an adult is stationary, the traditional costmap kept a minimum distance of 0.66 m from the human following the planned path without any consideration for the human, and the total distance traveled within 1.2 m of the human’s personal space was measured to be 3.16 m. While the minimum distance of 0.66 m is shown numerically, this only takes into account the distance between the center of the robot and the human, which means that, in reality, given the width of the Jackal is 430 mm, the human may feel like the robot came within about 0.45 m. This is the general distance a human would keep from family or friends. In contrast, when STAC was applied, the robot considered the personal space when planning its path, ensuring that the minimum distance from the person was 1.09 m, and invaded the personal space while moving 2.80 m. A 1.2 m personal space around the person was not perfectly protected because giving a high-risk value to a space of this size would cause the robot to stop and be unable to generate a path.
In the scenario where a child is standing, the traditional method approached the child with a minimum distance of 0.38 m, resulting from moving closer to the child than the adult due to the child’s smaller size. However, with STAC, the personal space applied around the child was set to be larger than that of the adult, allowing the robot to pass around the child more safely, resulting in a minimum distance between the robot and the child of 1.61 m. This means that a larger personal space was adopted even when compared to the experiment with the adult. When the child’s personal space was set to 1.7 m, the robot invaded this space while traveling 3.12 m with the traditional method, but STAC reduced this to 0.34 m. These results can be seen in Figure 12b,d, where the robot follows a safer trajectory around the child. It is important to note that planning a safe trajectory prioritizes safety even at the expense of some navigation efficiency, as shown by the increase in NT of about 3.35 s and PL of 0.16 m compared to the traditional method. This is a small tradeoff for safety. For the moving adult scenario, the traditional method resulted in driving at 0.41 m, closer than the stationary adult scenario, which caused the human to take action to avoid the robot. This approach did not respect the personal space, making the human feel uncomfortable. However, with STAC, the robot recognized the increased personal space in the direction of the human’s movement and moved away from it, maintaining a distance of 0.88 m from the adult. The experiment with the child showed similar results, maintaining a distance of at least 1.47 m from the child, which is significantly greater than the 0.70 m measured with the traditional method.
Overall, in the scenario of interacting with a single person, traditional methods could not maintain the required minimum distance due to their approach of considering people as obstacles and following the shortest path without taking any special avoidance measures. This approach was particularly challenging in environments with vulnerable individuals, such as children. Because of their small size, traditional methods only recognize children as small obstacles and tend to navigate at closer distances than adults. In contrast, STAC respects personal space, especially in interactions with children. We use cameras to identify children and apply appropriately sized costmap to them, even if they are too small to be detected by LiDAR sensors. Once a child is detected, the costmap is kept at the child’s last known location for a short time, even if the camera no longer detects the child, to ensure a safe distance between the child and the robot. Our approach allows the robot to respect better personal space in interactions with all people, not just children, and enables safe and comfortable navigation.

4.4.2. Human Group

The real group experiment scenario is shown in Figure 13; the experiments focused on setting the distance between the robot and the group members, with a distance of approximately 1 m between group members and either 3.5 or 6 m from the robot. The experiments were tested for two cases: a group consisting entirely of adults and a group including one child. The robot moves from a given starting position to the ’Robot goal’. When the group is standing and interacting with each other, the robot starts 3.5 m away from the group, and the robot goal is set at a position across the center of the group, as shown in Figure 13a,b. In the group moving scenario, the group members start about 6 m away from the robot and move to the ’Group goal’, where the robot goal is also located across the group’s center. It can be seen in Figure 13c,d. These experiments are designed to compare the robot’s behavior when only applying the traditional costmap and STAC methods. This allows us to evaluate how the robot reacts differently to a group consisting of only adults versus a group containing children and how efficiently the robot can access the goal set by crossing the group’s center.
The behavior of the robot and its global path are shown in Figure 14 for all scenarios in the group interaction experiments. The robot using the traditional costmap was utterly unconscious of group interaction and chose paths that crossed between group members, whether they were stationary or moving. This passing behavior between group members is particularly problematic in scenarios involving children. Passing by children is problematic because they can be more sensitive and unpredictable and require more space to ensure their safety and comfort. When a robot invades the personal space of a group, including children, it can lead to discomfort and safety issues. With STAC, on the other hand, the robot can regard both the personal space of the group members and the group space, which is the space between the group members. This allows the robot to choose a path that avoids disrupting the group’s interaction and moves without moving through the group.
The results of the experiments are detailed in Table 4. In the experiment with a group including a child, the Invasion Distance value was measured by adjusting the personal space required by the child from 1.7 m to 1.2 m, as opposed to the previous experiment. With a group of stationary adults, using the traditional method, the robot traveled at a minimum distance of 0.61 m from the person. However, with STAC, the robot traveled at a minimum distance of 1.10 m. This distance makes people more comfortable by moving through a path that largely avoids the group. Although STAC did not fully respect the personal space of 1.2 m for the group, choosing a path that avoided the group instead of driving through resulted in a safer and more comfortable interaction. In the scenario with a group including a child, there may not be much difference between the traditional method and STAC in the minimum distance (MD), measuring 0.44 and 0.62 m, respectively. This is because group space is adapted between group members, and each individual is given their own space. However, the significant difference is that STAC travels around the group space rather than having the robot pass directly between people. This approach makes the child and group members feel more comfortable with the navigation than having the robot pass directly between group members, especially for a group with a child.
In the moving adult group experiment, when using the traditional costmap, the robot chose a path that went through the group, avoiding people who were closer to the robot. The measured MD and ID in this scenario were 0.51 m and 1.14 m, respectively. However, when applying our proposed STAC, the robot was proactively aware of this and modified its path to avoid it, with a costmap that considered personal and group space. The MD was found to be 0.63 m, although it did not respect the human’s personal space of 1.2 m. This is a significant improvement as it allows the human to move without having to modify his/her path, considering that it is not always possible to ensure absolute personal space in a dynamic environment. When a child and an adult were moving together, the robot, using only the traditional costmap, traveled between them and maintained a minimum distance of 0.33 m from the person. However, the MD increased to 0.86 m for the robot using STAC. These results show that STAC is more respectful of personal space than traditional methods and is particularly effective at ensuring a safe distance between robots and people in groups containing a child, meaning that the robot will follow a safer trajectory around the child. Thus, STAC shows more respect for human interaction and works to provide driving trajectories that ensure the safety and comfort of all group members.

5. Conclusions

In this paper, we propose a new framework, the ’Social Type-Aware Navigation Framework’, for robots to classify individuals and groups within a social space and navigate while respecting the appropriate social distance. This framework utilizes sensors commonly mounted on robots, such as a camera and LiDAR, to detect and track people in real time, classify them by social type, and apply social distances for each type to make people feel comfortable and safe with the robot’s driving. The ‘Social Type-Aware Costmap’ proposed in this research is developed as a compatible plugin to the ROS-based navigation stack, which can be easily integrated into various robot systems and path-planning algorithms. The robot can recognize the behavior patterns based on age and group and adopt suitable avoidance strategies for each situation to ensure the safety of the interaction. This is expected to increase the social acceptance of robots, providing a new perspective in the field of human–robot interaction, with a particular focus on interactions with children. By recognizing vulnerable subjects such as children and maintaining an appropriate social distance for them, this research provides a basis for a robot to be perceived as a social being. Future research will include the consideration of not only children but also other vulnerable individuals such as the elderly and wheelchair users. Additionally, it is important to consider the size of the robot as a significant variable in developing the costmap and planning avoidance strategies. While our current study focused on a relatively small robot, future research should explore how different robot sizes affect human–robot interactions and the effectiveness of the proposed navigation framework. The goal is to further enhance the social acceptability of robots by enabling them to recognize and act on a wider range of people. However, the definition of personal space can vary depending on cultural and individual differences, so this is an area for continuous research and development in human–robot interaction. This is a significant challenge in enhancing the ability of robots to understand and adapt to social and cultural contexts.

Author Contributions

Conceptualization, S.K.; Methodology, S.K.; Software, S.K.; Validation, S.K.; Formal Analysis, S.K. and S.Y.; Investigation, S.Y. and D.K. (Daewon Kwak); Data Curation, S.K. and D.K. (Daewon Kwak); Visualization, S.K.; Project Administration, S.K.; Experimental Assistance, S.Y., D.K. (Daewon Kwak), and Y.J.; Writing—Original Draft Preparation, S.K.; Writing—Review and Editing, S.K., S.Y., D.K. (Daewon Kwak), and D.K. (Donghan Kim); Supervision, D.K. (Donghan Kim). All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by the MSIT (Ministry of Science and ICT), Korea, under the Convergence security core talent training business support program (IITP2023-RS-2023-00266615) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation), the BK21 plus program "AgeTech-Service Convergence Major" through the National Research Foundation (NRF) funded by the Ministry of Education of Korea [5120200313836], Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. RS-2022-00155911, Artificial Intelligence Convergence Innovation Human Resources Development (Kyung Hee University)), and the Ministry of Trade, Industry and Energy (MOTIE), South Korea, under Industrial Technology Innovation Program under Grant 20015440, 20025094.

Institutional Review Board Statement

All experiments involving child participants in this study were approved by the Institutional Review Board (IRB) of Kyung Hee University (approval number KHGIRB-23-467).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent was obtained from the legal guardians of the child participants.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to thank the anonymous reviewers and academic editor for their comments and suggestions. We extend our special thanks to Boram and Jak for their participation in the experiments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Triebel, R.; Arras, K.; Alami, R.; Beyer, L.; Breuers, S.; Chatila, R.; Chetouani, M.; Cremers, D.; Evers, V.; Fiore, M.; et al. Spencer: A socially aware service robot for passenger guidance and help in busy airports. In Field and Service Robotics: Results of the 10th International Conference; Springer: Berlin/Heidelberg, Germany, 2016; pp. 607–622. [Google Scholar]
  2. Hellou, M.; Lim, J.; Gasteiger, N.; Jang, M.; Ahn, H.S. Technical methods for social robots in museum settings: An overview of the literature. Int. J. Soc. Robot. 2022, 14, 1767–1786. [Google Scholar] [CrossRef]
  3. Khatib, O. Real-time obstacle avoidance for manipulators and mobile robots. Int. J. Robot. Res. 1986, 5, 90–98. [Google Scholar] [CrossRef]
  4. Fox, D.; Burgard, W.; Thrun, S. The dynamic window approach to collision avoidance. IEEE Robot. Autom. Mag. 1997, 4, 23–33. [Google Scholar] [CrossRef]
  5. Berg, J.; Lin, M.; Manocha, D. Reciprocal velocity obstacles for real-time multi-agent navigation. In Proceedings of the 2008 IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, 19–23 May 2008; pp. 1928–1935. [Google Scholar]
  6. Kruse, T.; Pandey, A.K.; Alami, R.; Kirsch, A. Human-aware robot navigation: A survey. Robot. Auton. Syst. 2013, 61, 1726–1743. [Google Scholar] [CrossRef]
  7. Hall, E.T.; Birdwhistell, R.L.; Bock, B.; Bohannan, P.; Diebold, A.R., Jr.; Durbin, M.; Edmonson, M.S.; Fischer, J.; Hymes, D.; Kimball, S.T.; et al. Proxemics [and comments and replies]. Curr. Anthropol. 1968, 9, 83–108. [Google Scholar] [CrossRef]
  8. Neggers, M.M.; Cuijpers, R.H.; Ruijten, P.A. Comfortable passing distances for robots. In Social Robotics: 10th International Conference, ICSR 2018, Qingdao, China, 28–30 November 2018; Proceedings 10; Springer: Berlin/Heidelberg, Germany, 2018; pp. 431–440. [Google Scholar]
  9. Pacchierotti, E.; Christensen, H.I.; Jensfelt, P. Evaluation of passing distance for social robots. In Proceedings of the ROMAN 2006—The 15th IEEE International Symposium on Robot and Human Interactive Communication, Hatfield, UK, 6–8 September 2006; pp. 315–320. [Google Scholar]
  10. Sisbot, E.A.; Marin-Urias, L.F.; Alami, R.; Simeon, T. A human aware mobile robot motion planner. IEEE Trans. Robot. 2007, 23, 874–883. [Google Scholar] [CrossRef]
  11. Kirby, R.; Simmons, R.; Forlizzi, J. Companion: A constraint-optimizing method for person acceptable navigation. In Proceedings of the ROMAN 2009—The 18th IEEE International Symposium on Robot and Human Interactive Communication, Toyama, Japan, 27 September–2 October 2009; pp. 607–612. [Google Scholar]
  12. O’Neal, E.E.; Jiang, Y.; Franzen, L.J.; Rahimian, P.; Yon, J.P.; Kearney, J.K.; Plumert, J.M. Changes in perception–action tuning over long time scales: How children and adults perceive and act on dynamic affordances when crossing roads. J. Exp. Psychol. Hum. Percept. Perform. 2018, 44, 18. [Google Scholar] [CrossRef]
  13. Parr, M.N.; O’Neal, E.E.; Zhou, S.; Williams, B.; Butler, K.M.; Chen, A.; Kearney, J.K.; Plumert, J.M. How children judge affordances when walking and bicycling across virtual roads: Does mode of locomotion matter? Dev. Psychol. 2023, 56, 1098. [Google Scholar] [CrossRef] [PubMed]
  14. Walters, M.L.; Dautenhahn, K.; Koay, K.L.; Kaouri, C.; Boekhorst, R.; Nehaniv, C.; Werry, I.; Lee, D. Close encounters: Spatial distances between people and a robot of mechanistic appearance. In Proceedings of the 5th IEEE-RAS International Conference on Humanoid Robots, Tsukuba, Japan, 5–7 December 2005; pp. 450–455. [Google Scholar]
  15. Leigh, A.; Pineau, J.; Olmedo, N.; Zhang, H. Person tracking and following with 2D laser scanners. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 726–733. [Google Scholar]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  18. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  19. Levi, G.; Hassner, T. Age and gender classification using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 34–42. [Google Scholar]
  20. Foggia, P.; Greco, A.; Percannella, G.; Vento, M.; Vigilante, V. A system for gender recognition on mobile robots. In Proceedings of the 2nd International Conference on Applications of Intelligent Systems, Las Palmas de Gran Canaria, Spain, 7–9 January 2019; pp. 1–6. [Google Scholar]
  21. Khalifa, A.; Abdelrahman, A.A.; Strazdas, D.; Hintz, J.; Hempel, T.; Al-Hamadi, A. Face recognition and tracking framework for human–robot interaction. Appl. Sci. 2022, 12, 5568. [Google Scholar] [CrossRef]
  22. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  23. Ince, O.F.; Park, J.; Song, J.; Yoon, B. Child and adult classification using ratio of head and body heights in images. Int. J. Comput. Commun. Eng. 2014, 3, 120. [Google Scholar] [CrossRef]
  24. Lin, J.-M.; Lin, W.-L.; Fan, C.-P. Age group classifier of adults and children with yolo-based deep learning pre-processing scheme for embedded platforms. In Proceedings of the 2022 IEEE 12th International Conference on Consumer Electronics (ICCE-Berlin), Berlin, Germany, 2–6 September 2022; pp. 1–3. [Google Scholar]
  25. Ramirez, O.A.I.; Varni, G.; Andries, M.; Chetouani, M.; Chatila, R. Modeling the dynamics of individual behaviors for group detection in crowds using low-level features. In Proceedings of the 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), New York, NY, USA, 26–31 August 2016; pp. 1104–1111. [Google Scholar]
  26. Chatterjee, I.; Steinfeld, A. Performance of a low-cost, human-inspired perception approach for dense moving crowd navi-gation. In Proceedings of the 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), New York, NY, USA, 26–31 August 2016; pp. 578–585. [Google Scholar]
  27. Taylor, A.; Chan, D.M.; Riek, L.D. Robot-centric perception of human groups. ACM Trans. Hum. Robot Interact. (THRI) 2020, 9, 1–21. [Google Scholar] [CrossRef]
  28. Linder, T.; Breuers, S.; Leibe, B.; Arras, K.O. On multi-modal people tracking from mobile platforms in very crowded and dynamic environments. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 5512–5519. [Google Scholar]
  29. Dondrup, C.; Bellotto, N.; Jovan, F.; Hanheide, M. Real-time multisensor people tracking for human-robot spatial interaction. In Proceedings of the Workshop on Machine Learning for Social Robotics at International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21 May 2015. [Google Scholar]
  30. Linder, T.; Girrbach, F.; Arras, K.O. Towards a robust people tracking framework for service robots in crowded, dynamic environments. In Proceedings of the Assistance and Service Robotics Workshop (ASROB-15) at the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015. [Google Scholar]
  31. Arras, K.O.; Grzonka, S.; Luber, M.; Burgard, W. Efficient people tracking in laser range data using a multi-hypothesis leg-tracker with adaptive occlusion probabilities. In Proceedings of the 2008 IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, 19–23 May 2008; pp. 1710–1715. [Google Scholar]
  32. Jafari, O.H.; Mitzel, D.; Leibe, B. Real-time rgb-d based people detection and tracking for mobile robots and head-worn cameras. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 5636–5643. [Google Scholar]
  33. Eirale, A.; Martini, M.; Chiaberge, M. Human-centered navigation and person-following with omnidirectional robot for indoor assistance and monitoring. Robotics 2022, 11, 108. [Google Scholar] [CrossRef]
  34. Nonaka, S.; Inoue, K.; Arai, T.; Mae, Y. Evaluation of human sense of security for coexisting robots using virtual reality. 1st report: Evaluation of pick and place motion of humanoid robots. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA’04), New Orleans, LA, USA, 26 April–1 May 2004; Volume 3, pp. 2770–2775. [Google Scholar]
  35. Neggers, M.M.; Cuijpers, R.H.; Ruijten, P.A.; IJsselsteijn, W.A. Determining shape and size of personal space of a human when passed by a robot. Int. J. Soc. Robot. 2022, 14, 561–572. [Google Scholar] [CrossRef]
  36. Lu, D.V.; Hershberger, D.; Smart, W.D. Layered costmaps for context-sensitive navigation. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 709–715. [Google Scholar]
  37. Ngo, H.Q.T.; Le, V.N.; Thien, V.D.N.; Nguyen, T.P.; Nguyen, H. Develop the socially human-aware navigation system using dynamic window approach and optimize cost function for autonomous medical robot. Adv. Mech. Eng. 2020, 12, 1687814020979430. [Google Scholar] [CrossRef]
  38. Patompak, P.; Jeong, S.; Nilkhamhang, I.; Chong, N.Y. Learning proxemics for personalized human–robot social interaction. Int. J. Soc. Robot. 2020, 12, 267–280. [Google Scholar] [CrossRef]
  39. Bera, A.; Randhavane, T.; Prinja, R.; Kapsaskis, K.; Wang, A.; Gray, K.; Manocha, D. The emotionally intelligent robot: Improving social navigation in crowded environments. arXiv 2019, arXiv:1903.03217. [Google Scholar]
  40. Gines, J.; Martin, F.; Vargas, D.; Rodriguez, F.J.; Matellan, V. Social navigation in a cognitive architecture using dynamic proxemic zones. Sensors 2019, 19, 5189. [Google Scholar] [CrossRef] [PubMed]
  41. Kendon, A. Conducting Interaction: Patterns of Behavior in Focused Encounters; CUP Archive; Cambridge University Press: Cambridge, UK, 1990; Volume 7. [Google Scholar]
  42. Yang, F.; Peters, C. Social-aware navigation in crowds with static and dynamic groups. In Proceedings of the 2019 11th International Conference on Virtual Worlds and Games for Serious Applications (VS-Games), Vienna, Austria, 4–6 September 2019; pp. 1–4. [Google Scholar]
  43. Haarslev, F.; Juel, W.K.; Kollakidou, A.; Kruger, N.; Bodenhagen, L. Human-aware robot behavior in healthcare facilities. In Proceedings of the IROS 2020 Workshop: Robots for Health and Elderly Care: An Honest Discourse on the Gap between Research and Practical Applications, Las Vegas, NV, USA, 25–29 October 2020. [Google Scholar]
  44. Daza, M.; Barrios-Aranibar, D.; Diaz-Amado, J.; Cardinale, Y.; Vilasboas, J. An approach of social navigation based on proxemics for crowded environments of humans and robots. Micromachines 2021, 12, 193. [Google Scholar] [CrossRef]
  45. Cai, K.; Chen, W.; Wang, C.; Song, S.; Meng, M.Q.-H. Human-aware path planning with improved virtual doppler method in highly dynamic environments. IEEE Trans. Autom. Sci. Eng. 2022, 20, 1304–1321. [Google Scholar] [CrossRef]
  46. Forer, S.; Banisetty, S.B.; Yliniemi, L.; Nicolescu, M.; Feil-Seifer, D. Socially-aware navigation using non-linear multi-objective optimization. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–9. [Google Scholar]
  47. Truong, X.-T.; Ngo, T.-D. Dynamic social zone based mobile robot navigation for human comfortable safety in social environments. Int. J. Soc. Robot. 2016, 8, 663–684. [Google Scholar] [CrossRef]
  48. Truong, X.-T.; Ngo, T.-D. “To approach humans?”: A unified framework for approaching pose prediction and socially aware robot navigation. IEEE Trans. Cogn. Dev. Syst. 2017, 10, 557–572. [Google Scholar] [CrossRef]
  49. Kobayashi, Y.; Sugimoto, T.; Tanaka, K.; Shimomura, Y.; Arjonilla Garcia, F.J.; Kim, C.H.; Yabushita, H.; Toda, T. Robot navigation based on predicting of human interaction and its reproducible evaluation in a densely crowded environment. Int. J. Soc. Robot. 2022, 14, 373–387. [Google Scholar] [CrossRef]
  50. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. Yolov7: Trainable bag-of-freebies sets new state-of the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  51. Shao, S.; Zhao, Z.; Li, B.; Xiao, T.; Yu, G.; Zhang, X.; Sun, J. Crowdhuman: A benchmark for detecting human in a crowd. arXiv 2018, arXiv:1805.00123. [Google Scholar]
  52. Lin, J.-M.; Lin, W.-L.; Fan, C.-P. Deep learning based efficient pedestrian age group estimation and tracking techniques for autonomous mover. In Proceedings of the 2023 5th International Conference on Computer Communication and the Internet (ICCCI), Fujisawa, Japan, 23–25 June 2023; pp. 15–19. [Google Scholar]
  53. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Fluids Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  54. costmap_2d. Available online: http://wiki.ros.org/costmap_2d (accessed on 28 November 2023).
  55. Haarslev, F.; Juel, W.K.; Kollakidou, A.; Kruger, N.; Bodenhagen, L. Context-aware social robot navigation. In Proceedings of the 18th International Conference, ICINCO 2021, Lieusaint–Paris, France, 6–8 July 2021; pp. 426–433. [Google Scholar]
  56. SocialTypeAwareCostmap. Available online: https://github.com/suminkxng/SocialTypeAwareCostmap (accessed on 18 June 2024).
  57. Navfn. Available online: https://wiki.ros.org/navfn (accessed on 6 December 2023).
  58. dwa_local_planner. Available online: https://wiki.ros.org/dwa_local_planner?distro=noetic (accessed on 6 December 2023).
  59. Jackal UGV. Available online: https://clearpathrobotics.com/jackal-small-unmanned-ground-vehicle/ (accessed on 6 February 2024).
  60. PedsimROS. Available online: https://github.com/srl-freiburg/pedsim_ros (accessed on 14 February 2024).
  61. Helbing, D.; Molnar, P. Social force model for pedestrian dynamics. Phys. Rev. E 1995, 51, 4282. [Google Scholar] [CrossRef]
Figure 3. Human detection and tracking result in a Gazebo environment.
Figure 3. Human detection and tracking result in a Gazebo environment.
Sensors 24 04862 g003
Figure 4. Example of the human position detected using the RGB-D camera in the robot coordinate system.
Figure 4. Example of the human position detected using the RGB-D camera in the robot coordinate system.
Sensors 24 04862 g004
Figure 5. Example of the personal space around a human using an asymmetric Gaussian distribution.
Figure 5. Example of the personal space around a human using an asymmetric Gaussian distribution.
Sensors 24 04862 g005
Figure 6. The result of applying an age-based layer. The personal spaces are shown at different sizes for a moving adult, a moving child, and a stationary child.
Figure 6. The result of applying an age-based layer. The personal spaces are shown at different sizes for a moving adult, a moving child, and a stationary child.
Sensors 24 04862 g006
Figure 7. Asymmetric Gaussian distribution illustrating (a) two individuals’ personal spaces and (b) the integration of personal and group spaces.
Figure 7. Asymmetric Gaussian distribution illustrating (a) two individuals’ personal spaces and (b) the integration of personal and group spaces.
Sensors 24 04862 g007
Figure 8. A comparison of the results with and without the STAC.
Figure 8. A comparison of the results with and without the STAC.
Sensors 24 04862 g008
Figure 9. (a) The simulated robot model used in the experiment, (b) the real robot, and (c) the field of view (FoV) of the sensors mounted on the robot. The gray area depicts the range of the LiDAR sensor, and the white area depicts the camera sensor.
Figure 9. (a) The simulated robot model used in the experiment, (b) the real robot, and (c) the field of view (FoV) of the sensors mounted on the robot. The gray area depicts the range of the LiDAR sensor, and the white area depicts the camera sensor.
Sensors 24 04862 g009
Figure 10. Trajectory results of the robot and humans in three simulated passing scenarios displayed in Rviz.
Figure 10. Trajectory results of the robot and humans in three simulated passing scenarios displayed in Rviz.
Sensors 24 04862 g010
Figure 11. Real-world experiment scenarios for individual–robot interaction.
Figure 11. Real-world experiment scenarios for individual–robot interaction.
Sensors 24 04862 g011
Figure 12. Experimental results in the real world with individuals. The left side of the figure is a snapshot of the Rviz visualization. The first row shows using the traditional costmap; the second row shows using STAC.
Figure 12. Experimental results in the real world with individuals. The left side of the figure is a snapshot of the Rviz visualization. The first row shows using the traditional costmap; the second row shows using STAC.
Sensors 24 04862 g012
Figure 13. Real-world experiment scenarios for group–robot interaction.
Figure 13. Real-world experiment scenarios for group–robot interaction.
Sensors 24 04862 g013
Figure 14. Experimental results in real-world group interaction. The left side of the figure is a snapshot of the Rviz visualization. The first row shows using the traditional costmap; the second row shows using STAC.
Figure 14. Experimental results in real-world group interaction. The left side of the figure is a snapshot of the Rviz visualization. The first row shows using the traditional costmap; the second row shows using STAC.
Sensors 24 04862 g014
Table 1. Parameter settings.
Table 1. Parameter settings.
ParameterSimulationReal World
A150125
σ 0 0.80.5
σ G 0.71.0
k v 55
A G 170125
k G 1.51.3
Table 3. Individual experiment results.
Table 3. Individual experiment results.
ScenarioMethodNT [s]PL [m]MD [m]ID [m]
Stationary AdultTraditional18.796.920.663.16
STAC (Ours)21.067.091.092.80
Stationary ChildTraditional19.856.560.383.12
STAC (Ours)23.206.721.610.34
Moving AdultTraditional17.897.340.411.31
STAC (Ours)21.577.700.880.83
Moving ChildTraditional20.026.730.703.22
STAC (Ours)21.407.301.470.62
Table 4. Human group experiment results.
Table 4. Human group experiment results.
ScenarioMethodNT [s]PL [m]MD [m]ID [m]
Standing GroupTraditional17.756.710.612.19
STAC (Ours)21.277.361.100.58
Standing C-GroupTraditional18.207.300.444.30
STAC (Ours)22.757.680.623.23
Moving GroupTraditional20.167.050.511.14
STAC (Ours)21.157.590.630.97
Moving C-GroupTraditional20.327.010.332.68
STAC (Ours)23.757.940.860.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kang, S.; Yang, S.; Kwak, D.; Jargalbaatar, Y.; Kim, D. Social Type-Aware Navigation Framework for Mobile Robots in Human-Shared Environments. Sensors 2024, 24, 4862. https://doi.org/10.3390/s24154862

AMA Style

Kang S, Yang S, Kwak D, Jargalbaatar Y, Kim D. Social Type-Aware Navigation Framework for Mobile Robots in Human-Shared Environments. Sensors. 2024; 24(15):4862. https://doi.org/10.3390/s24154862

Chicago/Turabian Style

Kang, Sumin, Sungwoo Yang, Daewon Kwak, Yura Jargalbaatar, and Donghan Kim. 2024. "Social Type-Aware Navigation Framework for Mobile Robots in Human-Shared Environments" Sensors 24, no. 15: 4862. https://doi.org/10.3390/s24154862

APA Style

Kang, S., Yang, S., Kwak, D., Jargalbaatar, Y., & Kim, D. (2024). Social Type-Aware Navigation Framework for Mobile Robots in Human-Shared Environments. Sensors, 24(15), 4862. https://doi.org/10.3390/s24154862

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop