Next Article in Journal
The Use of Digital Color Imaging and Machine Learning for the Evaluation of the Effects of Shade Drying and Open-Air Sun Drying on Mint Leaf Quality
Previous Article in Journal
Special Issue on 3D Information Technologies in Cultural Heritage Preservation and Popularization—Motivations, Works Overview, and the Future
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hey Max, Can You Help Me? An Intuitive Virtual Assistant for Industrial Robots

1
Robotics and Automation Group, Department of Materials and Production, Aalborg University, Fibigerstraede 16, 9220 Aalborg, Denmark
2
Center for Industrial Production, Department of Materials and Production, Aalborg University, Fibigerstraede 16, 9220 Aalborg, Denmark
3
Department of Technology and Business, University College of Northern Denmark, Sofiendalsvej 60, 9000 Aalborg, Denmark
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(1), 205; https://doi.org/10.3390/app13010205
Submission received: 2 October 2022 / Revised: 18 October 2022 / Accepted: 18 December 2022 / Published: 23 December 2022
(This article belongs to the Topic Human–Machine Interaction)

Abstract

:
Assisting employees in acquiring the knowledge and skills necessary to use new services and technologies on the shop floor is critical for manufacturers to adapt to Industry 4.0 successfully. In this paper, we employ a learning, training, assistance-formats, issues, tools (LTA-FIT) approach and propose a framework for a language-enabled virtual assistant (VA) to facilitate this adaptation. In our system, the human–robot interaction is achieved through spoken natural language and a dashboard implemented as a web-based application. This type of interaction enables operators of all levels to control a collaborative robot intuitively in several industrial scenarios and use it as a complementary tool for developing their competencies. Our proposed framework has been tested with 29 users who completed various tasks while interacting with the proposed VA and industrial robots. Through three different scenarios, we evaluated the usability of the system for LTA-FIT based on an established system usability scale (SUS) and the cognitive effort required by the users based on the standardised NASA-TLX questionnaire. The qualitative and quantitative results of the study show that users of all levels found the VA user friendly with low requirements for physical and mental effort during the interaction.

1. Introduction

Digitalization is changing the manufacturing world. The exponential growth in digital technologies provides manufacturing companies with possibilities for introducing new products, processes, and services, many of which have the potential to be disruptive. Hence, the manufacturing industry has adapted to Industry 4.0 values and is slowly transitioning to the Industry 5.0 era [1]. However, due to the complexity of the manufacturing tasks and the speed of the technological development, the adoption of new technologies remains a grand challenge [2].
As the nature of many proposed solutions typically involves multidisciplinary activities, operators with years of experience and domain knowledge are needed. However, in modern industrial settings, such expertise is not always readily available, and new, more flexible approaches for knowledge acquisition and dissemination of learned experience should be utilized [3]. One revolutionary model used to assist workers in their learning and training is the learning, training, assistance-formats, issues, tools (LTA-FIT) model proposed by [4]. The model offers a flexible approach to assist the re-qualification and training of workers in new digital tools.
In this work, we expand the proof-of-concept we presented previously [5], in which we proposed a virtual assistant (VA) built on a natural language processing architecture used as a digital interpretation of the LTA-FIT model in a smart factory. The system provides additional instructions to facilitate a basic understanding of the steps of an assembly process of a mobile phone mockup, assists the workers through the necessary training process, and then answers workers’ task-related questions. Because human operators are at the center of the LTA-FIT model, we performed an extensive user study to evaluate the system’s usability and the users’ acceptance, comfort, satisfaction, and task load. We used the proposed framework to complete an assembly task by using the eighth iteration of our autonomous industrial mobile manipulator (AIMM), named Little Helper 8 (LH8) (Figure 1). Little Helper is a series of mobile manipulators with an integrated industrial robot arm and mobile platform to offer flexible solutions to industrial settings [6]. The eighth iteration consists of a Franka Emika Panda collaborative robot mounted on a MiR200 omnidirectional mobile platform.

Contribution

In this study, we design and create a language-enabled VA for human–robot interaction in LTA-FIT of manufacturing tasks that enhance the user experience while maintaining a high level of system usability. Our specific contributions are:
  • a novel VA trained on the state-of-the-art bidirectional encoder representations from transformers (BERT) [7] model with a high human intent-recognition capability focusing on the industrial robots and satisfies downstream assistance needs of factories for employees to successfully obtain the acquired knowledge and skills for manufacturing tasks;
  • three common manufacturing scenarios are selected and analyzed in a genuine production setting with an emphasis on each level of LTA-FIT. To our knowledge, this is the first time an industrial-oriented VA has been integrated into an industrial setting with the goal of providing LTA-FIT for shop floor workers; and
  • a user study of human–machine interaction with both the VA and the robot. This analysis evaluates the system’s usability, the perceived comfort, and mental load while working on the task, and the comfortability, perceived competence, performance expectancy, and effort expectancy.
In the rest of the paper, we present an overview of related literature in Section 2, the architecture of the proposed framework in Section 3 and the selected LTA-FIT scenarios in Section 4. Section 5 and Section 6 present the experimental setup and analyze the obtained results while we reflect on the results and conclude the paper in Section 7.

2. Related Work

As we discussed before, the transition to the so-called Industry 5.0 era requires specialists and domain experts to be able to learn new concepts quickly and improve their qualifications intuitively [8]. A general term used often for such workplace training activities is work-based or work-integrated learning [9]. It is related to all learning generated directly through workplace considerations and concerns. It is typically an unplanned and informal process and either appears organically through regular development and problem-solving at work or as part of workplace training or coaching in response to identified work-related issues [10]. It has been proven to be more effective compared to traditional classroom teaching, mainly when it tackles real work issues, and it can improve professional work readiness, self-efficacy, and team skills [11].
In the context of industrial automation and manufacturing processes, the learning principles of work-based learning have been incorporated effectively into the concept of learning factories [12,13]. They constitute a paradigm of close collaboration between industry parties and universities to offer a flexible industrial environment where the involved stakeholders can introduce and test artificial intelligence (AI) techniques [14], swarm production concepts [15], and virtual recommissioning methods [16].
However, although learning factories can support proof-of-concept integration, they do not scale up adequately to real industrial challenges, as human operators, who should work in close collaboration with various machines or robots, lack the required qualifications [17]. Therefore, several employee qualification and competency-building models have been developed whereby a variety of personal, social, action-related, and domain-related competencies can be improved [18,19].
In our work in particular, where human–robot collaboration is the focus, we employ the LTA-FIT qualification model, and based on the lessons learned from [20] we propose a framework for a language-enabled virtual assistant (VA)-integrated system with the latest iteration of our AIMM [21] and deployed it in our own learning factory [22].
In recent years, voice-enabled interaction and VAs, in particular, have become an integral element in the interaction of humans with machines in various domains [23]. In the area of teaching and learning, Prgorskly et al. [24] proposed a virtual learning assistant as an assessment tool by which to support online learners’ self-regulation in online learning, whereas Preece et al. [25] investigated the effectiveness of a controlled natural language-based conversational interface for human–machine interaction regarding locally observed collective knowledge. In the manufacturing domain, Barbosa et al. [26] created a VA, which supported the training process of the technical personnel in industrial plant operators, Casillo et al. [27] built a chatbot for the training of employees in an Industry 4.0 scenario, and Zimmer et al. [28] developed a VA to assist the ramp-up process of an assembly system. In our previous work, we have presented a VA named Bot-X to assist the workers in handling a variety of complex manufacturing tasks, e.g., order processing, and production execution [29]. Naturally, VAs, dialogue systems, and conversational agents have also facilitated scenarios wherein humans work closely with robots. Maksymova et al. [30] used several models for voice control of an industrial robot in the context of a collaborative assembly whereas Bingol and Aydogmus [31] evaluated deep neural networks for the classification of commands in a natural speech-recognition system for a collaborative robot task.
Despite the vast amount of research that has been performed in all these domains, VAs are still not readily accepted as the preferred means of interaction with machines or robots. Many researchers have concluded that several challenges remain to be resolved. Russo et al. [32] focused their research on using dialogue systems and conversational agents for patients with dementia, where they concluded that the delay of the robot in responding to patients increased the cognitive load of the patients. Similarly, Bernard [33] and Rawassizadeh et al. [23] manifest that the use of VAs increases the cognitive requirements of the users. Additionally, Skantze [34] evaluated conversational agents in human–robot interaction (HRI) and concluded that natural timing in conversation systems is one of the essential factors for fluent human–human interaction, but it is also a challenging feature to adopt in VAs. He also emphasized that broken words and unclear sentences often hinder the performance of the VA and consequently increase the discomfort of the users.
Therefore, in this paper, we have a strong interest in evaluating our VA by human operators in an industrial scenario. We have enhanced our VA with natural timing capabilities, and we performed a user study focused on tracking the comfortability of the users and the usability and cognitive requirements of our overall system. Based on the learnings from other researchers [35,36,37], and our lessons learned from standardising HRI and human–robot collaboration (HRC) experiments before [38,39], we evaluated the system’s usability with the system usability scale (SUS) [40], the perceived mental load while working on the task, based on the NASA task load index (TLX) [41], the perceived competence, performance expectancy, and effort expectancy metrics inspired by [35], and the comfortability scale based on our uniquely designed questionnaire.

3. System Overview

This section describes the architecture of the proposed natural language-enabled VA, named Max. A preliminary development step is presented in our previous work [21]. Following the client-server (CS) style architecture, language and robot services are hosted on the Max server side. The Max client is composed of Microsoft cognitive speech service and the robot control module. RESTful API requests support the communication between the Max server and Max client. The Max client provides a voice interface (i.e., microphone and speaker) to interact with the human operator and leverages the RESTful APIs to communicate with the shop floor robots. The overall architecture is showcased in Figure 2 while a description of each component is presented in the remainder of the section.

3.1. Language Services

Intent Recognition. Two services are provided for human-intent recognition—rule-based keyword extraction and neural network-based intent recognition. Due to its simplicity and quick response time, learning and training scenarios rely heavily on the keyword-extraction approach. Max, on the other hand, should be able to precisely grasp the operator’s intent and directions in order to assist with everyday logistic activities in the assistance scenario.
We fine-tuned the pretrained basic BERT model (see language service part of Figure 3), which provides contextualized sentence representation and can learn the meaning of the words in the given context to predict the operator’s intent and key slots from the operator’s utterance. The model has 12 layers, 768 hidden units, 12 attention heads, and 110 M parameters fine-tuned with a feed-forward neural network (FFNN) layer and Softmax as the activation function to normalize the output of the model. A series of sequential operations (masking (15% of the words), tokenisation, and embeddings) are applied to the user’s utterance to generate the input representation. A special symbol, [CLS], is added as the first token to indicate the beginning of the user’s utterance. The embedded utterances are passed through 12 transformer encoders. The last FFNN takes the results of the BERT model as input and outputs a possible user’s intent ( X i n t e n t ) and slots ( X s l o t ). The model is trained and validated on a dialogue dataset containing the task-related conversations, and it achieves intent accuracy of 97.7% and slot F1 of 96.8% [5]. Compared with the rule-based keyword extraction, the pretrained BERT is a context-dependent model that can distinguish the word’s meaning in the given context. Therefore, operators may give the same command by using different words, e.g., “Hey Max, I need your help” or “Max, can you give me a hand?” when they need assistance without mentioning the exact keywords.
Dialogue State Tracking. A manufacturing task can take several steps to complete. Therefore, a VA may require several rounds or turns of conversation to obtain the necessary slots to complete the task. In this module, a JavaScript object notation (JSON) file called dialogue state rule (DSR) is predefined to specify all the slots (the optional slot marked with “optional”, otherwise marked as “Yes”) for completing each task. Max verifies and updates the state based on the requested slots. The dialogue continues until Max receives all the required slots.
Response Generation. Max’s response generation follows the template-based approach, i.e., mapping the nonlinguistic input (e.g., slots values from the user’s utterance) to the linguistic surface structure [42]. Furthermore, we introduce human–human conversation strategies in the interaction with the VA and the industrial robots. In this module, two dialogue strategies, namely lexical-semantic strategy and general diversion strategy [43], are integrated so Max can increase the task-completion rate while improving the user engagement. High engagement is also observed from active participation in human–human collaboration, e.g., taking the initiative in a task and smoothly switching the topic if it is not relevant to the current task. In this video (http://y2u.be/lYnh2cOeeE0, accessed on 1 October 2022), we demonstrate Max’s ability to initiate a request to the operator and start working with manufacturing tasks at the appropriate time, providing suggestions to the operator if the requested task cannot continue. Therefore, Max can complete the requested tasks and support active dialogues to enhance the user experience.

3.2. Robot Service

Grounding. Natural language grounding is essential for matching a word to the appropriate representation in terms of terminology and relationship in the learning and training scenarios. At the same time, it is responsible for assigning a vocal command to the associated robot’s action representation in the assistance scenario. To support learning scenarios, a knowledge graph (KG) [44], which stores interlinked descriptions of phone production, is designed based on the Neo4j system (https://neo4j.com/, accessed on 1 October 2022). It defines the entities (e.g., materials of producing a phone, operation process), attributes (e.g., color, size), and relationships (e.g., pair, need).
A knowledge graph is defined as
K G : = ( N , E , a N , r E , s , t ) ,
where
N ( K G ) is a finite set of nodes (i.e., entities)
E ( K G ) = { ( p , q ) N × N } where p , q N is a finite set of edges (i.e., relationship)
a N ( K G ) : N a N is a labeling function for nodes attribute
r E ( K G ) : E r E is a labeling function for relationship and
s ( K G ) , t ( K G ) : E N attaches the edge to its source and target node.
The grounding process can be explained with the following process. Considering that we have a corpus of operator’s utterance U which contains n extracted keywords K = { k 1 , . . . , k n } , we can find a partial function f to map the extracted keywords from operator’s utterance to their representations in the graph, f : K K G . The keywords refer to a predefined set of robot services and skill categories, which are stored as a JSON file in the database. The JSON file also lists the supported robot services and skills for the training and assistance scenarios.
Algorithm 1 provides an example of how the system interprets the operator’s dialogue related to the task “delivery_package” to provide the core skill parameters, e.g., “destination” for a package-delivery task.
Algorithm 1 Robot skill and response computation.
  1:
Input: the dialogue history, U s e r 0 , S y s A c t 0 ,..., U s e r i , and DSR
  2:
Output: the R o b o t _ S k i l l i , the R o b o t _ S k i l l _ P a r a m e t e r i and r e s _ p o n s e i
  3:
extract the current user intent, i n t e n t i , and previous intent, i n t e n t i 1 , from U s e r i and U s e r i 1 respectively
  4:
if i n t e n t i i n t e n t i 1 then
  5:
    Assign n u l l to the R o b o t _ S k i l l i and i n c o n s i s t e n t _ i n t e n t to r e s _ r e f e r e n c e i
  6:
    return
  7:
else
  8:
    Assign i n t e n t i to the R o b o t _ S k i l l i
  9:
    for Every dialogue D ( U s e r 0 , . . . , U s e r i 1 )  do
10:
        save the obtained slots of each turn to o b t _ s l o t s
11:
    end for
12:
end if
13:
extract the slots, s l o t s i , from current user utterance, U s e r i
14:
for Every s l o t s l o t s i  do
15:
    if  s l o t o b t _ s l o t s  then
16:
        update the o b t _ s l o t s with the new value of the s l o t
17:
    else
18:
        save the s l o t to o b t _ s l o t s
19:
    end if
20:
end for
21:
extract all the required slots ( r e q _ s l o t s ) from DSR regarding the requested i n t e n t i from U s e r i
22:
for Every s l o t r e q _ s l o t s  do
23:
    if  s l o t o b t _ s l o t s  then
24:
        Assign r e q u e s t e d _ s l o t to the R o b o t _ S k i l l _ P a r a m e t e r i and r e s _ r e f e r e n c e i
25:
        return
26:
    end if
27:
end for

3.3. Voice Interface

A voice-enabled interface is designed for Max’s client to enable efficient communication with low hardware requirements (i.e., headphones with a built-in microphone and Bluetooth wireless connection). In this module, we chose the Microsoft cognitive speech services, speech-to-text and text-to-speech, to enable Max’s client voice interface. The operator’s utterance transcripts are organized into a RESTful style HTTP request and sent to Max’s server. The response message is then extracted from the server’s reply (a JSON string), and a natural-sounding voice is generated. Such a voice interface frees the operator’s hands while providing natural communication for efficient HRI on the shop floor.

3.4. Web Interface

An interactive web interface is being developed to enhance the operator’s experience with LTA-FIT through Max. The interface is designed as a web interface on any web browser-enabled device, such as a mobile phone, tablet, or PC.
Figure 4 illustrates the main web interface. The interface’s message panel shows the operator’s real-time dialogue with Max. It enables the operator to keep track of the conversation history and verify that the Max accurately transcribes spoken orders and whether the Max delivers relevant responses or not. A robot services panel is included to tell the operator about the services that the requested robot supports. Additionally, a system status panel provides real-time information regarding activities or systems, such as network connections and the status of running processes.

3.5. Robot Control Module

Max is designed to be agnostic of robot hardware and aims to support various kinds of industrial robots, from mobile robots to industrial manipulators. Two subcomponents, a robot service management (RSM) component and a robot service execution (RSE) component, are implemented on the robot control module. RSM periodically synchronizes the local robot services and skills with Max’s server to update the robot services/skills registered on the local client side and the robot-controlling scripts. RSE invokes the corresponding robot control script, which specifies robot control functionalities (e.g., package delivery) and communication protocols for shop floor robots. A JSON file is maintained on RSM to ground the operator’s verbal commands (including operator’s intent and slots) to the requested robot service/skill. For example, “warehouse” is grounded to the coordinate “x”:, “y”:, “z” of the location of the mobile robot in the 2D map of the warehouse. The grounding results are used to match the corresponding robot control function and set up the input parameters.

4. Selected LTA-FIT Scenarios

In order to explore the capabilities of our proposed language-enabled VA, we examined three LTA-FIT scenarios. Here, our focus is to teach the production terminology to the operators, train them to assemble mock-up mobile phones, and assist them with the internal logistics.

4.1. Learning Scenario

According to the definition of the LTA-FIT model, learning focuses on providing the appropriate basic knowledge to the workers via instructions. Therefore, in the first scenario, the VA is leveraged to teach the worker the production terminology and the process of the assembly of a mock-up mobile phone (shown in Figure 5).
Such knowledge consists of a list of the raw materials, components, the respective quantities required to produce a phone, and the corresponding sequence of the phone-assembly task. Usually, such information is shared at the beginning of the training of new operators or during the introduction of new products in production.
In this case, the VA is expected to perform as a question answering (Q&A) system. To this end, we designed a repository of logical connections between the parts of the mock-up phone assembly with the means of the knowledge graph. The knowledge graph contains the core entities (e.g., printed circuit board (PCB), phone cover, fuse), attributes (e.g., the color of the cover), and the relationships (e.g., the order of the operations) regarding the mock-up phone assembly task (see Figure 6 [5]). The VA is able to understand the operator’s verbal requests (e.g., “how many processes are needed in a phone assembly?”), retrieve the relevant answer according to the knowledge graph, and generate responses to the operator (e.g., “There are eight operations in total.”).

4.2. Training Scenario

To solidify the gained knowledge from the learning level, the second scenario is selected for training operators to assemble the mock-up mobile phone. A live demonstration performed by an experienced and skilled operator is the most common way to carry out such a task in current manufacturing environment with a follow-up Q&A or hands-on learning session afterward. In this scenario, the VA performs as a task-oriented dialogue system. Based on the dialogue with the operator, the VA controls the industrial mobile manipulator and demonstrates the required steps for the assembly of the mock-up phone. The robot manipulator is programmed to perform the mock-up phone assembly automatically. However, the continuous assembly process may be challenging for operators to follow, and the correct sequence of the operations could be difficult to remember. Therefore, in addition to the entire demonstration, the VA is able to guide the operators throughout all the steps of the hands-on learning session.

4.3. Assistance Scenario

The third scenario targets the top level of the LTA-FIT, i.e., assistance. The purpose is to evaluate the performance of the VA when it comes to assisting the operators’ daily work and, specifically, internal logistics. In general, internal logistics focus on internal supply processes, transportation of materials and tools, and cargo distribution. Such tasks may require a mobile robot to deliver the goods to a location within the organization. A fleet manager or web application for robot control usually comes as a package with the chosen mobile robot to assist in scheduling and tracking the tasks. However, it could have a steep learning curve for new or inexperienced operators. To this end, we tested the VA on two practical assistance cases, including package delivery and system status checking.
The VA is expected to obtain the core information (e.g., a package delivery task may require the destination of the delivery and a recipient) through a dialogue with the operator. In such a case, a standard wireless headset is the only extra device we need to maintain the communication between the operator and VA. The operator’s verbal command is transmitted through the headset to the VA. Based on the needs of these two cases, real-time assistance in manufacturing tasks requires that the VA can understand the operator’s utterance with high accuracy and low latency. Therefore, in contrast to the learning and training scenarios, the assistance scenario focus on tracking the VA’s overall performance based on four quantitative metrics, i.e., intent error rate (IER), slot error rate (SER), task success rate (TSR), and average communication time (ACT).

5. Experimental Setup

The industrial feasibility of the three LTA-FIT scenarios was evaluated based on a set of experiments designed to assess the interactions of the human operators with the robot and the VA. The participants were expected to complete a task during the experiments while interacting with the system. The tasks for the learning and training scenarios were focused on the assembly of the mock-up mobile phone (shown in Figure 5), whereas the assistance scenario focuses on an internal logistic task. The list of identified tasks is presented in Table 1).
The purpose of our experiments was twofold: (1) exploration of the system’s behavior when used by a diverse user group; (2) collection of suggestions for improvement on the usability, workload, and overall human–machine interaction. Therefore, we were interested in the following research question: How can the adoption of a natural language-enable virtual assistant for LTA-FIT scenarios provide an easier implementation of Industry 4.0 principles in factories?

5.1. Evaluation Metrics

To assess and analyze the interaction between the VA, the robot, and the human operators, subjective and objective metrics were used for the examined scenarios. All participants used a postexperiment questionnaire to rate their interaction with the system anonymously in the learning and training scenarios. This questionnaire was built with standardized questions to evaluate the system’s usability, the perceived mental load while performing the task with the system and subjective metrics which cover a comfortability scale, the perceived competence, the performance, and effort expectancies. As mentioned in Section 4.3, four quantitative metrics were used to track the performance of the VA in the assistance scenario. These four metrics can assess Max’s ability to recognize the operators’ verbal commands, perform tasks successfully, and latency during the interaction with the users.

5.1.1. Metrics Used in the Learning and Training Scenarios

System Usability Scale. The SUS is a ten-item survey where the participants score the experiment on a five-point Likert scale (from strongly disagree to strongly agree). Afterward, an overall score of the system can be used to assess its usability of the system. The SUS score is calculated by following the instructions in [45], where each item is attributed a score from 0 (strongly disagree) to 4 (strongly agree). Therefore, a participant’s score can vary within the range of 0–100 and in our experiments, and we used the grouping proposed by [35] as: “80–100, users liked using the system”; “60–79, users accept the system”; “0–59, the users did not like to use the system”.
NASA-TLX questionnaire. The NASA-TLX questionnaire is commonly used to assess the mental workload that users experience while interacting with a system. It includes six subscales that the participants rate individually, i.e., mental demand, physical demand, temporal demand, performance, effort, and frustration [41,46].
Furthermore, the NASA-TLX questionnaire often employs a 20-point or 100-point scale for participants to grade the experiment. However, because there are many areas for the participant to score, there is a possibility for additional frustration while completing the questionnaire. Therefore, several researchers [47,48] support the idea that having a five-point format in the scale design could reduce the frustration level of the users and could thereby increase the response rate. Consequently, we used a five-point Likert scale format in our questionnaire to increase the quality of the responses.
Similarly, we divided the last subscale (frustration) into categories since the original question contains numerous adjectives that might be deemed synonyms. The new categories we used are: stressed, insecure, and annoyed. These are evaluated in relation to both the interaction with the VA and the robot system overall.
Subjective metrics. In addition to the objective and standardized SUS and NASA-TLX scales, we deemed it necessary to utilize some of the subjective metrics proposed by [35] and add our own comfortability scale. Similar to SUS, the participants score using a five-point Likert scale in a postexperiment questionnaire. These subjective metrics cover the following elements:
  • The comfortability scale measures the participant’s level of comfort with the system at the beginning and conclusion of the interaction.
  • Perceived competence refers to the participant’s conviction that they can complete the task with the help of the system.
  • Performance expectancy is the degree to which participants feel that utilizing the system would help them perform better in a work environment.
  • Effort expectancy defines the degree of easiness involved with using the system.

5.1.2. Metrics Used in the Assistance Scenario

Intent Error Rate (IER). This is the rate where the model cannot predict the requested intent of the user correctly. For instance, the requested intent is PACKAGE_DELIVERY, but the predicted intent from the BERT model might be POSITIONCHECK. IER is obtained by calculating the proportion of it occurring in all experiments.
Slot Error Rate (SER). This is the rate of incorrect slot values, i.e., the BERT model selects an incorrect value for the slot, and slot detection fails, i.e., the model cannot recognize the slot. Similar to IER, the number of occurrences of failing to recognize slot values is used to calculate the SER.
Task Success Rate (TSR). This corresponds to Max’s ability to retrieve all the required information from operator’s verbal commands to complete the task without encountering any intent or slot error problem.
Average Communication Time (ACT). This is the required time for a task to be completed on average. The average communication time equals the meantime from the beginning of the conversation to its end. In general, high IER or SER requires more communication time.

5.2. Physical Setup

The experiments were conducted in our learning factory at Aalborg University [22]. Figure 7 provides a schematic representation of the physical setup, together with a visual representation of the placement of the phone components on the robot table. A camera and an external microphone were placed 2.2 and 1.6 meters away from the robot table, respectively, to capture the participants’ interactions for further analysis.

5.3. Evaluation Procedure

At the beginning of the experiment, the users were introduced to the overall goal and purpose of the user studies. A short introduction to the scope of the survey was given, along with a brief explanation of the task at hand. Afterward, the users were introduced to the methods they could choose to interact with the system, and they were informed that the whole experiment would be recorded for archival purposes. To conduct the experiments, participants were asked to randomly choose a task and initiate a dialogue based on sample dialogues that were distributed to them. A sample task specification and the corresponding dialogue example of the phone assembly used for the training scenario are shown in Figure 8a, where the user requests from the system to demonstrate how to assemble the phone. In Figure 8b, the VA asks the user to verify that all components are in their default places before starting the demonstration.
In total, 29 people took part in the user study. Based on their demographics shown in Figure 9, they represent the age group of 20–25 years old primarily, they are mainly users who identify as males (72.4%), and 75.9% have an IT/engineering background. Moreover, 72.4% of all users had interacted with an industrial robot before, whereas almost 21% had previous interactions with a VA.

6. Experimental Results

6.1. System Usability Scale

In general terms, the users found the system easy to use, and they were satisfied with the interaction with the VA and the industrial manipulator. In our experiments, the SUS score varied between 50.0 and 90.0. The average score in the system usability scale was 68.53, which, as we mentioned in Section 5.1, means that the users accept the system. Obviously, they did not have an enthusiastic response to the system for multiple reasons that we explain in Section 7.

6.2. NASA-TLX

In terms of the mental workload that the users had to overcome while using the system, the obtained results with NASA-TLX showcase that the system requires low mental demand (MD) while the users can achieve high performance with minimal effort. Figure 10 depicts the obtained scores for each subscale of the NASA-TLX questionnaire.
As expected, physical demand (PD) scored the lowest with 1.31 because no physical interaction with the industrial manipulator was needed to complete the required tasks. At the same time, temporal demand (TP) also scored low with 1.41 because there were no time limits or time requirements for the completion of the tasks. Furthermore, it is important to mention that these scores are related to the evaluation of the overall system. However, the overall frustration subscale is divided into three categories, i.e., stressed, insecure, and annoyed, and evaluated for Max and the industrial manipulator separately. In the box-plot shown in Figure 11, we can see that the overall frustration is low for both systems. However, the annoyed category scores higher than expected for the VA. The reason for this reaction from the users were the network delays and lack of quick feedback from the VA when multiple queries took place during the experiments.

6.3. Subjective Metrics

In a similar fashion, as in the SUS and NASA-TLX metrics, the scores collected for the subjective metrics proved that the system was intuitive to use. More specifically, Figure 12 visualizes the mean score for each metric, where higher scores mean better evaluation of the system. Based on the scores obtained in the comfortability scale (CS), we can safely state that the users found the system comfortable to use, whereas by looking at the scores of perceived competence (PC), they were confident in their competencies to complete the tasks with the help of the system. At the same time, the high score of effort expectancy (EE) proves that the general impression of the users was that the system is easy to use; however, based on the score of performance expectancy (PE), there were still users who were not convinced that the proposed system would assist them adequately in the work environment.

6.4. Quantitative Performance Results

In order to assess the individual performance of the VA in the tasks of the assistance scenario (see Table 1), we performed each task 30 times. The evaluation was divided considering two different areas of interest; the language area and the task area. In the former area, we assessed how well Max can recognize the users’ intent and slot value during the dialogue, whereas in the latter we measured the rate where the VA could execute the requested tasks successfully and the average communication costs. As Table 2 shows, the VA managed to recognize the operator’s intent with minimum IER but struggled with recognizing the required slot values. Due to the complexity of the dialogues and the ambient noise of an industrial shop floor, the VA completed the two tasks with TSR of 0.50 and 0.76, respectively, whereas it required 31.08 and 15.28 s on average to complete the communication.

7. Discussion & Conclusions

In this work, we demonstrated a natural language-enabled VA, integrated with a robot platform to conduct multiple learning, training, and assistance scenarios in a real industrial environment. In an expansion of our prior work [5], we conducted user studies with an emphasis on user feedback on satisfaction and usability of our system.
The user study proved that adopting the proposed VA could provide an easier implementation of Industry 4.0 technologies in factories and support a more intuitive interaction with intelligent systems and robots. In particular, the system scores high in usability, user satisfaction, and comfortability metrics whereas it requires low cognitive effort from the vast majority of the users. At the same time, the VA alone performs with a high success rate and handles complex queries with minimum error rates.
Naturally, there are several lessons learned and areas for improvement. As many users in the user studies originated from a diverse geographic range, they spoke with various accents and dialects during their interaction with the system. This fact created an unexpected challenge for the VA to understand the verbal commands of certain participants. As a side effect, the VA would occasionally perform an incorrect command or would request the participant to repeat the instruction for clarification multiple times leading to higher annoyance and mental demands, as shown in Figure 11. However, toward the conclusion of the experiments, the participants appeared to have an easier time interacting with the system due to their increased familiarity with the system’s behavior. Additionally, most participants expressed satisfaction with the system, with many of them stating that it was “fun and intuitive” and that they would be interested in working with such a system in their workplace.
A significant area for improvement in our future work is the management of ambient noise. Further investigation into noise-suppression methods of steady-state background noise (e.g., ventilation, production) in the workplace can significantly improve the speech recognition accuracy, resulting in a higher task success rate and shorter communication times. Moreover, the addition of speech-enhancement techniques can also be essential to alleviate issues with various dialects and decrease the accent recognition error rate.
Despite the fact that the proposed user study confirmed the satisfaction and usability of the VA, the population of sampled users is limited, and the majority of them are lab engineers, researchers, and students with a robotics background. Future plans include organizing a large-scale user study with our industry partners. Together with the shop floor experts, we will specify their LTA tasks and collect and annotate a new dialogue corpus for training the BERT model in order to implement and assess the VA in the company. Additionally, the robot control algorithms will be adjusted based on the robots in use. Employees and shop floor specialists will be invited to evaluate the VA’s generality.
The source code of the proposed VA and the supplementary material of the user study are accessible at https://github.com/lcroy/Virtual-Assistant-Max, accessed on 1 October 2022) to support the reproducibility of the proposed framework.

Author Contributions

Conceptualization, O.M., C.L. and D.C.; methodology, C.L. and D.C.; algorithm, C.L. and D.P.; experiments, D.P. and C.L.; research, C.L. and D.C.; writing—original draft preparation, C.L., D.C. and D.P.; writing—review and editing, C.L., D.C., D.P., A.K.H., S.B. and O.M.; supervision, D.C. and C.L. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to acknowledge support by EU’s SMART EUREKA programme under grant agreement S0218-chARmER, Innovation Fund Denmark (Grant no. 9118-00001B), and the H2020-WIDESPREAD project no. 857061 “Networking for Research and Development of Human Interactive and Sensitive Robotics Taking Advantage of Additive Manufacturing—R2P2”. The research was also funded by the European Regional Development Fund and University College of Northern Denmark.

Institutional Review Board Statement

This experiments were reviewed and approved by the Ethical Committee of Faculty of the Engineering and Science, Department of Materials and Production, Aalborg University.

Informed Consent Statement

All participants signed an informed consent form to take part in this study.

Data Availability Statement

Not Applicable.

Acknowledgments

Authors wish to express their gratitude to all groups of users who graciously volunteered to be part of the testing phase of the prototype.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACTAverage Communication Time
AIArtificial intelligence
BERTBidirectional Encoder Representations from Transformers
CSClient-Server
CSComfortability Scale
DSRDialogue State Rule
EEEffort Expectancy
FFNNFeed-Forward Neural Network
HRCHuman-Robot Collaboration
HRIHuman-Robot Interaction
IERIntent Error Rate
JSONJavaScript Object Notation
KGKnowledge Graph
LH8Little Helper 8
LTA-FITLearning, Training, Assistant-Formats, Issues and Tools
MDMental Demand
PCPerceived Competence
PCBPrinted Circuit Board
PDPhysical Demand
PEPerformance Expectancy
Q&AQuestion Answering
RSERobot Service Execution
RSMRobot Service Management
SERSlot Error Rate
SUSSystem Usability Scale
TLXTask Load Index
TPTemporal Demand
TSRTask Success Rate
VAVirtual Assistant

References

  1. Xu, X.; Lu, Y.; Vogel-Heuser, B.; Wang, L. Industry 4.0 and Industry 5.0—Inception, conception and perception. J. Manuf. Syst. 2021, 61, 530–535. [Google Scholar] [CrossRef]
  2. Wang, L. A futuristic perspective on human-centric assembly. J. Manuf. Syst. 2022, 62, 199–201. [Google Scholar] [CrossRef]
  3. Adam, C.; Aringer-Walch, C.; Bengler, K. Digitalization in Manufacturing–Employees, Do You Want to Work There? In Proceedings of the Congress of the International Ergonomics Association, Florence, Italy, 26–30 August 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 267–275. [Google Scholar] [CrossRef]
  4. Rehe, G.; Schneider, R.; Thimm, D.D. Transformation towards Human Workforce 4.0 through the Qualification Model LTA-FIT. In Proceedings of the 53rd Hawaii International Conference on System Sciences, Maui, HI, USA, 7–10 January 2020. [Google Scholar] [CrossRef] [Green Version]
  5. Li, C.; Hansen, A.K.; Chrysostomou, D.; Bøgh, S.; Madsen, O. Bringing a Natural Language-enabled Virtual Assistant to Industrial Mobile Robots for Learning, Training and Assistance of Manufacturing Tasks. In Proceedings of the 2022 IEEE/SICE International Symposium on System Integration (SII), Narvik, Norway, 9–12 January 2022; pp. 238–243. [Google Scholar] [CrossRef]
  6. Schou, C.; Andersen, R.S.; Chrysostomou, D.; Bøgh, S.; Madsen, O. Skill-based instruction of collaborative robots in industrial settings. Robot. Comput.-Integr. Manuf. 2018, 53, 72–80. [Google Scholar] [CrossRef]
  7. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  8. Lu, Y.; Zheng, H.; Chand, S.; Xia, W.; Liu, Z.; Xu, X.; Wang, L.; Qin, Z.; Bao, J. Outlook on human-centric manufacturing towards Industry 5.0. J. Manuf. Syst. 2022, 62, 612–627. [Google Scholar] [CrossRef]
  9. Lester, S.; Costley, C. Work-based learning at higher education level: Value, practice and critique. Stud. High. Educ. 2010, 35, 561–575. [Google Scholar] [CrossRef] [Green Version]
  10. Ansari, F.; Erol, S.; Sihn, W. Rethinking human-machine learning in industry 4.0: How does the paradigm shift treat the role of human learning? Procedia Manuf. 2018, 23, 117–122. [Google Scholar] [CrossRef]
  11. Smith, C.; Worsfold, K. Unpacking the learning–work nexus:‘priming’as lever for high-quality learning outcomes in work-integrated learning curricula. Stud. High. Educ. 2015, 40, 22–42. [Google Scholar] [CrossRef]
  12. Lamancusa, J.S.; Jorgensen, J.E.; Zayas-Castro, J.L. The learning factory—A new approach to integrating design and manufacturing into the engineering curriculum. J. Eng. Educ. 1997, 86, 103–112. [Google Scholar] [CrossRef]
  13. Enke, J.; Tisch, M.; Metternich, J. Learning factory requirements analysis–Requirements of learning factory stakeholders on learning factories. Procedia Cirp 2016, 55, 224–229. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, W.; Cai, W.; Min, J.; Fleischer, J.; Ehrmann, C.; Prinz, C.; Kreimeier, D. 5G and AI Technology Application in the AMTC Learning Factory. Procedia Manuf. 2020, 45, 66–71. [Google Scholar] [CrossRef]
  15. Schou, C.; Avhad, A.; Bøgh, S.; Madsen, O. Towards the Swarm Production Paradigm. In Towards Sustainable Customization: Bridging Smart Products and Manufacturing Systems; Springer: Berlin/Heidelberg, Germany, 2021; pp. 105–112. [Google Scholar] [CrossRef]
  16. Mortensen, S.T.; Chrysostomou, D.; Madsen, O. A novel framework for virtual recommissioning in reconfigurable manufacturing systems. In Proceedings of the 2017 22nd IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Limassol, Cyprus, 12–15 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–4. [Google Scholar] [CrossRef]
  17. Wurhofer, D.; Meneweger, T.; Fuchsberger, V.; Tscheligi, M. Reflections on operators’ and maintenance engineers’ experiences of smart factories. In Proceedings of the 2018 ACM Conference on Supporting Groupwork, Sanibel Island, FL, USA, 7–10 January 2018; pp. 284–296. [Google Scholar] [CrossRef]
  18. Kohl, M.; Heimeldinger, C.; Brieke, M.; Fottner, J. Competency model for logistics employees in smart factories. In Proceedings of the International Conference on Applied Human Factors and Ergonomics, Washington, DC, USA, 24–28 July 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 133–145. [Google Scholar] [CrossRef]
  19. Simic, M.; Nedelko, Z. Development of competence model for Industry 4.0: A theoretical approach. Economic and Social Development: Book of Proceedings. 2019, pp. 1288–1298. Available online: https://www.esd-conference.com/upload/book_of_proceedings/Book_of_Proceedings_esdPrague2019_Online.pdf (accessed on 1 October 2022).
  20. Daling, L.M.; Schröder, S.; Haberstroh, M.; Hees, F. Challenges and requirements for employee qualification in the context of human-robot-collaboration. In Proceedings of the 2018 IEEE Workshop on Advanced Robotics and its Social Impacts (ARSO), Genova, Italy, 27–29 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 85–90. [Google Scholar] [CrossRef]
  21. Li, C.; Park, J.; Kim, H.; Chrysostomou, D. How Can I Help You? An Intelligent Virtual Assistant for Industrial Robots; HRI’21 Companion; Association for Computing Machinery: New York, NY, USA, 2021; pp. 220–224. [Google Scholar] [CrossRef]
  22. Nardello, M.; Madsen, O.; Møller, C. The smart production laboratory: A learning factory for industry 4.0 concepts. In Proceedings of the CEUR Workshop Proceedings. CEUR Workshop Proceedings, Copenhagen, Denmark, 28–30 August 2017; Volume 1898. [Google Scholar]
  23. Rawassizadeh, R.; Sen, T.; Kim, S.J.; Meurisch, C.; Keshavarz, H.; Mühlhäuser, M.; Pazzani, M. Manifestation of virtual assistants and robots into daily life: Vision and challenges. CCF Trans. Pervasive Comput. Interact. 2019, 1, 163–174. [Google Scholar] [CrossRef]
  24. Pogorskiy, E.; Beckmann, J.F.; Joksimović, S.; Kovanović, V.; West, R. Utilising a Virtual Learning Assistant as a Measurement and Intervention Tool for Self-Regulation in Learning. In Proceedings of the 2018 IEEE International Conference on Teaching, Assessment, and Learning for Engineering (TALE), Wollongong, NSW, Australia, 4–7 December 2018; pp. 846–849. [Google Scholar] [CrossRef]
  25. Preece, A.; Webberley, W.; Braines, D.; Zaroukian, E.G.; Bakdash, J.Z. Sherlock: Experimental Evaluation of a Conversational Agent for Mobile Information Tasks. IEEE Trans. Hum.-Mach. Syst. 2017, 47, 1017–1028. [Google Scholar] [CrossRef] [Green Version]
  26. da Silva Barbosa, A.; Silva, F.P.; dos Santos Crestani, L.R.; Otto, R.B. Virtual assistant to real time training on industrial environment. In Proceedings of the Transdisciplinary Engineering Methods for Social Innovation of Industry 4.0, Proceedings of the 25th ISPE Inc. International Conference on Transdisciplinary Engineering, Modena, Italy, 3–6 July 2018; IOS Press: Amsterdam, The Netherlands, 2018; Volume 7, p. 33. [Google Scholar] [CrossRef]
  27. Casillo, M.; Colace, F.; Fabbri, L.; Lombardi, M.; Romano, A.; Santaniello, D. Chatbot in Industry 4.0: An Approach for Training New Employees. In Proceedings of the 2020 IEEE International Conference on Teaching, Assessment, and Learning for Engineering (TALE), Takamatsu, Japan, 8–11 December 2020; pp. 371–376. [Google Scholar] [CrossRef]
  28. Zimmer, M.; Al-Yacoub, A.; Ferreira, P.; Lohse, N. Towards Human-Chatbot Interaction: A Virtual Assistant for the Ramp-Up Process. 2020. Available online: https://repository.lboro.ac.uk/articles/conference_contribution/Towards_human-chatbot_interaction_a_virtual_assistant_for_the_ramp-up_process/12091137/1/files/22231974.pdf (accessed on 1 October 2022).
  29. Li, C.; Yang, H.J. Bot-X: An AI-based virtual assistant for intelligent manufacturing. Multiagent Grid Syst. 2021, 17, 1–14. [Google Scholar] [CrossRef]
  30. Maksymova, S.; Matarneh, R.; Lyashenko, V.; Belova, N. Voice Control for an Industrial Robot as a Combination of Various Robotic Assembly Process Models. J. Comput. Commun. 2017, 5, 1–15. [Google Scholar] [CrossRef] [Green Version]
  31. Bingol, M.C.; Aydogmus, O. Performing predefined tasks using the human–robot interaction on speech recognition for an industrial robot. Eng. Appl. Artif. Intell. 2020, 95, 103903. [Google Scholar] [CrossRef]
  32. Russo, A.; D’Onofrio, G.; Gangemi, A.; Giuliani, F.; Mongiovi, M.; Ricciardi, F.; Greco, F.; Cavallo, F.; Dario, P.; Sancarlo, D.; et al. Dialogue systems and conversational agents for patients with dementia: The human–robot interaction. Rejuvenation Res. 2019, 22, 109–120. [Google Scholar] [CrossRef]
  33. Bernard, D. Cognitive interaction: Towards “cognitivity” requirements for the design of virtual assistants. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 210–215. [Google Scholar] [CrossRef]
  34. Skantze, G. Turn-taking in conversational systems and human-robot interaction: A review. Comput. Speech Lang. 2021, 67, 101178. [Google Scholar] [CrossRef]
  35. Weiss, A.; Huber, A.; Minichberger, J.; Ikeda, M. First application of robot teaching in an existing industry 4.0 environment: Does it really work? Societies 2016, 6, 20. [Google Scholar] [CrossRef]
  36. Meissner, A.; Trübswetter, A.; Conti-Kufner, A.S.; Schmidtler, J. Friend or foe? understanding assembly workers’ acceptance of human-robot collaboration. ACM Trans. Hum-Robot. Interact. THRI 2020, 10, 1–30. [Google Scholar] [CrossRef]
  37. Chacón, A.; Ponsa, P.; Angulo, C. Usability Study through a Human-Robot Collaborative Workspace Experience. Designs 2021, 5, 35. [Google Scholar] [CrossRef]
  38. Jost, C.; Le Pévédic, B.; Belpaeme, T.; Bethel, C.; Chrysostomou, D.; Crook, N.; Grandgeorge, M.; Mirnig, N. (Eds.) Human-Robot Interaction: Evaluation Methods and Their Standardization; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12, p. 385. [Google Scholar] [CrossRef]
  39. Rueben, M.; Elprama, S.A.; Chrysostomou, D.; Jacobs, A. Introduction to (Re) Using Questionnaires in Human-Robot Interaction Research. In Human-Robot Interaction: Evaluation Methods and Their Standardization; Springer: Berlin/Heidelberg, Germany, 2020; pp. 125–144. [Google Scholar] [CrossRef]
  40. Bangor, A.; Kortum, P.T.; Miller, J.T. An empirical evaluation of the system usability scale. Int. J. Hum. Comput. Interact. 2008, 24, 574–594. [Google Scholar] [CrossRef]
  41. Hart, S.G.; Staveland, L.E. Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research. Adv. Psychol. 1988, 52, 139–183. [Google Scholar] [CrossRef]
  42. Deemter, K.v.; Theune, M.; Krahmer, E. Real versus template-based natural language generation: A false opposition? Comput. Linguist. 2005, 31, 15–24. [Google Scholar] [CrossRef]
  43. Yu, Z.; Xu, Z.; Black, A.W.; Rudnicky, A. Strategy and policy learning for non-task-oriented conversational systems. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Los Angeles, CA, USA, 13–15 September 2016; pp. 404–412. [Google Scholar] [CrossRef]
  44. Fensel, D.; Şimşek, U.; Angele, K.; Huaman, E.; Kärle, E.; Panasiuk, O.; Toma, I.; Umbrich, J.; Wahler, A. Introduction: What is a knowledge graph? In Knowledge Graphs; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 1–10. [Google Scholar] [CrossRef]
  45. Brooke, J. Sus: A “quick and dirty’usability. Usability Eval. Ind. 1996, 189, 4–7. [Google Scholar]
  46. National Aeronautics and Space Administration. NASA TLX: Task Load Index. 2020. Available online: https://humansystems.arc.nasa.gov/groups/TLX/ (accessed on 1 October 2022).
  47. Babakus, E.; Mangold, W.G. Adapting the SERVQUAL scale to hospital services: An empirical investigation. Health Serv. Res. 1992, 26, 767–786. [Google Scholar]
  48. Chyung, S.Y.; Roberts, K.; Swanson, I.; Hankinson, A. Evidence-Based Survey Design: The Use of a Midpoint on the Likert Scale. Perform. Improv. 2017, 56, 15–27. [Google Scholar] [CrossRef]
Figure 1. The autonomous industrial mobile manipulator Little Helper, in which our proposed virtual assistant is integrated.
Figure 1. The autonomous industrial mobile manipulator Little Helper, in which our proposed virtual assistant is integrated.
Applsci 13 00205 g001
Figure 2. The architecture of the proposed natural language-enabled VA.
Figure 2. The architecture of the proposed natural language-enabled VA.
Applsci 13 00205 g002
Figure 3. The fine-tuned pretrained BERT model for intent prediction and slot extraction.
Figure 3. The fine-tuned pretrained BERT model for intent prediction and slot extraction.
Applsci 13 00205 g003
Figure 4. The web interface of the Max client running on a common web browser.
Figure 4. The web interface of the Max client running on a common web browser.
Applsci 13 00205 g004
Figure 5. The parts of the mock-up mobile phone used for the learning and training scenarios.
Figure 5. The parts of the mock-up mobile phone used for the learning and training scenarios.
Applsci 13 00205 g005
Figure 6. The graphic representation of the knowledge graph for the assembly of the mock-up mobile phone.
Figure 6. The graphic representation of the knowledge graph for the assembly of the mock-up mobile phone.
Applsci 13 00205 g006
Figure 7. An analysis of our experimental setup. (a) Schematic representation of the use case setup. Each user was positioned 30 cm away from the robot platform. The camera and microphone were positioned 1.5 m away and the screen where Max was showing messages for the interaction was placed 80 cm away from the user. (b) Phone components on the robot table.
Figure 7. An analysis of our experimental setup. (a) Schematic representation of the use case setup. Each user was positioned 30 cm away from the robot platform. The camera and microphone were positioned 1.5 m away and the screen where Max was showing messages for the interaction was placed 80 cm away from the user. (b) Phone components on the robot table.
Applsci 13 00205 g007
Figure 8. An example of a task and a dialogue for the training scenario. (a) A sample task specification for the assembly of the mock-up phone. (b) Example dialogue for the task of phone assembly.
Figure 8. An example of a task and a dialogue for the training scenario. (a) A sample task specification for the assembly of the mock-up phone. (b) Example dialogue for the task of phone assembly.
Applsci 13 00205 g008
Figure 9. Participants’ demographics. (a) Participants’ age range. (b) Participants’ gender. (c) Participants’ study/work background.
Figure 9. Participants’ demographics. (a) Participants’ age range. (b) Participants’ gender. (c) Participants’ study/work background.
Applsci 13 00205 g009
Figure 10. The obtained results based on the NASA-TLX questionnaire.
Figure 10. The obtained results based on the NASA-TLX questionnaire.
Applsci 13 00205 g010
Figure 11. Scores obtained for the overall frustration subscale of NASA-TLX.
Figure 11. Scores obtained for the overall frustration subscale of NASA-TLX.
Applsci 13 00205 g011
Figure 12. Scores obtained for the subjective metrics: comfortability scale (CS), perceived competence (PC), performance expectancy (PE), and effort expectancy (EE).
Figure 12. Scores obtained for the subjective metrics: comfortability scale (CS), perceived competence (PC), performance expectancy (PE), and effort expectancy (EE).
Applsci 13 00205 g012
Table 1. Identified tasks for the three LTA-FIT scenarios.
Table 1. Identified tasks for the three LTA-FIT scenarios.
CategoryTask DescriptionIntent Recognition
Learning1: Explain terminologyKeywords
Learning2: Reason operation relationshipsKeywords
Training3: Demonstrate auto phone assemblyKeywords
Training4: Guide phone assemblyKeywords
Assistance5: Assist package deliveryBERT Model
Assistance6: Check system statusBERT Model
Table 2. Quantitative results for the performance of the proposed VA in the tasks of the assistance scenario.
Table 2. Quantitative results for the performance of the proposed VA in the tasks of the assistance scenario.
Task IDIERSERTSRACT (in secs)
50.10.430.5031.08
60.20.060.7615.28
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, C.; Chrysostomou, D.; Pinto, D.; Hansen, A.K.; Bøgh, S.; Madsen, O. Hey Max, Can You Help Me? An Intuitive Virtual Assistant for Industrial Robots. Appl. Sci. 2023, 13, 205. https://doi.org/10.3390/app13010205

AMA Style

Li C, Chrysostomou D, Pinto D, Hansen AK, Bøgh S, Madsen O. Hey Max, Can You Help Me? An Intuitive Virtual Assistant for Industrial Robots. Applied Sciences. 2023; 13(1):205. https://doi.org/10.3390/app13010205

Chicago/Turabian Style

Li, Chen, Dimitrios Chrysostomou, Daniela Pinto, Andreas Kornmaaler Hansen, Simon Bøgh, and Ole Madsen. 2023. "Hey Max, Can You Help Me? An Intuitive Virtual Assistant for Industrial Robots" Applied Sciences 13, no. 1: 205. https://doi.org/10.3390/app13010205

APA Style

Li, C., Chrysostomou, D., Pinto, D., Hansen, A. K., Bøgh, S., & Madsen, O. (2023). Hey Max, Can You Help Me? An Intuitive Virtual Assistant for Industrial Robots. Applied Sciences, 13(1), 205. https://doi.org/10.3390/app13010205

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop