skip to main content
research-article
Open Access

Levels of Automation for a Mobile Robot Teleoperated by a Caregiver

Authors Info & Claims
Published:08 February 2022Publication History

Skip Abstract Section

Abstract

Caregivers in eldercare can benefit from telepresence robots that allow them to perform a variety of tasks remotely. In order for such robots to be operated effectively and efficiently by non-technical users, it is important to examine if and how the robotic system’s level of automation (LOA) impacts their performance. The objective of this work was to develop suitable LOA modes for a mobile robotic telepresence (MRP) system for eldercare and assess their influence on users’ performance, workload, awareness of the environment, and usability at two different levels of task complexity. For this purpose, two LOA modes were implemented on the MRP platform: assisted teleoperation (low LOA mode) and autonomous navigation (high LOA mode). The system was evaluated in a user study with 20 participants, who, in the role of the caregiver, navigated the robot through a home-like environment to perform control and perception tasks. Results revealed that performance improved in the high LOA when task complexity was low. However, when task complexity increased, lower LOA improved performance. This opposite trend was also observed in the results for workload and situation awareness. We discuss the results in terms of the LOAs’ impact on users’ attitude towards automation and implications on usability.

Skip 1INTRODUCTION Section

1 INTRODUCTION

Many older adults prefer aging in place within their own homes, which has been associated with increased well-being [41]. Moreover, as the ratio between the aging population and caregivers continually increases [40] and families live farther apart, this circumstance often becomes a necessity. Although many older adults manage to remain largely independent, they may require occasional help with activities of daily living from a caregiver [1]. One way to facilitate this is through the use of mobile robotic telepresence (MRP) systems [26]. MRP refers to the activity of remote controlling a robot that is able to move through its environment, thereby enabling its user (referred to as remote user or operator) to interact with other people (local users) within that same physical space [26]. As robotic hardware keeps evolving and sophisticated sensors and actuators become affordable, users will be enabled to perform more complex tasks—for example, as in the present use case, to support older adults in their daily lives [31] or when social distancing is required [17]. In the future, caregivers who cannot be physically present in the older adult’s residence will be able to use telepresence robots to visit. These systems afford a high level of independence and wide range of action, as seen in several projects focusing on the development of telepresence robots to support older adults in a range of daily activities [53]. Examples of previous work include developing a research platform (GiraffPlus [10]), design recommendations and functionalities (ExCITE EU project [34]), and developing social navigation capabilities (the TERESA project [44]). The present study focuses on two specific constructs, levels of automation (LOA) and task complexity, which were evaluated for a teleoperated mobile robot in terms of performance, workload, situation awareness (SA), and usability.

LOA defines the degree to which automation is employed in a given task [2, 14, 30]. Several short-term evaluations of MRP for telecare in home environments revealed maneuvering challenges and high workload when manually driving the robot. Hence, this study focuses on the design of LOA for navigation [9, 25]. If navigation is automated efficiently, it can free the operator to direct their attention to other tasks and subtasks [37]. Furthermore, as tasks become increasingly complex, it may be difficult for an operator to conduct the task sufficiently well manually [11]. Therefore, navigation is a critical function for which LOA can be introduced, especially in the context of different levels of task complexity.

SA is described as the degree to which an operator, at any point in time while controlling or monitoring a system (e.g., a robot), is aware of its current internal state and the state of the external environment in which it is located and acting [13, 15]. In the literature, SA is frequently defined as a qualitatively incremental variable spanning three levels: The ability to perceive the incoming and available data that is relevant to the current task or context, understanding the meaning and implications of that data as it pertains to the current state, and finally the ability to make predictions about how the situation is likely to develop in the (near) future if kept on its current trajectory [13]. Mental workload is a measure of the degree to which a person’s executive cognitive functions are being occupied at a given time [7], for example, as a result of focusing on one or several tasks which demand sustained allocation of these processes.

Task complexity has been identified in previous research as a critical factor impacting performance [11] and influencing the LOA design in human-robot interaction [2, 20]. It depends predominantly on properties of the task (objective complexity) and the perception of the human operator (subjective complexity) [39]. Further dimensions of complexity include component complexity —the number of distinct actions that the human operator must execute or number of informational cues that should be processed (e.g., the number and types of subtasks to be managed individually) [32]; coordinative complexity —the nature of relationships between task inputs and task products, the strength of these relationships, simultaneous action requirements, as well as the sequencing of inputs (e.g., timing, frequency, intensity and location requirements, level of difficulty) [8], as well as dynamic complexity —changes in the states of the environment, e.g., cause-effect chains, means-ends connections to which the human operator should adapt, criticality of changes, and the degree of human intervention required for these changes [5, 51].

Only few studies have considered task complexity in the design and evaluation of LOA for MRP systems specifically in the case of non-technical operators (e.g., caregivers) [18, 26, 27, 45]. Specific LOA designs suitable and feasible for such MRP scenarios are still lacking [47]. Previous work [24] evaluated semi-autonomy features in MRPs designed to relieve the pilot user of some of the mental and physical demand associated with maneuvering the robot. Other studies revealed that adjusting the robot’s autonomy in teleoperation tasks can help facilitate their use across a broader range of applications [22, 43]. In the context of a user interacting with a robot, such as in the case of an MRP task, it has been previously argued that dynamic adaptation of control and responsibilities is necessary to accommodate different users and successfully manage a variety of situations and tasks [38]. Autonomy is typically considered as a continuous or discrete spectrum, with direct human control and full autonomy at either end, and any number of intermediate levels in between [2]. However, this view is arguably more appropriate when considered at the task level, since a system may be performing multiple tasks simultaneously at different levels of autonomy. Moreover, most tasks can be decomposed further into subtasks and basic actions, which may be shared between a robot and human operator.

This work aims to develop and evaluate two LOA designs —assisted teleoperation (low LOA) mode and autonomous (high LOA) mode— which are functional, suitable, and adaptable for users in different task complexity scenarios. In general, the results of previous LOA experiments did not reveal clear-cut response trends regarding the influence of LOA on operator performance, workload, and SA [22]. The outcomes of these previous studies further suggest that some of our intuitions about automation and human performance have been imprecise or incomplete, partly due to LOA evaluations conducted in isolation from other influencing factors and limited number of studies conducted in this area [33]. In response to the recurring need for more accurate assessments of human performance variables in different contexts, the current research investigates the interaction of the LOA modes with different task complexity levels that may arise in the home environment during teleoperation. Navigating the robot in a more complex task with longer interactions, more waypoints and obstacles can lead to fatigue or loss of SA, on which the MRP system’s LOA design may have an impact. User attitude towards the LOA designs at different levels of complexity was investigated in a user study.

Participants took on the role of a caregiver controlling a telepresence robot using different LOA modes—to navigate through an assisted living apartment, check in on a person and perform a variety of observational tasks. The study comprised four conditions, combining two independent variables in a factorial format: LOA and task complexity. The objective was to evaluate the influence of the LOA modes on users’ performance, workload, SA, and usability in a teleoperated task with a focus on navigation.

The next section describes the different aspects of the MRP task in the present study, i.e., the functions to be carried out by the user, their respective allocation in the two LOAs, as well as the process of facilitating the coordination of the user-robot interaction. Section 3 provides an in-depth description of the study design and methodology, followed by results, discussion and outlook in Sections 4, 5, and 6, respectively.

Skip 2FUNCTION ALLOCATION AND LOA DEVELOPMENT IN MRP Section

2 FUNCTION ALLOCATION AND LOA DEVELOPMENT IN MRP

The function allocation and LOA development for the MRP were aimed to ensure that tasks (and subtasks) in the system are appropriately allocated to the human, robot, or both, at specific degrees of autonomy while allowing for adaptation as required under varying conditions. Fully manual and fully autonomous modes were not included in the design. Achieving fully autonomous operation is not practical at the moment, as tasks may change and systems are still not developed enough to handle dynamic changes in the home environment. Instead, intermediate levels where functions and tasks are shared between operator and robot are feasible.

The function allocations are based on estimated capacities of the user and robot in a given situation to ensure coordination and collaboration between the human and automation [12]. The functions in the MRP system were identified according to the four-stage “O-O-D-A loop” information model [4, 14, 36]. It involves functions related to (1) information acquisition (Observe), (2) information analysis (Orient), (3) decision selection (Decide), and (4) action implementation (Act), for the operation of the robot in the local environment and the remote operator user interface, as shown in Table 1.

Table 1.
  • System aspect refers to the different parts of the system to which the four stages apply and in which either the human, the robot, or both have a role to play.

Table 1. Functions in the MRP Task

  • System aspect refers to the different parts of the system to which the four stages apply and in which either the human, the robot, or both have a role to play.

These functions were utilized in defining two LOA modes, to qualify the MRP system for navigation focused on the specific task of obstacle detection and avoidance (Table 2). In this process, considerations were made regarding possible failures or emergencies. The role of handling such situations is referred to as a “contingency role” which involves taking certain actions in the light of such events. Details of the roles in each of the LOA modes are described below. Some roles such as monitoring, analysis, and selection of multimedia settings and communication are entirely under the operator’s control in both modes.

Table 2.
  • Cont.=contingency roles for fallback tasks, Coord.=coordinating responsibility demands such as in path planning and obstacle avoidance, Switch=LOA switching option.

Table 2. Summary of LOA Modes Implemented for the MRP

  • Cont.=contingency roles for fallback tasks, Coord.=coordinating responsibility demands such as in path planning and obstacle avoidance, Switch=LOA switching option.

2.1 Assisted Teleoperation Mode (Low LOA)

In the low LOA mode, the human observes the environment through a video stream. A map located on the side of the interface displays the robot’s position and is continually updated as the robot moves. The human observes the environment to ensure that the robot does not collide with any object or person as it is directed towards a goal. Meanwhile, the robot scans the environment with its laser scanner. If obstacles are detected within a certain range, it decelerates to avoid collision or to attenuate the impact. This is done simultaneously and autonomously, without the need for the human to activate, manipulate, or regulate the monitoring process during the task. This mode can be described as a guarded teleoperation mode.

The operator is further in charge of strategy generation and decision-making, identifying possible trajectories to reach the chosen destination and managing various multimedia settings related to the social interaction. These actions are jointly executed by the human and the robot. The human takes the action of activating controls related to starting the robot, positioning, and navigating the robot through the user interface, but the robot decelerates based on obstacle proximity as reported by its laser readings.

Additionally, it is primarily the human operator’s responsibility to monitor the system’s state through the user interface, to ensure that essential modules are functioning adequately. The robot only assists partially in monitoring the battery level, setting off a warning sound when it drops below a preset threshold.

2.2 Autonomous Navigation Mode (High LOA)

The robot is fully responsible for observing the environment to identify potential obstacles in its path, while the human operator only monitors the state of the robot through the user interface. The autonomous system identifies areas that are safe for navigation and within which the operator can select an appropriate goal location. The autonomous navigation, in turn, plots a viable path, if any can be found, towards the desired destination and subsequently follows it. As the robot moves through the environment, it continuously updates its local map to track dynamic obstacles, which it avoids when and where possible. This differs from the assisted teleoperation mode, where the map is also continuously updated but the robot only decelerates to avoid or minimize collision with obstacles. As a special feature in this LOA mode, the operator can decide to switch between LOA modes or cancel an active goal, as deemed appropriate or necessary throughout the interaction. The rationale for allowing participants to switch to the low LOA was that, upon arriving at a location, teleoperation is better suited for fine-tuning, particularly for adjusting the orientation to face areas of interest. While it is possible to do so with autonomous navigation mode, it may not be quite as precise as desired.

Skip 3METHODOLOGY Section

3 METHODOLOGY

3.1 Overview

User studies were conducted at the Center for Applied Autonomous Sensor Systems (AASS) at Örebro University in Sweden. A telepresence robot was deployed in a home-like testbed environment (called “PEIS-Home 2”) with a living room, kitchen, and bedroom (see Figure 1). The walls separating the rooms are only approximately one meter high, which allows experimenters to keep a view over the entire space. In a separate room outside of view and hearing range of the robot, participants were seated in front of a standard computer screen with a mouse and keyboard to control the robot through a user interface. A social humanoid robot (Pepper robot) [35] was placed in the living room to represent the older adult living in the home (henceforth referred to as the resident robot). Given its basic speech recognition function, we used it to evaluate how well the telepresence robot facilitates conversations, using a simple script which was provided for the participants. Pepper’s height is comparable to that of a sitting adult and thus representative of a typical interaction scenario. The description of the MRP system, the LOA modes implemented on the user interface, the tasks and the experimental design are presented in the following subsections.

Fig. 1.

Fig. 1. A cross-section of the home-like environment used for the study.

3.2 The System

The MRP system consists of a mobile robot platform with remote and local user interfaces developed to work through a server-client communication architecture utilizing WebRTC and rosbridge websocket. The user interfaces run within a standard web browser and are independent of the device’s operating system or any specific software. This is particularly relevant in the case of the operator client, as it makes the robot more widely accessible and less prone to issues such as missing updates. More details on the robot platform and user interfaces are provided as follows:

The robot platform. The platform is an extensively retrofitted Giraff telepresence robot [34] with a differential drive and screen with mechanical tilt function. Its height is approximately 1.65 m and the footprint of the base 0.55 by 0.62 m (see Figure 2). The most significant modifications to the original hardware configuration include the removing of the plastic cladding, replacing the battery, and adding a Hokuyo URG-04LX-UG01 2D laser scanner, as well as a wide-angle Structure Core RGBD camera. The latter was only used as an RGB webcam. The proprietary software was replaced with a custom-designed interface and back-end implementation based on the robot operating system (ROS). The autonomous navigation capability was realized via the ROS navigation stack, which provides the controller, local, and global planners, as well as 2D obstacle detection using the laser range data.

Fig. 2.

Fig. 2. The modified Giraff telepresence robot.

User Interfaces. The remote operator user interface (see Figure 3) running on the client device is divided into three sections: a left, central, and right panel. The left panel contains a small version of the operator’s own video feed, as is common in videoconferencing, multimedia controls, and the four conventional navigation control buttons. The central panel holds the video of the scene overlaid with a “pointing plate” used as a trackball-like driving mechanism. More details on the use is provided in the next paragraph. The right panel includes a robot-generated reactive map of the apartment and “goal buttons” to send the robot to specific locations within the home.

Fig. 3.

Fig. 3. The operator’s user interface: The camera image from the robot with the graphical indicator for the manual navigation is at the center, the operator’s own webcam, multimedia control buttons, and the conventional navigation buttons are located on the left of the screen; the interactive map and goal buttons are to the right. The same user interface was used for the high and low LOA mode but only specific features corresponding to the LOA in use were allowed for use in each experimental condition.

The two LOA modes are realized and accessed within this client user interface. In the low LOA mode, the first and main option for navigation control is operated via the pointing plate on the central panel of the interface. This pointing plate indicates the approximate position the robot would move towards, with a curved line originating from the center bottom to plot a rough trajectory towards that site (see Figure 3). If the mouse is moved across the central part of the interface containing the video scene, the plate and tail follow the mouse on a plane projection representing the floor. By pressing and holding down the mouse button, users accelerate the robot; releasing the button causes it to slow down and eventually come to a stop. The higher up (and thus farther ahead of the robot) the cursor is on the image while the mouse button is pressed, the faster the robot moves forward. Likewise, the angular velocity increases as the cursor is dragged farther to the left and right edges. This type of interface was chosen because it allows for precision with respect to linear and angular velocity (as opposed to keyboard keys) while relying on the mouse as a commonplace input device. The other, auxiliary navigation control option for the low LOA mode was the conventional navigation control buttons on the left panel of the interface, i.e., arrows for left, right, forward, and backward in the X-Y plane. This option has been introduced largely as a fallback system for simple actions such as backing up or rotating in place.

In the high LOA mode the user selects the desired goal on the reactive top-down map on the right side of the interface. The map displays the environment used for the study, along with annotations of relevant locations corresponding to specific rooms. By clicking on any accessible location on the map, the robot is given the task to plan a suitable path and navigate towards the destination. Dragging the mouse in any direction before releasing the button determines the orientation the robot is going to rotate towards upon reaching the goal. The secondary navigation control option in the high LOA mode is given through the goal buttons, labeled with the name of the different locations in the environment (e.g., kitchen, living room, and bedroom). The user can click on any of these buttons to send the robot to the desired location. Since this option only allows to move to a small set of discrete positions, it is not suited as a standalone function and rather intended to serve as a supporting feature. Nevertheless, both options rely on the same underlying mechanics (i.e., map coordinates and the ROS navigation stack).

Switching between modes is an implicit action, as using one LOA mode deactivates the other. For example, clicking on the video image for manual teleoperation (low LOA mode) cancels any active goal of the autonomous navigation (high LOA mode). Similarly, manual teleoperation (low LOA mode) is effectively deactivated as soon as no input is given, while, as mentioned, clicking on the map activates high LOA mode.

In addition to the remote user interface, the local user interface is displayed on the robot’s screen and includes the video stream for the local user to communicate with the remote operator.

3.3 Tasks

The tasks involved navigating the robot through the environment to specific locations while looking out for certain details in the surroundings. To assess participants’ SA, two items were placed on the floor in the doorway and the kitchen. The exact location of the items was randomized across the different trials. The robot always started at its charging station. An observation sheet was provided for the participants to take notes of their observations. Below, we describe the action sequences in both tasks:

Low complexity task (LTask). This task involves a small set of actions and input sequence requirements from the user, contributing to low component and coordinative complexity (earlier defined), respectively. The sequence of actions is as follows: 1. Locate the resident robot in the environment and drive up to it so that its face is well visible in the interface, 2. Communicate with the robot (a short script for the communication was provided, consisting mostly of greetings which the resident robot had the ability to respond to), 3. Navigate the robot back to its charging area.

High complexity task (HTask). This task involves a more distinct set of actions, as the user has to carry out longer input sequences. This increases the component and coordinative complexity compared to the low complexity task. The sequence of actions is as follows: 1. Locate the charger of the resident robot, 2. Check if the door to the home is closed, 3. Check for any obstacles on the floor that could cause a fall (two specific objects were placed on the way to the home’s entrance door and in the passage to the kitchen), 4. Check if the tap in the kitchen sink is running, 5. Navigate the robot to the charging area, and 6. Check if the robot’s charger is plugged in.

3.4 Hypotheses

Previous research revealed that performance is more likely to decline at higher robot autonomy with increasing task complexity [11]. Intuitively, it seems plausible to assume that longer and more complex tasks involve a higher potential for the occurrence of dynamic and unforeseen events which may require a degree of flexibility that (current) automation cannot provide. We therefore propose that:

H1: As task complexity increases, user performance will be higher in lower LOA (relative to higher LOA).

High task complexity often involves a higher probability of performance declines, uncertainty and failure [33]. Previous research revealed that users appear to have more confidence in their own ability to handle decisions at such higher levels of complexity compared to automation-generated decisions [15]. Increasing automation at high task complexity where more uncertainties can arise often shifts the workload towards monitoring, in an effort to ensure performance, as previously expressed by [3, 50]. In high-complexity situations, robot effectiveness is likely to decline if the robot is operating at higher autonomy, because there are higher chances for novel and dynamic conditions to arise for which the autonomous function was not prepared [16]. This calls for further investigation, particularly in the MRP scenario. We therefore define the second hypothesis to investigate this possibility as follows:

H2: As task complexity increases, workload will increase in higher LOA (relative to lower LOA).

Assessments of SA at different LOA [15] revealed that, as automation increases, users’ comprehension of the different variables related to the current task declines. This trend appears self-evident, as higher LOA reduces the human involvement, thus allowing operators to direct their attention towards other tasks. This relates to an out-of-the-loop performance problem which has been widely documented as a potential negative consequence of higher LOA [3, 28, 48]. These works mostly examined tasks with professional operators in operational domains such as aviation. Some studies reveal minimal and in some cases no impact on SA, as reported in the meta-studies conducted in [33]. In the current MRP scenario, where caregivers, who are not robotic experts are expected to control the robot remotely, there are considerable differences with respect to system demands, user expectations, and performance assessment. We therefore suggest that:

H3: As task complexity increases, SA will be higher in lower LOA (relative to higher LOA).

Many of the field studies conducted with MRP systems outlined the need for inclusion of autonomous features to improve the usability of the system and satisfaction [34]. Efforts to introduce autonomy to MRP platforms to reduce mental demand on users have been reported to be promising [23, 24]. The availability of more autonomous functions may increase users’ willingness to use the system when facing increasingly complex tasks or higher workload. Therefore, we propose that:

H4: As task complexity increases, the availability of LOA options will improve usability.

3.5 Experimental Design

The study was conducted in a within-participants format, with every participant performing a total of four randomized trials corresponding to four different conditions. The conditions resulted from the 2x2 factorial combination of the independent variables, task complexity (low, high) and LOA mode (low, high). In the high LOA conditions, participants were allowed to switch to low LOA at any point in the task, as they preferred. This is due to the fact that, upon arriving at a location, teleoperation is often better suited for small adjustments to the position and orientation of the robot. These adjustments are frequently made when looking around and inspecting areas of interest. Therefore, we refer to this condition in the experiment as adjustable-high, to better distinguish modes and experimental conditions.

The dependent variables were user performance, perceived workload, SA, and usability, as detailed in Section 3.8.

3.6 Participants

Twenty participants (7 female, 13 male) were recruited at Örebro University for the role of the caregiver (M \(=\) 29, SD \(=\) 6). A majority (14) had a technical background (i.e., computer science, engineering), 6 of whom had worked with robots before. All reported that they use a computer daily, while most (16) use it at work consistently. Nine stated that they do not play any video games at all, while most of the others play games that may have a positive effect on their understanding of and performance with controlling the robot’s interface (e.g., first person shooters, strategy games). All but two of the participants were affiliated with the university. All experiments were approved by the ethics review board at Örebro University.

3.7 Procedure

At the start of the experiment, after reading and signing the consent form, participants were asked to provide some background information regarding their age, gender, level of education, field of study/occupation as well as computer use and video gaming experience. Following this, they were briefed on the scenario, tasks and procedure. Before starting with the four main trials, users were introduced to the user interface and had the chance to practice with both operation modes until they felt sufficiently familiarized with the controls. This was defined as training to a basic use criterion—to navigate the robot to a specified location and back. Each trial was followed by a questionnaire enquiring about the experience with the condition. In between trials, while participants were occupied with the questionnaires, the experimenters would make subtle changes to the environment to reduce the learning effect. These changes concerned the locations or states of objects relevant to the tasks. After completion of all four trials, participants answered a final questionnaire in which they rated their overall experience with the robot and tasks. It further afforded the opportunity to provide free input, feedback, or remarks.

3.8 Measures

3.8.1 Objective Measures.

For each participant and trial, performance was measured in terms of task completion time, the number of subtasks completed, and the number of obstacles missed (which is a subset of the subtask completed). The frequency and reason for switching of the LOA mode (switching from high to low LOA mode) during task execution was also taken into account. Additionally, the number of collisions while operating was recorded. Collisions were defined as any contact with walls, furniture or the resident robot, while assuming that all collisions are unintentional. Though not part of the objective performance assessment, it should be mentioned that the robot did occasionally collide in the high LOA mode, since its laser scanner only detects obstacles on a 2D plane. Few objects protruding at different heights were not registered. On other rare occasions the robot would rotate in place as a recovery behavior before continuing on to the destination.

3.8.2 Subjective Measures.

The post-trial questionnaires included a total of 19 questions from three questionnaires as detailed below. The questions were related to perceived workload (rated on a 5-point Likert scale), SA (7-point Likert scale—according to the standard of the Situation Awareness Rating Technique (SART) [46], explained below) and usability (5-point Likert scale, see Table 3). Perceived workload was assessed using the NASA-Task Load Index (NASA-TLX) questionnaire [19], with overall perceived workload rating computed from the different workload dimensions. It has previously been employed in the evaluation of MRP systems [25]. SA was assessed using the 3D-SART version of the SA Rating Technique [46], which measures complexity of interaction, focus of attention and information quantity. Subjective assessment concerning the system’s usability was collected by means of the System Usability Scale (SUS) questionnaire [6]. The final questionnaire included participants’ assessments regarding the ease of use, as well as possible recommendations for how to develop the system further. Ease of use was evaluated with the Single Ease Questionnaire [42].

Table 3.

Table 3. Self-report Items from the Post-trial Questionnaire

3.9 Data Analysis

A generalized linear mixed model (GLMM) was applied to analyze the data with the LOA mode and task complexity as fixed modes, whereas the random effect was selected as the variances from the participants. The tests were designed as two-tailed with a significance level of 0.05.

Skip 4RESULTS Section

4 RESULTS

Objective performance results are presented as a plot in Figure 4, while more details on the interaction effects of LOA and task complexity on the subjective variables are provided in the following subsections:

Fig. 4.

Fig. 4. Performance of the participants in the objective measures. LTask-LLOA \( = \) low LOA in low task complexity, LTask-HLOA \( = \) adjustable-high LOA in low task complexity, HTask-LLOA \( = \) low LOA in high task complexity, HTask-HLOA \( = \) adjustable-high LOA in high task complexity, Completed \( = \) Participants who completed all subtasks, Missed \( = \) Participants who missed obstacles, Switched \( = \) Participants who switched to low LOA while in adjustable-high LOA condition, Collided \( = \) Participants who collided with obstacles.

4.1 Objective Performance

Subtasks Completed. In the low complexity task, all except one participant completed the subtasks using the low LOA mode and 17 out of 20 participants completed the subtasks using the high LOA mode. In the high complexity task 10 of the 20 participants completed the subtasks using low LOA mode while 11 participants completed the subtasks using the high LOA mode. In the high LOA mode, 5 participants used the map option actively for navigation while about 4 participants stayed with the use of the conventional arrow key controls in the low LOA mode (see Figure 4). The “fall risk” object which was placed on the floor in the robot’s path was detected by 9 out of 20 participants in the high LOA mode and by 8 out of 20 participants in the low LOA mode in the high complexity task. One participant in the high LOA mode missed the resident robot in the low complexity task (Figure 4).

Collisions. Three out of twenty participants had collisions with different objects in the environment (mainly with walls, a table, and chairs) while navigating using the low LOA mode in the low complexity task and 4 participants had a collision in the high complexity task (Figure 4).

Switching of LOA. Twelve out of the twenty participants switched in the adjustable-high LOA condition to low LOA, on average of 2 times in a trial and for various reasons, as detailed in Table 4. Three out of these twelve participants switched in the low complexity task, while 9 of the participants switched in the high complexity task. Six out of the twelve who switched to low LOA returned to the higher LOA after resolving the situation for which they made the initial switch (Figure 4). The remaining 6 of the 12 who switched to the low LOA kept switching between both modes until they completed the task. The reasons for this are also contained in the conditions in (Table 4). It should be mentioned that, if participants ended up executing more than half of the tasks with teleoperation in the adjustable-high LOA condition, it was counted as successful completion in low LOA mode. However, if the low LOA was only used for looking around and orienting the robot, it was considered as completed in the high LOA mode.

Table 4.
Condition for SwitchingExamples
Automation transparencyWhen they were not sure what the robot was doing while it
was navigating autonomously
Unexpected eventsWhen the robot got stuck between obstacles and could not get
out on its own
DissatisfactionWhen the robot did not position itself or accelerate as intended,
with the automationparticipants switched to the low LOA mode to fine-tune the
actions of the robot
Error handlingIf participants gave the wrong instruction to the robot and
wanted to correct the error
Safety considerationsSome of the participants stated that they were concerned about
the robot colliding with obstacles as it navigated on its own
Automation malfunctionsIf the robot did not perform as expected in navigating to the
desired location.

Table 4. Conditions Observed for Switching in the Adjustable-high Condition

4.2 Perceived Workload

Mental demand. LOA mode and task complexity had a significant interaction effect on participants’ mental demand (F(1,73) \(=\) 1.589, p \(=\) 0.009). At low task complexity, the low LOA mode placed a higher mental demand (M \(=\) 2.00, SD \(=\) 1.05) on the participants compared to the high LOA mode (M \(=\) 1.53, SD \(=\) 1.02). In high task complexity, however, both low LOA mode (M \(=\) 2.25, SD \(=\) 0.97) and high LOA mode (M \(=\) 2.21, SD \(=\) 1.18) had an equally high mental demand. The LOA mode alone, however, did not have a significant effect on the mental demand (F(1,73) \(=\) 1.910, p \(=\) 0.171), though at high task complexity (M \(=\) 2.23, SD \(=\) 1.06) there was a significantly higher mental demand (F(1,73) \(=\) 6.536, p \(=\) 0.013) compared to the low task complexity condition (M \(=\) 1.76, SD \(=\) 1.05).

Physical demand. The interaction of the LOA mode and task complexity was significant with respect to the physical demand on the participants (F(1,73) \(=\) 8.6, p \(=\) 0.004). Using the low LOA mode was similarly demanding in low task complexity (M \(=\) 1.68, SD \(=\) 1.00) and high task complexity (M \(=\) 1.65, SD \(=\) 1.04). However, while using the high LOA mode, participants reported a lower physical demand in low task complexity (M \(=\) 1.16, SD \(=\) 0.38) compared to the high task complexity (M \(=\) 1.79, SD \(=\) 0.98), where participants reported an increased physical demand. The task complexity as a main effect had a significant influence (F(1,73) \(=\) 8.6, p \(=\) 0.047) on the physical demand in a pattern similar to the mental demand, in that the high complexity task was more physically demanding (M \(=\) 1.72, SD \(=\) 0.99) than the low complexity task (M \(=\) 1.42, SD \(=\) 0.79). Regardless of this, LOA mode as a main effect was not significant in terms of physical demand (F(1,73) \(=\) 0.008, p \(=\) 0.930).

Temporal Demand. The interaction effect of the LOA mode and task complexity was not significant (F(1,73) \(=\) 1.534, p \(=\) 0.219), neither was the LOA mode as a main effect (F(1,73) \(=\) 0.383, p \(=\) 0.538). The task complexity was significant as a main effect (F(1,73) \(=\) 6.136, p \(=\) 0.016). The high complexity task placed a higher temporal demand (M \(=\) 2.05, SD \(=\) 1.12) on the participants compared to the low complexity task (M \(=\) 1.63, SD \(=\) 0.85).

Operator Performance. Significant interaction was observed between LOA mode and task complexity (F(1,73) \(=\) 4.376, p \(=\) 0.04) with respect to operator performance. In low task complexity, participants reported higher subjective performance (M \(=\) 4.32, SD \(=\) 1.00) using the high LOA mode compared to the low LOA mode (M \(=\) 3.74, SD \(=\) 1.28). At high task complexity, however, a reverse trend was observed—the low LOA mode showed higher reported performance results (M \(=\) 4.50, SD \(=\) 0.76) compared to the high LOA mode (M \(=\) 4.16, SD \(=\) 1.21). The effect of LOA mode alone (F(1,73) \(=\) 0.307, p \(=\) 0.581) or task complexity alone (F(1,73) \(=\) 1.878, p \(=\) 0.175) did not have a significant effect on the reported performance.

Effort. The interaction between the LOA mode and task was not significant (F(1,73) \(=\) 0.875, p \(=\) 0.353), neither was LOA mode alone as a main effect (F(1,73) \(=\) 0.008, p \(=\) 0.927). As expected, in the high complexity task (M \(=\) 2.41, SD \(=\) 1.04), participants reported more exertion of effort (F(1,73) \(=\) 14.047, p \(\lt\) 0.01) than in the low complexity task (M \(=\) 1.71, SD \(=\) 0.90).

Frustration. The interaction effect of LOA mode and task complexity was significant (F(1,73) \(=\) 6.701, p \(=\) 0.012). In the low complexity task, participants reported higher frustration while using the low LOA mode (M \(=\) 2.00, SD \(=\) 1.16) compared to the high LOA mode (M \(=\) 1.47, SD \(=\) 0.70). The opposite was the case in the high complex task, where results revealed higher frustration in using the high LOA mode (M \(=\) 2.32, SD \(=\) 1.25) compared to when using the low LOA mode (M \(=\) 1.75, SD \(=\) 1.12).

The overall perceived workload rating for all workload dimensions confirmed that the difference in task complexity was significant (F(1,73) = 17.51, p < 0.01). The high complexity task placed a higher demand (M \(=\) 49.23, SD \(=\) 12.90) on the participants compared to the low complexity task (M \(=\) 40.96, SD \(=\) 10.31).

The correlation between participants’ video gaming experience and the aggregated perceived workload scores, though slightly negative, was not found to be significant (r \(=\)\(-\)0.091, n \(=\) 77, p \(=\) 0.431).

4.3 Situation Awareness

Complexity of Interaction. The interaction of the LOA mode and task complexity was not significant (F(1,73) \(=\) 2.378, p \(=\) 0.127).The LOA mode (F(1,73) \(=\) 0.741, p \(=\) 0.392) and task complexity (F(1,73) \(=\) 2.298, p \(=\) 0.134) also did not significantly influence the interaction complexity. However, participants perceived the interaction with the system to be more complex in the high complexity task when using the high LOA mode, which was not the case while using the low LOA mode.

Focus of Attention. LOA mode and task complexity had an interaction effect on participants’ focus of attention (F(1,68) \(=\) 4.649, p \(=\) 0.035). While performing the low complexity task, participants concentrated their attention on more aspects of the interaction while using the low LOA mode (M \(=\) 3.63, SD \(=\) 1.74) compared to the condition where they used the high LOA mode (M \(=\) 2.79, SD \(=\) 1.93). This focus of attention was reversed in the high complex task, where participants concentrated on more aspects of the interaction in the high LOA (M \(=\) 3.58, SD \(=\) 1.92) mode than in the low LOA mode (M \(=\) 3.15, SD \(=\) 1.84).

Information Quantity. The perceived information quantity reported by the participants refers to the difference in information the participants gained about the environment for different LOAs at different task complexities. The interaction of LOA and task complexity did not have a significant effect on the perceived information quantity. (F(1,73) \(=\) 0.179, p \(=\) 0.674). However, the pattern of information gained with respect to the LOA modes changed in the different task complexities, similar to the trend for focus of attention. In the low complexity task, participants reported that they gained more information about the environment when using the low LOA mode (M \(=\) 5.58, SD \(=\) 1.07) compared to the high LOA mode (M \(=\) 5.53, SD \(=\) 1.12). In the high complexity task, however, the opposite was the case—participants reported to have gained more information while using the high LOA (M \(=\) 5.42, SD \(=\) 1.02) relative to when using the low LOA mode (M \(=\) 5.25, SD \(=\) 1.37).

4.4 System Usability

An interaction effect was observed between the LOA mode and task complexity with respect to the confidence in the system while using it (F(1,73) \(=\) 5.067, p \(=\) 0.027). In the low complexity task, participants reported higher confidence using the high LOA mode (M \(=\) 4.53, SD \(=\) 0.61) compared to the low LOA mode (M \(=\) 4.00, SD \(=\) 0.94), whereas in the high complexity task participants reported higher confidence using the assisted teleoperation mode (M \(=\) 4.15, SD \(=\) 0.81 vs. M \(=\) 4.00, SD \(=\) 1.00).

The aspect of the SUS questionnaire evaluating the integration of the system’s various functions was also significant (F(1,73) \(=\) 4.013, p \(=\) 0.049) with respect to the LOA mode. Participants considered the system functions more integrated in the high LOA mode (M \(=\) 4.24, SD \(=\) 0.75) compared to the low LOA mode (M \(=\) 3.92, SD \(=\) 0.87). The influence of the LOA mode and task complexity was not significant in the other aspects of the SUS questionnaire. The LOA mode, however, significantly influenced the aggregated usability score for all the participants (F(1,73) \(=\) 4.174, p \(=\) 0.045). The usability scores were higher in the high LOA mode (M \(=\) 59.00, SD \(=\) 5.07) compared to the low LOA mode (M \(=\) 56.87, SD \(=\) 6.01). Although the mean scores were lower than the SUS-recommended 68% [6], 85% of the participants considered the system easy to use.

Skip 5DISCUSSION Section

5 DISCUSSION

The presented study examines LOA implementation and use in the context of an MRP system operated through a user interface. The primary contribution was the development and evaluation of two LOA designs—assisted teleoperation (low LOA) mode and autonomous (high LOA) mode to support users in different task complexity situations. Some of the results should be taken with caution due to the relatively small number of participants; it is possible that more comprehensive user studies provide a more definitive picture on the relation between LOA and task complexity. In addition, we acknowledge that we did not evaluate dependency on user background factors, such as participants’ age, gender, and video gaming experience, which may have introduced bias into the results.

5.1 Impact of LOA in MRP Systems

The influence of the implemented LOA modes was observed only between the different task complexity levels. As the task complexity increased, the LOA which engaged participants more in managing the robot’s functions yielded higher performance compared to the LOA with higher robot autonomy, in line with H1.

Previous studies [49] and Onnasch et al. [33], involving both expert and non-expert users revealed an overall improvement in performance with increasing automation for routine tasks. However, it was stated that in other situations involving tasks with more situational demands, critical decisions, and action implementation, the performance declined with higher automation. This can be associated with those situations in our study in which the higher LOA (autonomous navigation mode) produced higher performance in the low complexity (and less demanding) task. In the high complexity task, which demanded more critical decisions and actions and in which more automation failures occurred, the lower LOA (assisted teleoperation mode) yielded higher performance.

An increase in workload was observed as the task complexity increased while using the high LOA mode (in line with H2). The reason for this might be connected with the frustration experienced by the participants in the higher task complexity when the automation failed or did not perform as expected (as seen in the frustration dimension of the NASA-TLX). In these situations, participants switched to lower automation (as was possible in the adjustable-high LOA condition), perhaps to facilitate easier handling of some of these challenges as noted in [29] where a lower LOA was found to better facilitate easier interaction. The switching, however, incurs some switch costs [22, 52] which could have contributed to the frustration observed in the high LOA mode. This switch cost, may also have contributed to the reason why most of the participants did not switch when the task complexity was low.

Consistent with previous findings that lower LOA tends to improve the SA of users [15, 33], in the low complexity task, the assisted teleoperation mode appeared to provide better SA in terms of focus of attention and the information participants gained about the environment (in line with H3). Moreover, participants seemed to miss fewer details about the environment, as seen in the objective measure assessing missed objects. This relation was reversed in the more complex task, which required a higher degree of awareness—a higher level of automation produced higher SA, as participants seemed to be better able to detect the obstacles. This concurs with the view of [22], arguing that the outcome of LOA implementations may vary with different task demands and advocating for the characterization of these LOA models in different tasks, contexts, and situations in order to collate the prevalent trends for model improvements.

5.2 Users’ Attitudes Towards Automation

Most participants used the automated functions in the adjustable-high LOA condition as they were guided to use it. Their confidence in the automation appeared to increase with use, as observed in Section 4.4. However, along with this rise in confidence, one of the observed behaviors was overreliance on the automation without recognizing its limitations.

On the other hand, disuse of the automation was also observed, especially when the autonomous function failed, delayed, or did not perform as expected. Some of the participants who switched to the low LOA mode mentioned that the robot did not provide sufficient information in such situations to enable them to take prompt actions. This might be attributable to a sense of responsibility for system outcome, which can positively affect error detection tendencies by the human operator [21, 22].

Another observation was a form of satisfying behavior. Several participants did not exhaustively explore all control options, even though they could potentially yield better performance, e.g., the use of the map for autonomous navigation or the pointing plate (trackball-like interface) in the low LOA mode. Some simply continued using the conventional arrow control keys to navigate the robot since they were familiar with it and it presumably gave them a “good enough” outcome. This behavior of accepting a readily identifiable operational solution that meets some minimum level of performance can mediate the use of automation, as noted by Kaber et al. [21]. But it also has the tendency to lead to suboptimal solutions that could be detrimental to performance in general [22]. Another group of participants was observed using the low LOA for looking around during adjustable-high LOA trials, i.e., autonomous navigation, to a specific location and then low LOA for orientation and inspection within short distances. This indicates the benefit of the adjustable LOA, which afforded the possibility of switching to a different LOA at will to better support specific task demands.

The need for exploration is, in part, the reason why we decided to allow switching between LOAs in the adjustable-high LOA condition. For example, when looking for a specific object, it could be a good strategy to set a goal in the approximate area of interest where the object is and then switch upon arrival. Once there, a participant would, if necessary, rotate in place via teleoperation to inspect the site more closely. In another case, the participant might see something interesting while an autonomous navigation goal is active. By clicking on the video (and thereby issuing a teleoperation command), the navigation goal is automatically canceled and, as before, the interesting object can be examined closer, either by moving towards it with teleoperation or by setting a new goal in front of it. Although the inherent objective of the autonomous navigation, which is to reach the goal it is given in the most direct way possible (while avoiding collisions), may be considered as conflicting with the user’s tasks to explore, we believe that it is this low-effort switching which negates the apparent mismatch of objectives, as the goal of the navigation stack in the autonomous function can be frequently updated and realigned to suit the purpose of exploration.

Presumptions were also made by the participants regarding the level of SA that the camera afforded them. For example, in sequence 3 of the high complexity task, the participants were expected to inspect the whole floor area of the home environment for obstacles that could cause a fall. Some participants skipped the step, possibly because they presumed to have seen most of the floor area of the local environment while carrying out the preceding sequences. These instances were responsible for certain subtasks not being completed. The discrepancies observed between the reported SA ratings in these trials and the objective results concerning missed objects highlight this point. It further points to the necessity of evaluating the behavioral patterns of users, which could include assumptions to be considered in automation design (as suggested in [22]).

The usability assessment outcome was lower than the recommended 68% in SUS [6]. This may be due to the complexity of the user interface, which the initial training did not completely overcome. The participants’ comments reveal various areas of the design which could be amended to enhance the usability of the system. The comments are related to improved feedback from the interface, error handling, video quality of the robot’s cameras, and availability of zooming capabilities. Responses with regards to the ease of use and learnability of the LOA modes, however, revealed satisfactory outcomes. This highlights the potential for improved usability in complex tasks using the LOA options that are available (in line with H4).

Users seemed to adapt the control options within the LOA modes successfully, as suggested by their responses regarding satisfaction (detailed in Section 4.4 in the usability results). This aligned with the goal of including the control options in the design process to achieve seamless transitioning between control modes. Overall, participants’ responses concerning the control options provided within each LOA mode in the different task complexity conditions revealed a relatively high degree of satisfaction.

Skip 6CONCLUSIONS AND FUTURE WORK Section

6 CONCLUSIONS AND FUTURE WORK

The user study yielded valuable insights into participants’ preferences and which characteristics of the operator interface related to LOA should be modified to enhance the user experience and performance. The evaluation of the impact of LOA and task complexity on performance, workload, SA, and usability is therefore a contribution that emerges as a building block for further studies which may include other forms and dimensions of complexity changes. It aids the design of automation which is based on empirical work rather than ad hoc designs relying on speculations and presumptions. Furthermore, the effect of real-time switching between LOA modes during task operations has been investigated in relation to task complexity, as a basis for further development of this strategy in LOA development and continuous adaptation to various task demands.

It is recommended that human operators be kept in the loop in all LOA modes. In the present context, this translates to using LOA modes that keep the users more involved in the task (such as the low LOA mode) to avoid overreliance, while improving detection of potential failures or conflicts. In addition, the higher LOA mode can be improved in the future to minimize the need for switching and fine-tuning. In situations when it is desired or required, however, switching should be effortless and seamless, incurring as little penalty to users’ mental workload, SA, and trust as possible (following the results in Section 4.4).

The users’ attitudes towards the automation (Section 5.2) can inform some recommendations for further LOA development in MRP systems, which include clarity of feedback from the robot to ensure that users remain aware of its actions at all times and in all LOA modes. Options for error handling are further recommended to be included in the LOA modes as part of the fallback mechanisms to employ when the robot fails (according to the comments in Section 4.4). This encourages users to detect and resolve errors, provided the tools for resolution are made available.

Ongoing work is focused on increased transparency of the high LOA mode and visual markers indicating close proximity to obstacles in the front and on the sides. We will further investigate effective measures to increase users’ spatial awareness of the robot’s immediate surroundings. This will require tests with different types of interfaces and cameras, along with their positioning on the platform. Moreover, future work should investigate additional interaction functions, e.g., zooming functions for the camera and extensive feedback from the robot. By making these adjustments, we expect the performance and usability with the direct teleoperation mode to improve and task completion time to decrease. In this way, we hope to provide a step forward in advancing MRP systems into markets and enabling them to become viable tools that add value to people’s everyday lives.

REFERENCES

  1. [1] Allaban Anas Abou, Wang Maozhen, and Padır Taşkın. 2020. A systematic review of robotics research in support of in-home care for older adults. Information 11, 2 (2020), 75.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Beer Jenay, Fisk Arthur D., and Rogers Wendy A.. 2014. Toward a framework for levels of robot autonomy in human-robot interaction. Journal of Human-Robot Interaction 3, 2 (2014), 74. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. [3] Billings Charles E.. 1991. Human-centered Aircraft Automation: A Concept and Guidelines. Vol. 103885. National Aeronautics and Space Administration, Ames Research Center.Google ScholarGoogle Scholar
  4. [4] Boyd John R.. 1996. The essence of winning and losing. Unpublished Lecture Notes 12, 23 (1996), 123125.Google ScholarGoogle Scholar
  5. [5] Braarud Per Øivind. 2001. Subjective task complexity and subjective workload: Criterion validity for complex team tasks. International Journal of Cognitive Ergonomics 5, 3 (2001), 261273.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Brooke John. 1996. SUS-A quick and dirty usability scale. Usability Evaluation in Industry 189, 194 (1996), 47.Google ScholarGoogle Scholar
  7. [7] Cain Brad. 2007. A Review of the Mental Workload Literature. Technical Report. Defence Research And Development Toronto (Canada).Google ScholarGoogle Scholar
  8. [8] Campbell Donald J.. 1988. Task complexity: A review and analysis. Academy of Management Review 13, 1 (1988), 4052.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Cesta Amedeo, Cortellessa Gabriella, Orlandini Andrea, and Tiberio Lorenza. 2012. Evaluating telepresence robots in the field. In Proceedings of the International Conference on Agents and Artificial Intelligence. Springer, 433448.Google ScholarGoogle Scholar
  10. [10] Coradeschi S., Cesta A., Cortellessa G., Coraci L., Gonzalez J., Karlsson L., Furfari F., Loutfi A., Orlandini A., Palumbo F., Pecora F., Rump S. von, Stimec A., Ullberg J., and Otslund B.. 2013. GiraffPlus: Combining social interaction and long term monitoring for promoting independent living. In Proceedings of the 2013 6th International Conference on Human System Interactions. IEEE, 578585. DOI:DOI: DOI: https://doi.org/10.1109/HSI.2013.6577883Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Crandall Jacob W. and Goodrich Michael A.. 2002. Characterizing efficiency of human robot interaction: A case study of shared-control teleoperation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vol. 2. IEEE, 12901295.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Dekker Sidney W. A. and Woods David D.. 2002. MABA-MABA or abracadabra? Progress on human–automation co-ordination. Cognition, Technology & Work 4, 4 (2002), 240244.Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Endsley Mica R.. 2001. Designing for situation awareness in complex systems. In Proceedings of the 2nd International Workshop on Symbiosis of Humans, Artifacts and Environment. 114.Google ScholarGoogle Scholar
  14. [14] Endsley Mica R. and Kaber David B.. 1999. Level of automation effects on performance, situation awareness and workload in a dynamic control task. Ergonomics 42, 3 (1999), 462492.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Endsley Mica R. and Kiris Esin O.. 1995. The out-of-the-loop performance problem and level of control in automation. Human Factors 37, 2 (1995), 381394.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Goodrich Michael A., Olsen Dan R., Crandall Jacob W., and Palmer Thomas J.. 2001. Experiments in adjustable autonomy. In Proceedings of the IJCAI Workshop on Autonomy, Delegation and Control: Interacting with Intelligent Agents. 16241629.Google ScholarGoogle Scholar
  17. [17] Greenstone Michael and Nigam Vishan. 2020. Does social distancing matter? University of Chicago, Becker Friedman Institute for Economics Working Paper2020-26 (2020).Google ScholarGoogle Scholar
  18. [18] Gutman Dana, Olatunji Samuel, and Edan Yael. 2021. Evaluating levels of automation in human–robot collaboration at different workload levels. Applied Sciences 11, 16 (2021), 7340.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Hart Sandra G. and Staveland Lowell E.. 1988. Development of NASA-TLX (task load index): Results of empirical and theoretical research. In Proceedings of the Advances in Psychology. Vol. 52. Elsevier, 139183.Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Honig Shanee S., Oron-Gilad Tal, Zaichyk Hanan, Sarne-Fleischmann Vardit, Olatunji Samuel, and Edan Yael. 2018. Toward socially aware person-following robots. IEEE Transactions on Cognitive and Developmental Systems 10, 4 (2018), 936954.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Kaber David, Kaufmann Karl, Alexander Amy L., Kim Sang-Hwan, Naylor James T., III Lawrence J. Prinzel, Jr Carl Pankok, and Gil Guk-Ho. 2013. Testing and validation of a psychophysically defined metric of display clutter. Journal of Aerospace Information Systems 10, 8 (2013), 359368.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Kaber David B.. 2018. Issues in human–automation interaction modeling: Presumptive aspects of frameworks of types and levels of automation. Journal of Cognitive Engineering and Decision Making 12, 1 (2018), 724.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Kiselev Andrey, Kristoffersson Annica, and Loutfi Amy. 2014. Combining semi-autonomous navigation with manned behaviour in a cooperative driving system for Mobile robotic telepresence. In Proceedings of the European Conference on Computer Vision. Springer, 1728.Google ScholarGoogle Scholar
  24. [24] Kiselev Andrey, Kristoffersson Annica, Melendez Francisco, Galindo Cipriano, Loutfi Amy, Gonzalez-Jimenez Javier, and Coradeschi Silvia. 2015. Evaluation of using semi-autonomy features in mobile robotic telepresence systems. In Proceedings of the 2015 IEEE 7th International Conference on Cybernetics and Intelligent Systems and IEEE Conference on Robotics, Automation and Mechatronics. IEEE, 147152.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Kiselev Andrey and Loutfi Amy. 2012. Using a mental workload index as a measure of usability of a user interface for social robotic telepresence. In Proceedings of the 2nd Workshop of Social Robotic Telepresence in Conjunction with IEEE International Symposium on Robot and Human Interactive Communication 2012.Google ScholarGoogle Scholar
  26. [26] Kristoffersson Annica, Coradeschi Silvia, and Loutfi Amy. 2013. A review of mobile robotic telepresence. Advances in Human-Computer Interaction 2013 (2013), 3.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] Michaud Francois, Boissy Patrick, Labonte Daniel, Corriveau Helene, Grant Andrew, Lauria Michel, Cloutier Richard, Roux Marc-André, Iannuzzi Daniel, and Royer Marie-Pier. 2007. Telepresence robot for home care assistance. In Proceedings of the AAAI Spring Symposium: Multidisciplinary Collaboration for Socially Assistive Robotics. 5055.Google ScholarGoogle Scholar
  28. [28] Moray Neville. 1986. Monitoring behavior and supervisory control. In Proceedings of the Handbook of Perception and Human Performance. K. R. Boff, L. Kaufman, and J. P. Thomas (Eds.), Vol. 2.Google ScholarGoogle Scholar
  29. [29] Olatunji Samuel, Markfeld Noa, Gutman Dana, Givati Shai, Sarne-Fleischmann Vardit, Oron-Gilad Tal, and Edan Yael. 2019. Improving the interaction of older adults with a socially assistive table setting robot. In Proceedings of the International Conference on Social Robotics. Springer, 568577.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Olatunji Samuel, Oron-Gilad Tal, Markfeld Noa, Gutman Dana, Sarne-Fleischmann Vardit, and Edan Yael. 2021. Levels of automation and transparency: Interaction design considerations in socially assistive robots for older adults. IEEE Transactions on Human-Machine Systems 51, 6 (2021), 673–683.Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Olatunji Samuel, Oron-Gilad Tal, Sarne-Fleischmann Vardit, and Edan Yael. 2020. User-centered feedback design in person-following robots for older adults. Paladyn, Journal of Behavioral Robotics 11, 1 (2020), 86103.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Olsen Dan R. and Goodrich Michael A.. 2003. Metrics for evaluating human-robot interactions. In Proceedings of the PERMIS, Vol. 2003. 4.Google ScholarGoogle Scholar
  33. [33] Onnasch Linda, Wickens Christopher D., Li Huiyang, and Manzey Dietrich. 2014. Human performance consequences of stages and levels of automation: An integrated meta-analysis. Human Factors 56, 3 (2014), 476488.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Orlandini Andrea, Kristoffersson Annica, Almquist Lena, Björkman Patrik, Cesta Amedeo, Cortellessa Gabriella, Galindo Cipriano, Gonzalez-Jimenez Javier, Gustafsson Kalle, Kiselev Andrey, Loutfi Amy, Melendez Francisco, Nilsson Malin, Hedman Lasse Odens, Odontidou Eleni, Ruiz-Sarmiento Jose Raul, Scherlund Mårten, Tiberio Lorenza, Rump Stephen von, and Coradeschi Silvia. 2016. ExCITE Project: A Review of Forty-two Months of Robotic Telepresence Technology Evolution. Presence Teleoperators and Virtual Environments 25, 3 (July 2016), 204221. DOI: https://doi.org/10.1162/PRES_a_00262Google ScholarGoogle Scholar
  35. [35] Pandey Amit Kumar and Gelin Rodolphe. 2018. A mass-produced sociable humanoid robot: Pepper: The first machine of its kind. IEEE Robotics & Automation Magazine 25, 3 (2018), 4048.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Parasuraman Raja, Sheridan Thomas B., and Wickens Christopher D.. 2000. A model for types and levels of human interaction with automation. IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans 30, 3 (2000), 286297. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Parasuraman Raja, Sheridan Thomas B., and Wickens Christopher D.. 2008. Situation awareness, mental workload, and trust in automation: Viable, empirically supported cognitive engineering constructs. Journal of Cognitive Engineering and Decision Making 2, 2 (2008), 140160.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Potenza André and Saffiotti Alessandro. 2018. One robot and two humans: Some notes on shared autonomy in the case of robotic telepresence. In Proceedings of the IJCAI2018, Workshop on Autonomy in Teams.Google ScholarGoogle Scholar
  39. [39] Rasmussen Martin, Standal Martin Inge, and Laumann Karin. 2015. Task complexity as a performance shaping factor: A review and recommendations in standardized plant analysis risk-human reliability analysis (SPAR-H) adaption. Safety Science 76 (2015), 228238.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Redfoot Donald, Feinberg Lynn, and Houser Ari N.. 2013. The Aging of the Baby Boom and the Growing Care Gap: A Look at Future Declines in the Availability of Family Caregivers. AARP Public Policy Institute Washington, DC.Google ScholarGoogle Scholar
  41. [41] Rueangsirarak Worasak, Atkins Anthony S., Sharp Bernadette, Chakpitak Nopasit, and Meksamoot Komsak. 2012. Fall-risk screening system framework for physiotherapy care of elderly. Expert Systems with Applications 39, 10 (2012), 88598864. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. [42] Sauro J.. 2012. 10 things to know about the Single Ease Question (SEQ). Measuring U, 2012 (2012).Google ScholarGoogle Scholar
  43. [43] Sheridan Thomas B. and Verplank William L.. 1978. Human and Computer Control of Undersea Teleoperators. Technical Report. Massachusetts Inst of Tech Cambridge Man-Machine Systems Lab.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Shiarlis Kyriacos, Messias Joao, Someren Maarten van, Whiteson Shimon, Kim Jaebok, Vroon Jered Hendrik, Englebienne Gwenn, Truong Khiet Phuong, Pérez-Higueras Noé, Pérez-Hurtado Ignacio, Vanessa Evers , Fernando Caballero, Luis Merino, Jie Shen, Stavros Petridis, Lasse Hedman, Maja Pantic, Mårten Scherlund, Koster Raphael, and Hervé Michel. 2015. TERESA: A socially intelligent semi-autonomous telepresence system. In Proceedings of the Workshop on Machine Learning for Social Robotics 2015.Google ScholarGoogle Scholar
  45. [45] Steinfeld Aaron, Fong Terrence, Kaber David, Lewis Michael, Scholtz Jean, Schultz Alan, and Goodrich Michael. 2006. Common metrics for human-robot interaction. In Proceedings of the 1st ACM SIGCHI/SIGART Conference on Human-robot Interaction. 3340. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. [46] Taylor Richard M.. 2017. Situational awareness rating technique: The development of a tool for aircrew systems design. In Proceedings of the Situational Awareness. Routledge, 111128.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Vagia Marialena, Transeth Aksel A., and Fjerdingen Sigurd A.. 2016. A literature review on the levels of automation during the years. What are the different taxonomies that have been proposed? Applied Ergonomics 53, Part A (2016), 190202.Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Wickens Christopher D., Hollands Justin G., Banbury Simon, and Parasuraman Raja. 2015. Engineering Psychology and Human Performance. Psychology Press.Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Wickens Christopher D., Li Huiyang, Santamaria Amy, Sebok Angelia, and Sarter Nadine B.. 2010. Stages and levels of automation: An integrated meta-analysis. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Vol. 54. Sage Publications Sage CA: Los Angeles, CA, 389393.Google ScholarGoogle ScholarCross RefCross Ref
  50. [50] Wiener Earl L.. 1985. Cockpit automation: In need of a philosophy. SAE Transactions (1985), 952958.Google ScholarGoogle Scholar
  51. [51] Wood Robert E.. 1986. Task complexity: Definition of the construct. Organizational Behavior and Human Decision Processes 37, 1 (1986), 6082.Google ScholarGoogle ScholarCross RefCross Ref
  52. [52] Wylie Glenn and Allport Alan. 2000. Task switching and the measurement of “switch costs”. Psychological Research 63, 3 (2000), 212233.Google ScholarGoogle ScholarCross RefCross Ref
  53. [53] Zafrani Oded and Nimrod Galit. 2019. Towards a holistic approach to studying human–robot interaction in later life. The Gerontologist 59, 1 (2019), e26–e36.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Levels of Automation for a Mobile Robot Teleoperated by a Caregiver

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Human-Robot Interaction
        ACM Transactions on Human-Robot Interaction  Volume 11, Issue 2
        June 2022
        308 pages
        EISSN:2573-9522
        DOI:10.1145/3505213
        Issue’s Table of Contents

        Copyright © 2022 Copyright held by the owner/author(s).

        This work is licensed under a Creative Commons Attribution International 4.0 License.

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 8 February 2022
        • Accepted: 1 November 2021
        • Revised: 1 September 2021
        • Received: 1 July 2020
        Published in thri Volume 11, Issue 2

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format