Human operator cognitive availability aware Mixed-Initiative control

Page created by Martin Walters
 
CONTINUE READING
Human operator cognitive availability aware Mixed-Initiative control
Human operator cognitive availability aware Mixed-Initiative control
                                                             Giannis Petousakis1 , Manolis Chiou2 , Grigoris Nikolaou1 , and Rustam Stolkin2

                                            Abstract— This paper presents a Cognitive Availability         suggestions [8]. Hence, the agents lack the ability to override
                                         Aware Mixed-Initiative Controller for remotely operated mobile    each other’s actions. Additionally, the robot’s AI initiative
                                         robots. The controller enables dynamic switching between          is often limited within a specific LOA, e.g. in safe mode
                                         different levels of autonomy (LOA), initiated by either the AI
                                         or the human operator. The controller leverages a state-of-the-   [7]. Research on MI systems that are able to switch LOA
                                         art computer vision method and an off-the-shelf web camera        dynamically is fairly limited. Moreover, some of the MI
arXiv:2108.11885v1 [cs.RO] 26 Aug 2021

                                         to infer the cognitive availability of the operator and inform    systems proposed are not experimentally evaluated e.g. [9].
                                         the AI-initiated LOA switching. This constitutes a qualitative       Our previous work [10] proposed a novel AI ”expert-
                                         advancement over previous Mixed-Initiative (MI) controllers.      guided” MI controller and identified some major challenges:
                                         The controller is evaluated in a disaster response experiment,
                                         in which human operators have to conduct an exploration           the conflict for control between the operator and the robot’s
                                         task with a remote robot. MI systems are shown to effectively     AI and the design of context aware MI controllers. In this
                                         assist the operators, as demonstrated by quantitative and         paper we extend the MI controller proposed in [10] in
                                         qualitative results in performance and workload. Additionally,    order to tackle the above mentioned challenges by using
                                         some insights into the experimental difficulties of evaluating    information on the cognitive availability of the human oper-
                                         complex MI controllers are presented.
                                                                                                           ator and by improving the controller’s assumptions. We call
                                                                I. I NTRODUCTION                           our approach Cognitive Availability Aware Mixed-Initiative
                                            In this paper we follow a Variable Autonomy (VA) ap-           (CAA-MI) control. Cognitive availability indicates if the
                                         proach in order to improve Human-Robot Systems (HRS)              operator’s attention is available to focus on controlling the
                                         that are being increasingly used in high risk, safety-critical    robot.
                                         applications such as disaster response. VA in this context           In our work, cognitive availability inference is based on
                                         refers to systems with autonomous capabilities of varying         the operator’s head pose estimation provided by a state-of-
                                         scales.                                                           the-art deep learning computer vision algorithm. A low cost,
                                            The task of manually controlling the robot combined            off-the-self webcamera mounted on the Operator Control
                                         with various operator straining factors, increase the cogni-      Unit (OCU) provides video streaming to the computer vision
                                         tive workload of the operator [1]. Moving towards robotic         algorithm in real time. The head pose estimation provides
                                         systems that can actively assist the operators and reduce         input to the fuzzy CAA-MI controller where the cognitive
                                         their workload can lead in reduced errors and improved            availability inference and the LOA switching decisions are
                                         performance [1], [2]. The advantage of such HRS lies in           made. Somewhat related to our paper is the work of Gateau
                                         the complementing capabilities of humans and Artificial           et al. [8] which uses cognitive availability to inform decisions
                                         Intelligence (AI). In a VA system control can be traded           on asking the operator for help. Gateau et al. [8] uses
                                         between a human operator and a robot by switching between         a specialized eye tracker and does not involve cognitive
                                         different Levels of Autonomy [3], e.g. the robot can navigate     availability informed LOA switching.
                                         autonomously while the operator is multitasking. Levels              This work is explicitly making use of operator’s cognitive
                                         of Autonomy (LOAs) refer to the degree to which the               availability in MI controller’s LOA switching decision pro-
                                         robot, or any artificial agent, takes its own decisions and       cess and is contributing by: a) extending the MI controller
                                         acts autonomously [4]. This work addresses the problem of         in [10] to explicitly include cognitive availability and active
                                         dynamically changing LOA during task execution using a            LOA information; b) including those parameters provides
                                         form of VA called Mixed-Initiative control. Mixed-initiative      a qualitative advancement that tackles assumptions of the
                                         (MI) refers to a VA system where both the human operator          original controller; c) providing proof of concept on using
                                         and the robot’s AI have authority to initiate actions and to      state-of-the-art computer vision methods for informing MI
                                         change the LOA [5].                                               control on operator’s status.
                                            In many VA systems found in literature the LOA is chosen       II. E XPERT- GUIDED M IXED -I NITIATIVE CONTROL WITH
                                         during the initialization of the system and cannot change                   COGNITIVE AVAILABILITY INFERENCE
                                         on the fly [6], [7]. Other systems allow only the operator
                                         to initiate actions, although these are based on system’s            The MI control problem this paper addresses is allowing
                                                                                                           dynamic LOA switching by either the operator or the robot’s
                                           1 University   of West Attica, Greece gspetou@gmail.com,        AI towards improving the HRS performance. In this work we
                                         nikolaou@uniwa.gr                                                 assume a HRS which has two LOAs: a) teleoperation, where
                                           2 Extreme Robotics Lab (ERL) and National Center for Nuclear
                                         Robotics (NCNR), University of Birmingham, UK {m.chiou,           the human operator has manual control over the robot via a
                                         r.stolkin}@bham.ac.uk                                             joystick; b) autonomy, where the operator clicks on a desired
Human operator cognitive availability aware Mixed-Initiative control
B. Cognitive availability via head pose estimation
                                                                    In this work we consider the information on whether the
                                                                 operator is attending or not at the Graphical User Interface
                                                                 (GUI) to be directly relevant to his current cognitive avail-
                                                                 ability. The CAA-MI controller presented here infers the
                                                                 operator’s cognitive availability by monitoring the operator’s
                                                                 head pose via a computer vision algorithm. By introducing
                                                                 cognitive availability perception capabilities to the HRS
                                                                 system, we make a qualitative change to the robot’s ability to
                                                                 perceive information about the operator and his availability
                                                                 to control the robot.
     Fig. 1.   The block diagram of the CAA-MI control system.      By using a computer vision technique for inferring the
                                                                 cognitive availability of the operator we avoid intrusive
                                                                 methods such as biometric sensors (e.g. EEG) and the use
location on the map with the robot autonomously executing        of cumbersome apparatus. Our aim is to keep the approach
a trajectory to that location.                                   generalizable and versatile. The latest head pose estimation
   Operator’s initiated LOA switches are based on their          techniques are robust in real time use and do not require
judgment. Previous work [3], [11] showed that humans are         specialized image capturing equipment [12], [13], contrary
able to determine when they need to change LOA in order to       to most state-of-the-art gaze tracking implementations (e.g.
improve performance. AI initiated LOA switches are based         [14]). We chose head pose estimation over gaze tracking
on a fuzzy inference engine. The controller’s two states         preferring perception of more pronounced natural cues.
are: a) switch LOA; b) do not switch LOA. The controller            Our focus is on using a robust CV algorithm to provide
uses four input variables: a) an online task effectiveness       cognitive availability input to the controller rather than eval-
metric for navigation, the goal-directed motion error; b)        uating the various CV algorithms. The Deepgaze algorithm
the cognitive availability information provided via the deep     [12] is used in our system for head pose estimation. It makes
learning computer vision algorithm; c) the currently active      use of state-of-the-art methods such as a pre-trained Convo-
LOA; d) the current speed of the robot. Please refer to Figure   lutional Neural Network (CNN), is open source and actively
1 for the block diagram of the system.                           maintained. In our implementation, deepgaze measures the
                                                                 operator’s head rotation on the vertical axis (yaw). The head
A. Goal-directed motion error                                    rotation was filtered through an exponential moving average
                                                                 (EMA) [15] in order to reduce noise in the measurements.
   The CAA-MI controller uses an expert-guided approach
                                                                 The parameters of the EMA were calculated by analyzing
to initiate LOA switches. It assumes the existence of an AI
                                                                 data regarding the attention of the operator in the secondary
task expert that given a navigational goal can provide the
                                                                 task in previous experiments. Lastly, baseline data on what
expected task performance. The comparison between the run-
                                                                 values of rotation constitute attending or not at the GUI
time performance of the system with the expected expert
                                                                 were gathered in a pilot experiment and in a standardized
performance yields an online task effectiveness metric. This
                                                                 way. These were mapped into the fuzzy input for cognitive
metric expresses how effectively, the system, performs the
                                                                 availability.
navigation task.
   The controller’s task effectiveness metric is the goal-       C. The fuzzy rule base
directed motion error (refereed to as “error” from now on)          A fuzzy bang-bang controller is used with the Largest
and is the difference between the robot’s current motion (i.e.   of Maxima defuzzification method. We have introduced the
speed in this case) and the motion of the robot required to      active LOA and the Cognitive availability of the operator
achieve its goal according to an expert. To provide more         as fuzzy inputs. A hierarchical fuzzy approach is used for
context we extract the expert motion performance from a          the CAA-MI controller, meaning that the first in order rule
concurrently active navigation system which is given an          activated has priory over the others. The cognitive availability
idealized (unmapped obstacle and noise-free) view of the         of the operator and the active LOA take precedence in the
robot’s world. In essence, the expert provides an idealized      fuzzy rule activated. The rules added to the original MI
model of possible robot behavior that can be seen as an upper    controller make sure that when the operator is not attending
bound on system performance.                                     the screen the robot will operate in autonomy, with the rest
   Additionally, our controller encodes expert knowledge         of the rule base architecture being similar to the previous MI
from human operators data learned using machine learning         controller.
techniques from a previous experiment [3] on: a) what is
considered to be a large enough error to justify a LOA                        III. E XPERIMENTAL EVALUATION
switch; and b) the time window in which that error needs            This experiment evaluates the CAA-MI controller and
to accumulate. For detailed information the reader is encour-    investigates its potential advantages over the previously built
aged to read our previous work [10].                             MI controller [10] (referred simply as MI controller for the
the MI controller. Counterbalancing was used in the order
                                                                                  of the two controllers for different participants. Each par-
                                                                                  ticipant underwent extensive standardized training ensuring
                                                                                  a minimum skill level and understanding of the system.
                                                                                  Participants were instructed to perform the primary task as
                                                                                  quickly and safely as possible. They were also instructed that
                  (a)                                   (b)                       when presented with the secondary task, they should do it as
                                                                                  quickly and as accurately as possible. They were explicitly
Fig. 2. 2(a): The experimental apparatus: composed of a mouse and a               told that they should give priority to the secondary task over
joystick, a laptop and a desktop computer, a screen showing the GUI; a
web-camera mounted on the screen; and a laptop presenting the secondary           the primary task and should only perform the primary task
task. 2(b): A typical example of the secondary task.                              if the workload allowed. Lastly, after every trial, participants
                                                                                  completed a raw NASA-TLX workload questionnaire.

                                                                                                            IV. R ESULTS
                                                                                  A. Statistical analysis
                                                                                     The CAA-MI controller displayed a trend, though not
                                                                                  statistically significant (Wilcoxon signed-rank tests and t-
                                                                                  tests), of improved performance over the MI controller in
                                                                                  terms of lower number of LOA switches, more secondary
                                                                                  task number of correct answers and less perceived workload
                                                                                  (NASA-TLX), as demonstrated in table I.

                                                                                                               TABLE I
Fig. 3. The GUI: left, video feed from the camera, the LOA in use and               TABLE SHOWING DESCRIPTIVE STATISTICS FOR ALL THE METRICS .
the status of the robot; right, the map showing the position of the robot, the
current goal (blue arrow), the AI planned path (green line), the obstacles’           metric                descriptive statistics
laser reflections (red) and the walls (black). The letters denote the waypoints       primary task          MI: M = 242 sec, SD = 21.8
to be explored.                                                                       completion time       CAA-MI: M = 243.5 sec, SD = 24.4
                                                                                                            MI: M = 7.6, SD = 3.1
                                                                                      secondary task no.
                                                                                                            CAA-MI: M = 8.5, SD = 3.5
                                                                                      of correct answers
                                                                                                            Baseline: M = 8.6, SD = 3.1
rest of the paper). The Gazebo simulated robot was equipped                           secondary task        MI: M = 84.2, SD = 15.6
                                                                                      accuracy (% of        CAA-MI: M = 87.2, SD = 18.9
with a laser range finder and a RGB camera. The software                              correct answers)      Baseline: M = 82.9, SD = 14.2
was developed in Robot Operating System (ROS) and is                                  NASA-TLX
                                                                                                            MI: M = 34.9, SD = 17.4
described in more detail in [10]. The ROS code for the CAA-                           (Lower score means
                                                                                                            CAA-MI: M = 30.4, SD = 14.97
MI controller is provided under MIT license in our repository                         less workload)
                                                                                      number of             MI: M = 6.4, SD = 3.8
[16]. The robot was controlled via an Operator Control Unit                           LOA switches          CAA-MI: M = 5.5, SD = 3.4
(OCU) (see Figure 2(a)) including a GUI (Figure 3). The                               number of AI          MI: M = 2, SD = 1.6
experiment’s test arena was approximately 24m × 24m. The                              LOA switches          CAA-MI: M = 1.75, SD = 1.5
software used for the secondary task was the OpenSesame
[17] and the images used as stimuli were previously validated
for mental rotation tasks in [18].                                                B. Discussion
   The participants were tasked with a primary exploration                           Performance towards the primary task was on the same
task and a cognitively demanding secondary task. The pri-                         level for both controllers. This can be attributed to the
mary task was the exploration of the area along a set of                          varied exploration strategies of the operators as supported
waypoints in order to identify the number of cardboard                            by anecdotal evidence. Additionally, it is possible that the
boxes placed on those waypoints. In the secondary task                            performance degradation periods did not last enough to
the operators were presented with pairs of 3D objects. The                        amplify the advantages of the CAA-MI controller compared
operators were required to verbally state whether or not the                      to the MI one. By increasing secondary task duration and
two objects were identical or mirrored.                                           complexity, clearer insight to the advantages of the CAA-
   Artificially generated sensor noise was used to degrade the                    MI controller can be provided. Given the above evidence,
performance of the autonomous navigation. The secondary                           a trade-off must be found between having a meaningful
task was used to degrade the performance of the operator.                         exploration task (i.e. to be used for system evaluation) and
Each of these performance degrading situations occurred                           minimizing variance in operator strategies.
once concurrently (i.e. overlapping) but at random.                                  During MI controller trials and while the secondary task
   A total of 8 participants participated in a within-groups                      overlapped with the sensor noise, the robot would switch
design in which every participant performed two identical                         from autonomy to teleoperation. This negatively affected
trials, one with each controller: the CAA-MI controller and                       operators’ focus on the secondary task with some of them
expressing their frustration. In contrast, this was not ob-                                     R EFERENCES
served in the case of the CAA-MI controller since the robot         [1] J. L. Casper and R. R. Murphy, “Human-Robot Interactions During
could perceive that the operator was unavailable and hence              the Robot-Assisted Urban Search and Rescue Response at the World
maintained, or switched to, autonomy. Supporting evidence               Trade Center,” IEEE Transactions on Systems, Man, and Cybernetics,
                                                                        Part B: Cybernetics, vol. 33, no. 3, pp. 367–385, 2003.
can be found in the NASA-TLX frustration scale which                [2] H. A. Yanco, A. Norton, W. Ober, D. Shane, A. Skinner, and
while not statistically significant, showed lower frustration           J. Vice, “Analysis of Human-robot Interaction at the DARPA Robotics
in CAA-MI (M = 27.5, SD = 21.9) compared to MI                          Challenge Trials,” Journal of Field Robotics, vol. 32, no. 3, pp. 420–
                                                                        444, 2015.
(M = 33.25, SD = 18.3). Participants also performed better          [3] M. Chiou, R. Stolkin, G. Bieksaite, N. Hawes, K. L. Shapiro, and T. S.
on the secondary task. In conjunction with the above, a trend           Harrison, “Experimental analysis of a variable autonomy framework
for lower LOA switches has been observed in the case of                 for controlling a remotely operating mobile robot,” IEEE International
                                                                        Conference on Intelligent Robots and Systems, pp. 3581–3588, 2016.
the CAA-MI controller. This is a possible indication of more        [4] T. B. Sheridan and W. L. Verplank, “Human and computer control of
efficient LOA switching compared to the MI as the operators             undersea teleoperators,” MIT Man-Machine Systems Laboratory, 1978.
relied more on the CAA-MI system’s capabilities.                    [5] S. Jiang and R. C. Arkin, “Mixed-Initiative Human-Robot Interaction:
                                                                        Definition , Taxonomy , and Survey,” in IEEE International Confer-
   Lastly, note that our previous work [3], [10] has already            ence on Systems, Man, and Cybernetics (SMC), 2015, pp. 954–961.
demonstrated that dynamic LOA switching (e.g. MI control)           [6] C. W. Nielsen, D. A. Few, and D. S. Athey, “Using mixed-initiative
is significantly better than pure teleoperation or/and pure             human-robot interaction to bound performance in a search task,” in
                                                                        IEEE International Conference on Intelligent Sensors, Sensor Net-
autonomy. Hence, it can be inferred that the CAA-MI                     works and Information Processing, 2008, pp. 195–200.
controller is also advantageous over pure teleoperation or          [7] D. J. Bruemmer, D. A. Few, R. L. Boring, J. L. Marble, M. C.
pure autonomy.                                                          Walton, and C. W. Nielsen, “Shared Understanding for Collaborative
                                                                        Control,” IEEE Transactions on Systems, Man, and Cybernetics - Part
            V. C ONCLUSIONS AND FUTURE WORK                             A: Systems and Humans, vol. 35, no. 4, pp. 494–504, 2005.
                                                                    [8] T. Gateau, C. P. C. Chanel, M.-h. Le, and F. Dehais, “Considering
   This paper presented a Cognitive Availability Aware                  Human’s Non-Deterministic Behavior and his Availability State When
Mixed-Initiative (CAA-MI) controller and its experimental               Designing a Collaborative Human-Robots System,” in IEEE Interna-
evaluation. The CAA-MI controller’s advantage lies with the             tional Conference on Intelligent Robots and Systems (IROS), 2016, pp.
                                                                        4391–4397.
use of operator’s cognitive availability status into the AI’s       [9] D. J. Bruemmer, J. L. Marble, D. D. Dudenhoeffer, M. O. Anderson,
LOA switching decision process.                                         and M. D. Mckay, “Mixed-initiative control for remote characterization
   Research in MI control often requires integrating different          of hazardous environments,” in 36th Annual Hawaii International
                                                                        Conference on System Sciences, 2003, pp. 9 pp.–.
advanced algorithms in robotics and AI. Thus, the use of           [10] M. Chiou, N. Hawes, and R. Stolkin, “Mixed-Initiative variable
a state-of-the-art deep learning computer vision algorithm,             autonomy for remotely operated mobile robots,” arXiv preprint
demonstrated proof of concept that such algorithms can                  1911.04848, 2019. [Online]. Available: http://arxiv.org/abs/1911.
                                                                        04848
provide reliable and low cost advancements to MI control.          [11] M. Chiou, G. Bieksaite, N. Hawes, and R. Stolkin, “Human-Initiative
   Experimental results showed that the CAA-MI controller               Variable Autonomy: An Experimental Analysis of the Interactions
performed at least as good as the MI controller in the primary          Between a Human Operator and a Remotely Operated Mobile Robot
                                                                        which also Possesses Autonomous Capabilities,” AAAI Fall Sympo-
exploration task. Also trends found in the results (e.g.                sium Series: Shared Autonomy in Research and Practice, pp. 304–310,
NASA-TLX frustration scale, number of LOA switches) and                 2016.
anecdotal evidence indicate a more efficient LOA switching         [12] M. Patacchiola and A. Cangelosi, “Head pose estimation in the wild
                                                                        using Convolutional Neural Networks and adaptive gradient methods,”
in certain situations compared to the MI controller. This is            Pattern Recognition, vol. 71, pp. 132–143, 2017.
due to the qualitative advantage that the operator’s cognitive     [13] N. Ruiz, E. Chong, and J. M. Rehg, “Fine-grained head pose esti-
status information gives to the CAA-MI compared to the MI               mation without keypoints,” IEEE Computer Society Conference on
                                                                        Computer Vision and Pattern Recognition Workshops, vol. 2018-June,
controller.                                                             pp. 2155–2164, 2018.
   Concluding, we identify the current experimental                [14] J. Lemley, A. Kar, A. Drimbarean, and P. Corcoran, “Convolutional
paradigms as a limitation in testing the full potential of              neural network implementation for eye-gaze estimation on low-quality
                                                                        consumer imaging systems,” IEEE Transactions on Consumer Elec-
more complex MI systems. This is despite the recent                     tronics, vol. 65, no. 2, pp. 179–187, 2019.
advances on experimental frameworks [3]. Evaluating                [15] R. G. Brown, Smoothing, forecasting and prediction of discrete time
MI Human-Robot Systems in realistic scenarios involves                  series. Courier Corporation, 1963.
                                                                   [16] G. Petousakis and M. Chiou, “Cognitive availability aware
difficult intrinsic confounding factors that are hard to predict        mixed-initiative controller code,” https://github.com/uob-erl/fuzzy mi
or overcome. Hence, further research should devise a more               controller.git, 2019.
complex experimental protocol along the lines proposed in          [17] S. Mathôt, D. Schreij, and J. Theeuwes, “OpenSesame: An open-
                                                                        source, graphical experiment builder for the social sciences,” Behavior
the discussion, i.e. finding a trade-off between realism and            Research Methods, vol. 44, no. 2, pp. 314–324, 2012.
meaningful scientific inference.                                   [18] G. Ganis and R. Kievit, “A New Set of Three-Dimensional Shapes
                                                                        for Investigating Mental Rotation Processes : Validation Data and
                   ACKNOWLEDGMENT                                       Stimulus Set,” Journal of Open Psychology Data, vol. 3, no. 1, 2015.
  This work was supported by the following grants of
UKRI-EPSRC: EP/P017487/1 (Remote Sensing in Extreme
Environments); EP/R02572X/1 (National Centre for Nuclear
Robotics); EP/P01366X/1 (Robotics for Nuclear Environ-
ments). Stolkin was also sponsored by a Royal Society
Industry Fellowship.
You can also read