Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Jul 6:8:692811.
doi: 10.3389/frobt.2021.692811. eCollection 2021.

Becoming Team Members: Identifying Interaction Patterns of Mutual Adaptation for Human-Robot Co-Learning

Affiliations

Becoming Team Members: Identifying Interaction Patterns of Mutual Adaptation for Human-Robot Co-Learning

Emma M van Zoelen et al. Front Robot AI. .

Abstract

Becoming a well-functioning team requires continuous collaborative learning by all team members. This is called co-learning, conceptualized in this paper as comprising two alternating iterative stages: partners adapting their behavior to the task and to each other (co-adaptation), and partners sustaining successful behavior through communication. This paper focuses on the first stage in human-robot teams, aiming at a method for the identification of recurring behaviors that indicate co-learning. Studying this requires a task context that allows for behavioral adaptation to emerge from the interactions between human and robot. We address the requirements for conducting research into co-adaptation by a human-robot team, and designed a simplified computer simulation of an urban search and rescue task accordingly. A human participant and a virtual robot were instructed to discover how to collaboratively free victims from the rubbles of an earthquake. The virtual robot was designed to be able to real-time learn which actions best contributed to good team performance. The interactions between human participants and robots were recorded. The observations revealed patterns of interaction used by human and robot in order to adapt their behavior to the task and to one another. Results therefore show that our task environment enables us to study co-learning, and suggest that more participant adaptation improved robot learning and thus team level learning. The identified interaction patterns can emerge in similar task contexts, forming a first description and analysis method for co-learning. Moreover, the identification of interaction patterns support awareness among team members, providing the foundation for human-robot communication about the co-adaptation (i.e., the second stage of co-learning). Future research will focus on these human-robot communication processes for co-learning.

Keywords: co-adaptation; co-learning; emergent interactions; human-robot collaboration; human-robot team; interaction patterns.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

FIGURE 1
FIGURE 1
A storyboard describing how a human-robot team might free an earthquake victim from underneath a pile of rocks. In this storyboard, we can see how the robot picks up a large rock, unaware that this will cause another rock to fall on the head of the victim (panel C). The human notices the issue, and steps in to prevent the rock from falling (panel D). This event can help the robot learn about the task, and that it apparently made a mistake. The human can learn about the capabilities of the robot, namely that it didn’t understand how the rocks would fall and that it would cause harm.
FIGURE 2
FIGURE 2
The USAR task environment programmed in MATRX. It shows a victim underneath a pile of rocks, and a human and a robot representing the team members. The dashed red square (above the human’s head) represents the hand of the human that can be moved to pick up rocks. The dashed blue square represents the hand of the robot. Scene (A) was used as level 1 in the experiment, while scene (B) was used as level 2.
FIGURE 3
FIGURE 3
A flowchart showing the rule-based decision making the agent would go through when using Macro-Action 1.
FIGURE 4
FIGURE 4
A flowchart showing the rule-based decision making the agent would go through when using Macro-Action 2.
FIGURE 5
FIGURE 5
A flowchart showing the rule-based decision making the agent would go through when using Macro-Action 3.
FIGURE 6
FIGURE 6
An overview of the representation of the learning problem embedded in the experiment. It shows the different runs that a participant went through (5 runs for level 1, 3 runs for level 2), as well as how the runs were separated into 4 phases defined by the Phase Variables. The colors show how in R1.1, R1.2 and R1.3, the robot usually used O1—picking up all, O2—passive large rocks and O3—breaking respectively in each phase. From R1.4 onwards, the robot would choose a Macro-action based on the learned Q-values. The Future Run portrays the behavior that the robot would engage in if there were another run, based on the Q-values after R2.3.
FIGURE 7
FIGURE 7
The Collaboration Fluency scores per run in the experiment for all participants.
FIGURE 8
FIGURE 8
An overview of how often certain Macro-actions were chosen by the robot across all participants per phase (A) and per run (B).
FIGURE 9
FIGURE 9
An overview of how many participants used specific behavioral strategies per run.
FIGURE 10
FIGURE 10
An overview of how often certain Macro-actions were chosen by the robot across all participants per run, split up by the level adaptation the participant showed: (A) shows participants who adapted by actively using O2—passive large rocks and/or O3—breaking, (B) shows participants who adapted by balancing waiting and acting, and (C) shows participants who did not adapt.

References

    1. Ansari F., Erol S., Sihn W. (2018). Rethinking Human-Machine Learning in Industry 4.0: How Does the Paradigm Shift Treat the Role of Human Learning? Proced. Manufacturing 23, 117–122. 10.1016/j.promfg.2018.04.003 - DOI
    1. Baxter P., Ashurst E., Read R., Kennedy J., Belpaeme T. (2017). Robot Education Peers in a Situated Primary School Study: Personalisation Promotes Child Learning. PLoS ONE 12 (5), e0178126. 10.1371/journal.pone.0178126 - DOI - PMC - PubMed
    1. Belpaeme T., Vogt P., van den Berghe R., Bergmann K., Göksun T., de Haas M., et al. (2018). Guidelines for Designing Social Robots as Second Language Tutors. Int. J. Soc. Robotics 10 (3), 325–341. 10.1007/s12369-018-0467-6 - DOI - PMC - PubMed
    1. Bosch K. v. d., Schoonderwoerd T., Blankendaal R., Neerincx M. (2019). Six Challenges for Human-AI Co-learning. Lecture Notes Comput. Sci. 11597, 572–589. 10.1007/978-3-030-22341-0_45 - DOI
    1. Burke C. S., Stagl K. C., Salas E., Pierce L., Kendall D. (2006). Understanding Team Adaptation: A Conceptual Analysis and Model. J. Appl. Psychol. 91 (December), 1189–1207. 10.1037/0021-9010.91.6.1189 - DOI - PubMed

LinkOut - more resources