Quality-diversity based semi-autonomous teleoperation using reinforcement learning

  • Sangbeom Park
  • , Taerim Yoon
  • , Joonhyung Lee
  • , Sunghyun Park
  • , Sungjoon Choi*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)

Abstract

Recent successes in robot learning have significantly enhanced autonomous systems across a wide range of tasks. However, they are prone to generate similar or the same solutions, limiting the controllability of the robot to behave according to user intentions. These limited robot behaviors may lead to collisions and potential harm to humans. To resolve these limitations, we introduce a semi-autonomous teleoperation framework that enables users to operate a robot by selecting a high-level command, referred to as option. Our approach aims to provide effective and diverse options by a learned policy, thereby enhancing the efficiency of the proposed framework. In this work, we propose a quality-diversity (QD) based sampling method that simultaneously optimizes both the quality and diversity of options using reinforcement learning (RL). Additionally, we present a mixture of latent variable models to learn multiple policy distributions defined as options. In experiments, we show that the proposed method achieves superior performance in terms of the success rate and diversity of the options in simulation environments. We further demonstrate that our method outperforms manual keyboard control for time duration over cluttered real-world environments.

Original languageEnglish
Article number106543
JournalNeural Networks
Volume179
DOIs
Publication statusPublished - 2024 Nov

Bibliographical note

Publisher Copyright:
© 2024 The Authors

Keywords

  • Quality-diversity
  • Reinforcement learning
  • Shared autonomy
  • Teleoperation

ASJC Scopus subject areas

  • Cognitive Neuroscience
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Quality-diversity based semi-autonomous teleoperation using reinforcement learning'. Together they form a unique fingerprint.

Cite this