Learning to Solve Tasks with Exploring Prior Behaviour

Ruiqi Zhu, Siyuan Li, Tianhong Dai, Chongjie Zhang, Oya Celiktutan

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

Abstract

Demonstrations are widely used in Deep Reinforcement Learning (DRL) for facilitating solving tasks with sparse rewards. However, the tasks in real-world scenarios can often have varied initial conditions from the demonstration, which would require additional prior behaviours. For example, consider we are given the demonstration for the task of \emph{picking up an object from an open drawer}, but the drawer is closed in the training. Without acquiring the prior behaviours of opening the drawer, the robot is unlikely to solve the task. To address this, in this paper we propose an Intrinsic Rewards Driven Example-based Control \textbf{(IRDEC)}. Our method can endow agents with the ability to explore and acquire the required prior behaviours and then connect to the task-specific behaviours in the demonstration to solve sparse-reward tasks without requiring additional demonstration of the prior behaviours. The performance of our method outperforms other baselines on three navigation tasks and one robotic manipulation task with sparse rewards. Codes are available at this https URL. https://github.com/Ricky-Zhu/IRDEC.
Original languageEnglish
Title of host publication2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
PublisherIEEE Explore
Number of pages7
ISBN (Electronic)978-1-6654-9190-7
DOIs
Publication statusPublished - 13 Dec 2023
EventIEEE?RSJ International Conference on Intelligent Robots and Systems - Detroit, United States
Duration: 1 Oct 20235 Oct 2023
https://ieee-iros.org/

Conference

ConferenceIEEE?RSJ International Conference on Intelligent Robots and Systems
Country/TerritoryUnited States
CityDetroit
Period1/10/235/10/23
Internet address

Fingerprint

Dive into the research topics of 'Learning to Solve Tasks with Exploring Prior Behaviour'. Together they form a unique fingerprint.

Cite this