Optimizing Reinforcement Learning with Limited HRI Demonstrations: A Task-Oriented Weight Update Method with Analysis of Multi-Head and Layer Feature Combinations

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

To address the challenge of training reinforcement learning (RL) networks with limited data in Human-Robot Interaction (HRI), we introduce a novel task-oriented update method that combines meta-inverse reinforcement learning (Meta-IRL) and transformer encoder architectures. Our approach utilizes Meta-IRL to emulate expert trajectories, improving training efficiency by minimizing ineffective interactions. Through a systematic exploration of transformer encoder components, including varying heads and layers, we optimize feature extraction for sparse data representations. Experimental validation shows performance improvements in reinforcement learning networks trained with limited HRI expert demonstration data from Applied Behavior Analysis (ABA) scenarios.

Article activity feed