arXiv Open Access 2024

Policy Decorator: Model-Agnostic Online Refinement for Large Policy Model

Xiu Yuan Tongzhou Mu Stone Tao Yunhao Fang Mengke Zhang +1 lainnya
Lihat Sumber

Abstrak

Recent advancements in robot learning have used imitation learning with large models and extensive demonstrations to develop effective policies. However, these models are often limited by the quantity, quality, and diversity of demonstrations. This paper explores improving offline-trained imitation learning models through online interactions with the environment. We introduce Policy Decorator, which uses a model-agnostic residual policy to refine large imitation learning models during online interactions. By implementing controlled exploration strategies, Policy Decorator enables stable, sample-efficient online learning. Our evaluation spans eight tasks across two benchmarks-ManiSkill and Adroit-and involves two state-of-the-art imitation learning models (Behavior Transformer and Diffusion Policy). The results show Policy Decorator effectively improves the offline-trained policies and preserves the smooth motion of imitation learning models, avoiding the erratic behaviors of pure RL policies. See our project page (https://policydecorator.github.io) for videos.

Topik & Kata Kunci

Penulis (6)

X

Xiu Yuan

T

Tongzhou Mu

S

Stone Tao

Y

Yunhao Fang

M

Mengke Zhang

H

Hao Su

Format Sitasi

Yuan, X., Mu, T., Tao, S., Fang, Y., Zhang, M., Su, H. (2024). Policy Decorator: Model-Agnostic Online Refinement for Large Policy Model. https://arxiv.org/abs/2412.13630

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓