Jul 24, 2023
Speaker · 0 followers
Speaker · 0 followers
Speaker · 0 followers
Speaker · 0 followers
Agents must be able to adapt quickly as an environment changes. We find that existing model-based reinforcement learning agents are unable to do this well, in part because of the way that they leverage their experience to train a world model. In particular, strategies that are effective for updating world models in unchanging environments, such as uniform sampling from a replay buffer, can become ineffective in changing environments. Here, we present Curious Replay, a method for improving the ability of model-based systems to adaptively explore and perform tasks in changing environments. We demonstrate our method's success in a nonstationary version of the widely-used Deepmind Control Suite, an exploration paradigm inspired by animal behavior, and the Crafter benchmark.Agents must be able to adapt quickly as an environment changes. We find that existing model-based reinforcement learning agents are unable to do this well, in part because of the way that they leverage their experience to train a world model. In particular, strategies that are effective for updating world models in unchanging environments, such as uniform sampling from a replay buffer, can become ineffective in changing environments. Here, we present Curious Replay, a method for improving the ab…
Professional recording and live streaming, delivered globally.
Presentations on similar topic, category or speaker
Jiequn Han, …
Sungbin Shin, …
Haotong Qin, …
Zuheng Xu, …