Regularizing Trajectory Optimization with Denoising Autoencoders (Paper Explained)

Опубликовано: 24 Май 2020
на канале: Yannic Kilcher
5,471
191

Can you plan with a learned model of the world? Yes, but there's a catch: The better your planning algorithm is, the more the errors of your world model will hurt you! This paper solves this problem by regularizing the planning algorithm to stay in high probability regions, given its experience.

https://arxiv.org/abs/1903.11981

Interview w/ Harri:    • Harri Valpola: System 2 AI and Planni...  

Abstract:
Trajectory optimization using a learned model of the environment is one of the core elements of model-based reinforcement learning. This procedure often suffers from exploiting inaccuracies of the learned model. We propose to regularize trajectory optimization by means of a denoising autoencoder that is trained on the same trajectories as the model of the environment. We show that the proposed regularization leads to improved planning with both gradient-based and gradient-free optimizers. We also demonstrate that using regularized trajectory optimization leads to rapid initial learning in a set of popular motor control tasks, which suggests that the proposed approach can be a useful tool for improving sample efficiency.

Authors: Rinu Boney, Norman Di Palo, Mathias Berglund, Alexander Ilin, Juho Kannala, Antti Rasmus, Harri Valpola

Links:
YouTube:    / yannickilcher  
Twitter:   / ykilcher  
BitChute: https://www.bitchute.com/channel/yann...
Minds: https://www.minds.com/ykilcher


Смотрите видео Regularizing Trajectory Optimization with Denoising Autoencoders (Paper Explained) онлайн без регистрации, длительностью часов минут секунд в хорошем качестве. Это видео добавил пользователь Yannic Kilcher 24 Май 2020, не забудьте поделиться им ссылкой с друзьями и знакомыми, на нашем сайте его посмотрели 5,471 раз и оно понравилось 191 людям.