Regularizing Trajectory Optimization with Denoising Autoencoders (Paper Explained)

Published: 24 May 2020
on channel: Yannic Kilcher
5,471
191

Can you plan with a learned model of the world? Yes, but there's a catch: The better your planning algorithm is, the more the errors of your world model will hurt you! This paper solves this problem by regularizing the planning algorithm to stay in high probability regions, given its experience.

https://arxiv.org/abs/1903.11981

Interview w/ Harri:    • Harri Valpola: System 2 AI and Planni...  

Abstract:
Trajectory optimization using a learned model of the environment is one of the core elements of model-based reinforcement learning. This procedure often suffers from exploiting inaccuracies of the learned model. We propose to regularize trajectory optimization by means of a denoising autoencoder that is trained on the same trajectories as the model of the environment. We show that the proposed regularization leads to improved planning with both gradient-based and gradient-free optimizers. We also demonstrate that using regularized trajectory optimization leads to rapid initial learning in a set of popular motor control tasks, which suggests that the proposed approach can be a useful tool for improving sample efficiency.

Authors: Rinu Boney, Norman Di Palo, Mathias Berglund, Alexander Ilin, Juho Kannala, Antti Rasmus, Harri Valpola

Links:
YouTube:    / yannickilcher  
Twitter:   / ykilcher  
BitChute: https://www.bitchute.com/channel/yann...
Minds: https://www.minds.com/ykilcher


Watch video Regularizing Trajectory Optimization with Denoising Autoencoders (Paper Explained) online without registration, duration hours minute second in high quality. This video was added by user Yannic Kilcher 24 May 2020, don't forget to share it with your friends and acquaintances, it has been viewed on our site 5,471 once and liked it 191 people.