Adaptive Probabilistic Trajectory Optimization via Efficient Approximate Inference release_hj4ghkf5djbwbcbd5ualtgw4v4

by Yunpeng Pan, Xinyan Yan, Evangelos Theodorou, Byron Boots

Released as a article .

2016  

Abstract

Robotic systems must be able to quickly and robustly make decisions when operating in uncertain and dynamic environments. While Reinforcement Learning (RL) can be used to compute optimal policies with little prior knowledge about the environment, it suffers from slow convergence. An alternative approach is Model Predictive Control (MPC), which optimizes policies quickly, but also requires accurate models of the system dynamics and environment. In this paper we propose a new approach, adaptive probabilistic trajectory optimization, that combines the benefits of RL and MPC. Our method uses scalable approximate inference to learn and updates probabilistic models in an online incremental fashion while also computing optimal control policies via successive local approximations. We present two variations of our algorithm based on the Sparse Spectrum Gaussian Process (SSGP) model, and we test our algorithm on three learning tasks, demonstrating the effectiveness and efficiency of our approach.
In text/plain format

Archived Files and Locations

application/pdf  2.1 MB
file_glql73xovffplf36snjm2gkx2u
arxiv.org (repository)
web.archive.org (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article
Stage   submitted
Date   2016-09-11
Version   v2
Language   en ?
arXiv  1608.06235v2
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: c13cd13d-c3ea-4023-8273-ab07c8e09186
API URL: JSON