A survey on policy search algorithms for learning robot controllers in a handful of trials
K Chatzilygeroudis, V Vassiliades… - IEEE Transactions …, 2019 - ieeexplore.ieee.org
Most policy search (PS) algorithms require thousands of training episodes to find an
effective policy, which is often infeasible with a physical robot. This survey article focuses on …
effective policy, which is often infeasible with a physical robot. This survey article focuses on …
Bayesian optimization using domain knowledge on the ATRIAS biped
Robotics controllers often consist of expert-designed heuristics, which can be hard to tune in
higher dimensions. Simulation can aid in optimizing these controllers if parameters learned …
higher dimensions. Simulation can aid in optimizing these controllers if parameters learned …
Robot learning with crash constraints
In the past decade, numerous machine learning algorithms have been shown to
successfully learn optimal policies to control real robotic systems. However, it is common to …
successfully learn optimal policies to control real robotic systems. However, it is common to …
Using simulation to improve sample-efficiency of Bayesian optimization for bipedal robots
Learning for control can acquire controllers for novel robotic tasks, paving the path for
autonomous agents. Such controllers can be expert-designed policies, which typically …
autonomous agents. Such controllers can be expert-designed policies, which typically …
Using parameterized black-box priors to scale up model-based policy search for robotics
K Chatzilygeroudis, JB Mouret - 2018 IEEE international …, 2018 - ieeexplore.ieee.org
The most data-efficient algorithms for reinforcement learning in robotics are model-based
policy search algorithms, which alternate between learning a dynamical model of the robot …
policy search algorithms, which alternate between learning a dynamical model of the robot …
Robust walking based on MPC with viability guarantees
Model predictive control (MPC) has shown great success for controlling complex systems,
such as legged robots. However, when closing the loop, the performance and feasibility of …
such as legged robots. However, when closing the loop, the performance and feasibility of …
Deep kernels for optimizing locomotion controllers
Sample efficiency is important when optimizing parameters of locomotion controllers, since
hardware experiments are time consuming and expensive. Bayesian Optimization, a sample …
hardware experiments are time consuming and expensive. Bayesian Optimization, a sample …
Beyond basins of attraction: Quantifying robustness of natural dynamics
S Heim, A Spröwitz - IEEE Transactions on Robotics, 2019 - ieeexplore.ieee.org
Properly designing a system to exhibit favorable natural dynamics can greatly simplify
designing or learning the control policy. However, it is still unclear what constitutes favorable …
designing or learning the control policy. However, it is still unclear what constitutes favorable …
On the design of LQR kernels for efficient controller learning
Finding optimal feedback controllers for nonlinear dynamic systems from data is hard.
Recently, Bayesian optimization (BO) has been proposed as a powerful framework for direct …
Recently, Bayesian optimization (BO) has been proposed as a powerful framework for direct …
Learning to control highly accelerated ballistic movements on muscular robots
High-speed and high-acceleration movements are inherently hard to control. Applying
learning to the control of such motions on anthropomorphic robot arms can improve the …
learning to the control of such motions on anthropomorphic robot arms can improve the …