Leveraging user preference in the design and evaluation of lower-limb exoskeletons and prostheses

KA Ingraham, M Tucker, AD Ames, EJ Rouse… - Current Opinion in …, 2023 - Elsevier
The field of wearable robotics has seen major advances in recent years, largely owing to an
intense focus on optimizing device behavior to accomplish a narrow set of objectives. This …

The role of user preference in the customized control of robotic exoskeletons

KA Ingraham, CD Remy, EJ Rouse - Science robotics, 2022 - science.org
User preference is a promising objective for the control of robotic exoskeletons because it
may capture the multifactorial nature of exoskeleton use. However, to use it, we must first …

User preference optimization for control of ankle exoskeletons using sample efficient active learning

UH Lee, VS Shetty, PW Franks, J Tan… - Science Robotics, 2023 - science.org
One challenge to achieving widespread success of augmentative exoskeletons is accurately
adjusting the controller to provide cooperative assistance with their wearer. Often, the …

Learning multimodal rewards from rankings

V Myers, E Biyik, N Anari… - Conference on robot …, 2022 - proceedings.mlr.press
Learning from human feedback has shown to be a useful approach in acquiring robot
reward functions. However, expert feedback is often assumed to be drawn from an …

Learning reward functions from diverse sources of human feedback: Optimally integrating demonstrations and preferences

E Bıyık, DP Losey, M Palan… - … Journal of Robotics …, 2022 - journals.sagepub.com
Reward functions are a common way to specify the objective of a robot. As designing reward
functions can be extremely challenging, a more promising approach is to directly learn …

Safety-aware preference-based learning for safety-critical control

R Cosner, M Tucker, A Taylor, K Li… - … for dynamics and …, 2022 - proceedings.mlr.press
Bringing dynamic robots into the wild requires a tenuous balance between performance and
safety. Yet controllers designed to provide robust safety guarantees often result in …

Towards modeling and influencing the dynamics of human learning

R Tian, M Tomizuka, AD Dragan, A Bajcsy - Proceedings of the 2023 …, 2023 - dl.acm.org
Humans have internal models of robots (like their physical capabilities), the world (like what
will happen next), and their tasks (like a preferred goal). However, human internal models …

Aprel: A library for active preference-based reward learning algorithms

E Bıyık, A Talati, D Sadigh - 2022 17th ACM/IEEE International …, 2022 - ieeexplore.ieee.org
Reward learning is a fundamental problem in human-robot interaction to have robots that
operate in alignment with what their human user wants. Many preference-based learning …

Learning reward functions from scale feedback

N Wilde, E Bıyık, D Sadigh, SL Smith - arXiv preprint arXiv:2110.00284, 2021 - arxiv.org
Today's robots are increasingly interacting with people and need to efficiently learn
inexperienced user's preferences. A common framework is to iteratively query the user about …

Active reward learning from online preferences

V Myers, E Bıyık, D Sadigh - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
Robot policies need to adapt to human preferences and/or new environments. Human
experts may have the domain knowledge required to help robots achieve this adaptation …