Leveraging user preference in the design and evaluation of lower-limb exoskeletons and prostheses
The field of wearable robotics has seen major advances in recent years, largely owing to an
intense focus on optimizing device behavior to accomplish a narrow set of objectives. This …
intense focus on optimizing device behavior to accomplish a narrow set of objectives. This …
The role of user preference in the customized control of robotic exoskeletons
User preference is a promising objective for the control of robotic exoskeletons because it
may capture the multifactorial nature of exoskeleton use. However, to use it, we must first …
may capture the multifactorial nature of exoskeleton use. However, to use it, we must first …
User preference optimization for control of ankle exoskeletons using sample efficient active learning
One challenge to achieving widespread success of augmentative exoskeletons is accurately
adjusting the controller to provide cooperative assistance with their wearer. Often, the …
adjusting the controller to provide cooperative assistance with their wearer. Often, the …
Learning multimodal rewards from rankings
Learning from human feedback has shown to be a useful approach in acquiring robot
reward functions. However, expert feedback is often assumed to be drawn from an …
reward functions. However, expert feedback is often assumed to be drawn from an …
Learning reward functions from diverse sources of human feedback: Optimally integrating demonstrations and preferences
Reward functions are a common way to specify the objective of a robot. As designing reward
functions can be extremely challenging, a more promising approach is to directly learn …
functions can be extremely challenging, a more promising approach is to directly learn …
Safety-aware preference-based learning for safety-critical control
Bringing dynamic robots into the wild requires a tenuous balance between performance and
safety. Yet controllers designed to provide robust safety guarantees often result in …
safety. Yet controllers designed to provide robust safety guarantees often result in …
Towards modeling and influencing the dynamics of human learning
Humans have internal models of robots (like their physical capabilities), the world (like what
will happen next), and their tasks (like a preferred goal). However, human internal models …
will happen next), and their tasks (like a preferred goal). However, human internal models …
Aprel: A library for active preference-based reward learning algorithms
Reward learning is a fundamental problem in human-robot interaction to have robots that
operate in alignment with what their human user wants. Many preference-based learning …
operate in alignment with what their human user wants. Many preference-based learning …
Learning reward functions from scale feedback
Today's robots are increasingly interacting with people and need to efficiently learn
inexperienced user's preferences. A common framework is to iteratively query the user about …
inexperienced user's preferences. A common framework is to iteratively query the user about …
Active reward learning from online preferences
Robot policies need to adapt to human preferences and/or new environments. Human
experts may have the domain knowledge required to help robots achieve this adaptation …
experts may have the domain knowledge required to help robots achieve this adaptation …