Validating metrics for reward alignment in human-autonomy teaming

L Sanneman, JA Shah - Computers in Human Behavior, 2023 - Elsevier
Alignment of human and autonomous agent values and objectives is vital in human-
autonomy teaming settings which require collaborative action toward a common goal. In …

A Safe Preference Learning Approach for Personalization with Applications to Autonomous Vehicles

R Karagulle, N Aréchiga, A Best… - IEEE Robotics and …, 2024 - ieeexplore.ieee.org
This letter introduces a preference learning method that ensures adherence to given
specifications, with an application to autonomous vehicles. Our approach incorporates the …

Scaling Learning-based Policy Optimization for Temporal Logic Tasks by Controller Network Dropout

N Hashemi, B Hoxha, D Prokhorov… - ACM Transactions on …, 2024 - dl.acm.org
This article introduces a model-based approach for training feedback controllers for an
autonomous agent operating in a highly non-linear (albeit deterministic) environment. We …

Signal Temporal Logic-Guided Apprenticeship Learning

AG Puranic, JV Deshmukh… - 2024 IEEE/RSJ …, 2024 - ieeexplore.ieee.org
Apprenticeship learning crucially depends on effectively learning rewards, and hence
control policies from user demonstrations. Of particular difficulty is the setting where the …

A Preference Learning Approach to Develop Safe and Personalizable Autonomous Vehicles

R Karagulle, N Aréchiga, A Best, J DeCastro… - arXiv preprint arXiv …, 2023 - arxiv.org
This work introduces a preference learning method that ensures adherence to traffic rules for
autonomous vehicles. Our approach incorporates priority ordering of signal temporal logic …

Transparent Value Alignment: Foundations for Human-Centered Explainable AI in Alignment

L Sanneman - 2023 - dspace.mit.edu
Alignment of autonomous agents' values and objectives with those of humans can greatly
enhance these agents' ability to act flexibly to safely and reliably meet humans' goals across …