Safe rlhf: Safe reinforcement learning from human feedback

J Dai, X Pan, R Sun, J Ji, X Xu, M Liu, Y Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
With the development of large language models (LLMs), striking a balance between the
performance and safety of AI systems has never been more critical. However, the inherent …

Is sora a world simulator? a comprehensive survey on general world models and beyond

Z Zhu, X Wang, W Zhao, C Min, N Deng, M Dou… - arXiv preprint arXiv …, 2024 - arxiv.org
General world models represent a crucial pathway toward achieving Artificial General
Intelligence (AGI), serving as the cornerstone for various applications ranging from virtual …

World models for autonomous driving: An initial survey

Y Guan, H Liao, Z Li, J Hu, R Yuan, Y Li… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
In the rapidly evolving landscape of autonomous driving, the capability to accurately predict
future events and assess their implications is paramount for both safety and efficiency …

Robust Training of Federated Models with Extremely Label Deficiency

Y Zhang, Z Yang, X Tian, N Wang, T Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Federated semi-supervised learning (FSSL) has emerged as a powerful paradigm for
collaboratively training machine learning models using distributed data with label deficiency …

Feasibility Consistent Representation Learning for Safe Reinforcement Learning

Z Cen, Y Yao, Z Liu, D Zhao - arXiv preprint arXiv:2405.11718, 2024 - arxiv.org
In the field of safe reinforcement learning (RL), finding a balance between satisfying safety
constraints and optimizing reward performance presents a significant challenge. A key …

OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning

Y Yao, Z Cen, W Ding, H Lin, S Liu, T Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Offline safe reinforcement learning (RL) aims to train a policy that satisfies constraints using
a pre-collected dataset. Most current methods struggle with the mismatch between imperfect …

Dynamic Model Predictive Shielding for Provably Safe Reinforcement Learning

A Banerjee, K Rahmani, J Biswas, I Dillig - arXiv preprint arXiv …, 2024 - arxiv.org
Among approaches for provably safe reinforcement learning, Model Predictive Shielding
(MPS) has proven effective at complex tasks in continuous, high-dimensional state spaces …

FOSP: Fine-tuning Offline Safe Policy through World Models

C Cao, Y Xin, S Wu, L He, Z Yan, J Tan… - arXiv preprint arXiv …, 2024 - arxiv.org
Model-based Reinforcement Learning (RL) has shown its high training efficiency and
capability of handling high-dimensional tasks. Regarding safety issues, safe model-based …

Leveraging Approximate Model-based Shielding for Probabilistic Safety Guarantees in Continuous Environments

AW Goodall, F Belardinelli - arXiv preprint arXiv:2402.00816, 2024 - arxiv.org
Shielding is a popular technique for achieving safe reinforcement learning (RL). However,
classical shielding approaches come with quite restrictive assumptions making them difficult …

[PDF][PDF] Evaluation of DreamerV3 for defective plant search in a simulated crop

T Bonte - 1924 - edepot.wur.nl
The use of autonomous robots is increasing in agriculture; however, their development is
still slow as each task usually needs a specific hard-coded behaviour. Hard coding of all …