Recent advances in reinforcement learning (RL) enable its use on increasingly complex tasks, but the lack of formal safety guarantees still limits its application in safety-critical settings. A common practical approach is to augment the RL policy with a safety filter that overrides unsafe actions to prevent failures during both training and deployment. However, safety filtering is often perceived as sacrificing performance and hindering the learning process. We show that this perceived safety–performance tradeoff is not inherent and prove, for the first time, that enforcing safety with a sufficiently permissive safety filter does not degrade asymptotic performance. We formalize RL safety with a safety-critical Markov decision process (SC-MDP), which requires categorical, rather than high-probability, avoidance of catastrophic failure states. Additionally, we define an associated filtered MDP in which all actions result in safe effects, thanks to a safety filter that is considered to be a part of the environment. Our main theorem establishes that (i) learning in the filtered MDP is safe categorically, (ii) standard RL convergence carries over to the filtered MDP, and (iii) any policy that is optimal in the filtered MDP—when executed through the same filter—achieves the same asymptotic return as the best safe policy in the SC-MDP, yielding a complete separation between safety enforcement and performance optimization. We validate the theory on Safety Gymnasium with representative tasks and constraints, observing zero violations during training and final performance matching or exceeding unfiltered baselines. Together, these results shed light on a long-standing question in safety-filtered learning and provide a simple, principled recipe for safe RL: train and deploy RL policies with the most permissive safety filter that is available.
@article{oh2025provably,title={Provably Optimal Reinforcement Learning under Safety Filtering},author={Oh, Donggeon David and Nguyen, Duy P and Hu, Haimin and Fisac, Jaime F},journal={arXiv preprint arXiv:2510.18082},year={2025}}
RSS
Safety with Agency: Human-Centered Safety Filter with Application to AI-Assisted Motorsports
Donggeon David
Oh*, Justin
Lidard*, Haimin
Hu, and
8 more authors
In Proceedings of Robotics: Science and Systems, 2025
We propose a human-centered safety filter (HCSF) for shared autonomy that significantly enhances system safety without compromising human agency. Our HCSF is built on a neural safety value function, which we first learn scalably through black-box interactions and then use at deployment to enforce a novel state-action control barrier function (Q-CBF) safety constraint. Since this Q-CBF safety filter does not require any knowledge of the system dynamics for both synthesis and runtime safety monitoring and intervention, our method applies readily to complex, black-box shared autonomy systems. Notably, our HCSF’s CBF-based interventions modify the human’s actions minimally and smoothly, avoiding the abrupt, last-moment corrections delivered by many conventional safety filters. We validate our approach in a comprehensive in-person user study using Assetto Corsa—a high-fidelity car racing simulator with black-box dynamics—to assess robustness in "driving on the edge" scenarios. We compare both trajectory data and drivers’ perceptions of our HCSF assistance against unassisted driving and a conventional safety filter. Experimental results show that 1) compared to having no assistance, our HCSF improves both safety and user satisfaction without compromising human agency or comfort, and 2) relative to a conventional safety filter, our proposed HCSF boosts human agency, comfort, and satisfaction while maintaining robustness.
@inproceedings{oh2025safety,title={Safety with Agency: Human-Centered Safety Filter with Application to AI-Assisted Motorsports},author={Oh, Donggeon David and Lidard, Justin and Hu, Haimin and Sinhmar, Himani and Lazarski, Elle and Gopinath, Deepak and Sumner, Emily S and DeCastro, Jonathan A and Rosman, Guy and Leonard, Naomi Ehrich and and Fisac, Jaime Fern{\'a}ndez},booktitle={Proceedings of Robotics: Science and Systems},year={2025}}
Short & Exciting Intro (1 min, muted)RSS 2025 Talk (4 min, narrated by myself!)
2023
CDC
Safety-critical control under multiple state and input constraints and application to fixed-wing UAV
Donggeon David
Oh*, Dongjae
Lee*, and H Jin
Kim
In 2023 62nd IEEE Conference on Decision and Control (CDC), 2023
This study presents a framework to guarantee safety for a class of second-order nonlinear systems under multiple state and input constraints. To facilitate real-world applications, a safety-critical controller must consider multiple constraints simultaneously, while being able to impose general forms of constraints designed for various tasks (e.g., obstacle avoidance). With this in mind, we first devise a zeroing control barrier function (ZCBF) using a newly proposed nominal evading maneuver. By designing the nominal evading maneuver to 1) be continuously differentiable, 2) satisfy input constraints, and 3) be capable of handling other state constraints, we deduce an ultimate invariant set, a subset of the safe set that can be rendered forward invariant with admissible control inputs. Thanks to the development of the ultimate invariant set, we then propose a safety-critical controller, which is a computationally tractable one-step model predictive controller (MPC) with guaranteed recursive feasibility. We validate the proposed framework in simulation, where a fixed-wing UAV tracks a circular trajectory while satisfying multiple safety constraints including collision avoidance, bounds on flight speed and flight path angle, and input constraints.
@inproceedings{oh2023safety,title={Safety-critical control under multiple state and input constraints and application to fixed-wing UAV},author={Oh, Donggeon David and Lee, Dongjae and Kim, H Jin},booktitle={2023 62nd IEEE Conference on Decision and Control (CDC)},pages={1748--1755},year={2023},organization={IEEE}}
ICRA
Stable Contact Guaranteeing Motion/Force Control for an Aerial Manipulator on an Arbitrarily Tilted Surface
Jeonghyun
Byun, Byeongjun
Kim, Changhyeon
Kim, and
2 more authors
In 2023 IEEE International Conference on Robotics and Automation (ICRA), 2023
This study aims to design a motion/force controller for an aerial manipulator which guarantees the tracking of time-varying motion/force trajectories as well as the stability during the transition between free and contact motions. To this end, we model the force exerted on the end-effector as the Kelvin-Voigt linear model and estimate its parameters by recursive least-squares estimator. Then, the gains of the disturbance-observer (DOB)-based motion/force controller are calculated based on the stability conditions considering both the model uncertainties in the dynamic equation and switching between the free and contact motions. To validate the proposed controller, we conducted the time-varying motion/force tracking experiments with different approach speeds and orientations of the surface. The results show that our controller enables the aerial manipulator to track the time-varying motion/force trajectories.
@inproceedings{byun2023stable,title={Stable Contact Guaranteeing Motion/Force Control for an Aerial Manipulator on an Arbitrarily Tilted Surface},author={Byun, Jeonghyun and Kim, Byeongjun and Kim, Changhyeon and Oh, Donggeon David and Kim, H Jin},booktitle={2023 IEEE International Conference on Robotics and Automation (ICRA)},pages={5345--5351},year={2023},organization={IEEE}}
2022
ICCAS
Real-Time Trajectory Generation of a Quadrotor UAV with Load Suspended from a Pulley
Donggeon David
Oh, Jeonghyun
Byun, and Dongjae
Lee
In 2022 22nd International Conference on Control, Automation and Systems (ICCAS), 2022
Quadrotor UAV carrying a payload with a cable can be a prominent setup for aerial transportation. However, it is challenging for the suspended load system to navigate through tunnel-like obstacles if the vertical gap is narrower than the length of the cable. We aim to resolve this problem by adopting the pulley mechanism which controls the cable length. This paper presents a model predictive control based real-time trajectory generation algorithm for a quadrotor with its load suspended from a pulley, or the quadrotor-pulley system. We design a cost function that can be generally applied for avoiding obstacles with rectangular surfaces. Using the cost function, an optimal control problem is formulated, which is then solved by implementing model predictive control algorithm with sequential quadratic programming solver. We employ the state-of-the-art real-time iteration scheme to promote computational efficiency of the algorithm. The presented algorithm is demonstrated with simulations, where trajectories for the quadrotor-pulley system were successfully generated. Collision-free trajectories through tunnels with various heights and lengths were generated in real-time, thus validating the robustness of the algorithm.
@inproceedings{oh2022real,title={Real-Time Trajectory Generation of a Quadrotor UAV with Load Suspended from a Pulley},author={Oh, Donggeon David and Byun, Jeonghyun and Lee, Dongjae},booktitle={2022 22nd International Conference on Control, Automation and Systems (ICCAS)},pages={1309--1314},year={2022},organization={IEEE}}