Safe Reinforcement Learning via Observation Shielding
dc.contributor.author | Mccalmon, Joe | |
dc.contributor.author | Liu, Tongtong | |
dc.contributor.author | Goldsmith, Reid | |
dc.contributor.author | Cyhaniuk, Andrew | |
dc.contributor.author | Halabi, Talal | |
dc.contributor.author | Alqahtani, Sarra | |
dc.date.accessioned | 2022-12-27T19:22:57Z | |
dc.date.available | 2022-12-27T19:22:57Z | |
dc.date.issued | 2023-01-03 | |
dc.description.abstract | Reinforcement Learning (RL) algorithms have shown success in scaling up to large problems. However, deploying those algorithms in real-world applications remains challenging due to their vulnerability to adversarial perturbations. Existing RL robustness methods against adversarial attacks are weak to large perturbations - a scenario that cannot be ruled out for RL adversarial threats, as is the case for deep neural networks in classification tasks. This paper proposes a method called observation-shielding RL (OSRL) to increase the robustness of RL against large perturbations using predictive models and threat detection. Instead of changing the RL algorithms with robustness regularization or retrain them with adversarial perturbations, we depart considerably from previous approaches and develop an add-on safety feature for existing RL algorithms during runtime. OSRL builds on the idea of model predictive shielding, where an observation predictive model is used to override the perturbed observations as needed to ensure safety. Extensive experiments on various MuJoCo environments (Ant, Hooper) and the classical pendulum environment demonstrate that our proposed OSRL is safer and more efficient than state-of-the-art robustness methods under large perturbations. | |
dc.format.extent | 10 | |
dc.identifier.doi | 10.24251/HICSS.2023.799 | |
dc.identifier.isbn | 978-0-9981331-6-4 | |
dc.identifier.other | 63533de5-8b7f-47c4-ac6c-c2d3d4dde1ab | |
dc.identifier.uri | https://hdl.handle.net/10125/103433 | |
dc.language.iso | eng | |
dc.relation.ispartof | Proceedings of the 56th Hawaii International Conference on System Sciences | |
dc.rights | Attribution-NonCommercial-NoDerivatives 4.0 International | |
dc.rights.uri | https://creativecommons.org/licenses/by-nc-nd/4.0/ | |
dc.subject | Cyber Operations, Defense, and Forensics | |
dc.subject | adversarial examples | |
dc.subject | reinforcement learning | |
dc.subject | robustness | |
dc.subject | safety | |
dc.subject | shielding | |
dc.title | Safe Reinforcement Learning via Observation Shielding | |
dc.type.dcmi | text | |
prism.startingpage | 6603 |
Files
Original bundle
1 - 1 of 1