(preprint) Compositional Shielding and Reinforcement Learning for Multi-Agent Systems
Asger Horn Brorholt, Kim Guldstrand Larsen, and Christian Schilling
To appear at Autonomous Agents and Multiagent Systems, AAMAS 2025
I am a PhD student at Department of Computer Science, Aalborg University. My supervisors are
Kim Guldstrand Larsen and Chistian Schilling.
I focus my research on shielded AI, where a shield compontent filters the actions of an
AI-agent, such that the system as a whole is guaranteed to be safe, even when the AI-agent alone is not.
My pronouns are they/them.
(I enjoy attending game jams. Check out the games I've made at itch.io.)
Asger Horn Brorholt, Kim Guldstrand Larsen, and Christian Schilling
To appear at Autonomous Agents and Multiagent Systems, AAMAS 2025
Asger Horn Brorholt, Andreas Holck Høeg-Petersen, Kim Guldstrand Larsen, and Christian Schilling
Verification and Learning for Assured Autonomy, ISoLA & AISoLA 2024
We consider the problem of synthesizing safety strategies for control systems, also known as shields. Since the state space is infinite, shields are typically computed over a finite-state abstraction, with the most common abstraction being a rectangular grid. However, for many systems, such a grid does not align well with the safety property or the system dynamics. That is why a coarse grid is rarely sufficient, but a fine grid is typically computationally infeasible to obtain. In this paper, we show that appropriate state-space transformations can still allow to use a coarse grid at almost no computational overhead. We demonstrate in three case studies that our transformation-based synthesis outperforms a standard synthesis by several orders of magnitude. In the first two case studies, we use domain knowledge to select a suitable transformation. In the third case study, we instead report on results in engineering a transformation without domain knowledge.
Asger Horn Brorholt, Peter Gjøl Jensen, Kim Guldstrand Larsen, Florian Lorber, Christian Schilling
Bridging the Gap Between AI and Reality - 1st International Conference, AISoLA 2023
Safe and optimal controller synthesis for switched-controlled hybrid systems, which combine differential equations and discrete changes of the system’s state, is known to be intricately hard. Reinforcement learning has been leveraged to construct near-optimal controllers, but their behavior is not guaranteed to be safe, even when it is encouraged by reward engineering. One way of imposing safety to a learned controller is to use a shield, which is correct by design. However, obtaining a shield for non-linear and hybrid environments is itself intractable. In this paper, we propose the construction of a shield using the so-called barbaric method, where an approximate finite representation of an underlying partition-based two-player safety game is extracted via systematically picked samples of the true transition function. While hard safety guarantees are out of reach, we experimentally demonstrate strong statistical safety guarantees with a prototype implementation and Uppaal Stratego. Furthermore, we study the impact of the synthesized shield when applied as either a pre-shield (applied before learning a controller) or a post-shield (only applied after learning a controller). We experimentally demonstrate superiority of the pre-shielding approach. We apply our technique on a range of case studies, including two industrial examples, and further study post-optimization of the post-shielding approach.