Poster
An Optimistic Algorithm for online CMDPS with Anytime Adversarial Constraints
Jiahui Zhu · Kihyun Yu · Dabeen Lee · Xin Liu · Honghao Wei
East Exhibition Hall A-B #E-1903
Modern AI systems—like those used in self-driving cars, robots, and cybersecurity—must learn to make good decisions while following safety rules. But in the real world, these safety rules can be unpredictable: they may change over time, be unclear, or even be set up in a way that tries to trick the system. This kind of hostile or adversarial behavior makes safe learning especially difficult.In this paper, we present a new learning algorithm that helps AI systems stay safe even when the safety constraints are changing in both adversarial and stochastic ways. Our method does not assume that the system already knows what is safe, and it does not require ideal conditions to work. Instead, it learns and adapts on the challenging and uncertain environments.We show that this approach can learn effectively while also minimizing safety violations over time. This work provides stronger and more realistic guarantees for using AI safely in high-risk settings.