Poster
Universal Approximation Theorem of Deep Q-Networks
Qian Qi
West Exhibition Hall B2-B3 #W-1009
Imagine teaching a computer to make smart decisions in situations that change smoothly over time, like guiding a self-driving car through flowing traffic or managing a complex power grid. Many current AI learning methods work well for situations with distinct steps, like a board game. Our research explores how a popular AI technique, called Deep Q-Networks (DQNs), can learn in these more fluid, continuous environments.We show two key things. First, these DQNs are theoretically powerful enough to learn the "best possible moves" (or, more technically, figure out the optimal decision-making strategy) with high accuracy, no matter how complicated the continuous task is. Second, we demonstrate that the common ways these DQNs are trained will, under the right conditions, actually lead them to correctly learn these best moves.To achieve this, we've developed a new mathematical foundation that connects these AI learning methods with established theories from control engineering. This work helps build a more solid understanding of how DQNs behave and learn in real-world scenarios that don't just jump from one step to the next, paving the way for more reliable and effective AI in dynamic systems.