Poster
Concurrent Reinforcement Learning with Aggregated States via Randomized Least Squares Value Iteration
Yan Chen · Jerry Bai · Yiteng Zhang · Maria Dimakopoulou · Shi Dong · Qi Sun · Zhengyuan Zhou
West Exhibition Hall B2-B3 #W-817
Reinforcement learning (RL) trains computer programs (agents) to make optimal decisions through trial-and-error interactions with complex environments. One significant challenge is helping multiple agents explore efficiently when working together, as current methods mainly address single-agent scenarios. In our study, we developed a new approach that helps groups of agents collectively learn faster and more efficiently. We adapted a method called concurrent randomized least-squares value iteration (RLSVI) with aggregated states, allowing multiple agents to explore concurrently by sharing and simplifying their knowledge of the environment. Our theoretical results demonstrate that each agent learns faster as more agents participate, significantly speeding up overall learning. Additionally, our method greatly reduces the memory requirements compared to existing work, making it practical for systems with limited resources. Compared to previous research, our approach achieves better performance through simpler computations and innovative theoretical insights. Through experiments, we confirmed that our theory translates into practical performance improvements. This work advances our understanding of multi-agent learning, providing a scalable solution for real-world applications.