Poster
The Courage to Stop: Overcoming Sunk Cost Fallacy in Deep Reinforcement Learning
Jiashun Liu · Johan Obando-Ceron · Pablo Samuel Castro · Aaron Courville · Ling Pan
West Exhibition Hall B2-B3 #W-721
In this paper, we observe the sunk cost fallacy in mainstream deep RL algorithms. This fallacy manifests as agents executing each episode blindly, incur significant interaction costs, and may pollute the data distribution in the buffer, limiting their learning potential. This observation highlights an overlooked aspect within the community: the inability of existing algorithms to stop before entering into suboptimal trajectories may be a hidden factor that limits their performance. To address this issue, we propose a direct optimization approach called LEAST to quantify the current situation and control the termination of sampling. LEAST effectively mitigates the sunk cost fallacy without the need for additional networks and enhances learning efficiency across diverse scenarios. We hope that the problem presented in this paper could inspire future research aimed at optimizing sampling and learning efficiency.