Poster
Improving Rationality in the Reasoning Process of Language Models through Self-playing Game
Pinzheng Wang · Juntao Li · Zecheng Tang · Haijia Gui · Min zhang
East Exhibition Hall A-B #E-2512
Large language models (LLMs) can solve complex reasoning problems, but they often struggle to recognize when their own answers are wrong—even when they do know better. This limits their reliability in critical tasks.We propose a method called the Critic Discernment Game (CDG) to improve their self-evaluation skills. In this game, three roles interact: the Prover, who answers a question step-by-step; the Helpful Critic, who points out flaws in wrong answers without giving away the fix; and the Misleading Critic, who tries to trick the Prover into changing a correct answer by suggesting a fake error.The Prover must learn to accept helpful feedback while ignoring misleading suggestions—strengthening its ability to judge reasoning quality independently. Inspired by training strategies like those used in AlphaGo, this game-like setup encourages deeper reflection.Our results show that CDG helps models better detect their own reasoning mistakes, a step toward building AI systems that are more trustworthy, cautious, and robust in real-world decision-making.