Poster
A Classification View on Meta Learning Bandits
Mirco Mutti · Jeongyeol Kwon · Shie Mannor · Aviv Tamar
West Exhibition Hall B2-B3 #W-908
In a healthcare application we may perform various tests to asses a patient condition and then decide on the best treatment to give. When human design the decision strategies, they aim for the assessment to be fast, since the patient's health is at stake, and easy to interpret for a physician overseeing the process. Instead, when the problem is tackled with AI, the resulting strategy is often opaque, hard to interpret for the end user. In this paper, we provide a template to compute decision strategies that are efficient and easy to interpret, which may be employed in the described healthcare scenario or other applications in which interpretability is important. Similar to the human approach, the strategy prescribes a sequence of simple tests to gather sufficient information, then to take optimal decisions with the given information. We analyize the proposed template both theoretically, through a formal study of its efficiency, and empirically, through numerical validation in synthetic domains. We believe our work is a first step in the direction of improving interpretability of decision strategies obtained with AI.