ChatGPT's Information Seeking Strategy: Insights From the 20-Questions Game

Leonardo Bertolazzi, Davide Mazzaccara, Filippo Merlo, Raffaella Bernardi

Paper

In Sessions:

INLG Oral Session 3: Leveraging Large Language Models for NLG: (Thursday, 10:30 CEST, Sun II , Watch on Zoom , Chat on Discord )


Abstract: Large Language Models, and ChatGPT in particular, have recently grabbed the attention of the community and the media. Having reached high language proficiency, attention has been shifting toward its reasoning capabilities. In this paper, our main aim is to evaluate ChatGPT's question generation in a task where language production should be driven by an implicit reasoning process. To this end, we employ the 20-Questions game, traditionally used within the Cognitive Science community to inspect the information seeking-strategy's development. This task requires a series of interconnected skills: asking informative questions, stepwise updating the hypothesis space, and stopping asking questions when enough information has been collected. We build hierarchical hypothesis spaces, exploiting feature norms collected from humans vs. ChatGPT itself, and we inspect the efficiency and informativeness of ChatGPT's strategy. Our results show that ChatGPT's performance gets closer to an optimal agent only when prompted to explicitly list the updated space stepwise.