<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Interplay Between Language Generation and Reasoning: Information Seeking Games</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rafaella Bernardi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Rafaella Bernardi is Associate Professor at CIMeC (Center for Mind/Brain Science) and DISI (Department of Information Engineering and Computer Science), University of Trento. Through her career, she worked both with symbolic and connectionist AI approaches. She studied at the Universities of Utrecht and Amsterdam specialising in Logic and Language, in 1999 she joined the international PhD Programme at the University of Utrecht and wrote a dissertation on categorial type logic (defended in</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Trento</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <fpage>6</fpage>
      <lpage>7</lpage>
      <abstract>
        <p>Large Language Models, and ChatGPT in particular, have recently grabbed the attention of the community and the media. Having reached high language proficiency, attention has been shifting toward its reasoning capabilities. It has been shown that ChatGPT can carry out some simple deductive reasoning steps when provided with a series of facts out of which it is tasked to draw some inferences. In this talk, I am going to argue for the need of models whose language generation is driven by an implicit reasoning process. To support my claim, I will present our evaluation of ChatGPT on the 20-Questions game, traditionally used within the Cognitive Science community to inspect the information seeking-strategy's development. This task requires a series of interconnected skills: asking informative questions, stepwise updating the hypothesis space by computing some simple deductive reasoning steps, and stopping asking questions when enough information has been collected. Thus, it is a perfect testbed to monitor the language and reasoning interplay in LLMs, shed lights on their strength and their weakness, and lay the ground for models that think while speaking.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Abstract</title>
    </sec>
    <sec id="sec-2">
      <title>2. Short Biography</title>
      <p>Free University of Bozen-Bolzano (2002-20011), she worked on Natural Language Interfaces
to Structured Data. In 2011, she started working on Distributional Semantics investigating its
compositional properties and its integration with Computer Vision models. Since then she has
mostly worked on Multimodal Models in interactive settings (e.g visual dialogues). She has
recently been the EU representative within the ACL Sponsorship Board, and she is member of
the ELLIS Trento unit.</p>
      <p>Personal Website. http://disi.unitn.it/~bernardi/</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>