<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Semantic Search Engine for the Exploration of Explainability Phenomena</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Meisam Booshehri</string-name>
          <email>mbooshehri@techfak.uni-bielefeld.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christian Kullik</string-name>
          <email>ckullik@techfak.uni-bielefeld.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Philipp Cimiano</string-name>
          <email>cimiano@cit-ec.uni-bielefeld.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>SEMANTiCS'25: International Conference on Semantic Systems</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Semantic Computing Group, Faculty of Technology, Bielefeld University</institution>
          ,
          <addr-line>Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Transregional Collaborative Research Center 318 “Constructing Explainability”</institution>
          ,
          <addr-line>Paderborn and Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>3</fpage>
      <lpage>8</lpage>
      <abstract>
        <p>In this paper, we present a novel semantic search engine designed to facilitate the exploration of explainability phenomena in various contexts. The search engine is built upon an ontology specifically developed to capture explanatory moves, enabling the semantical indexing and retrieval of related concepts. This ontology integrates multiple existing annotation schemas, mapping them into a unified framework that describes diverse explanatory behaviors. We detail the development and implementation of both the search engine and the ontology, outlining their architecture and the methodology behind the integration of annotation schemas. Additionally, we present the results of a usability evaluation that assesses the efectiveness and user experience of the search engine. Our ifndings provide insights into the potential of semantic search for advancing the study and understanding of explainability phenomena.</p>
      </abstract>
      <kwd-group>
        <kwd>eXplainable AI</kwd>
        <kwd>Semantic Search</kwd>
        <kwd>Explanations</kwd>
        <kwd>Ontology design</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Recently, there has been a resurgence of interest in eXplainable Artificial Intelligence (XAI)—a subfield
of AI focused on “mak[ing] a system’s behavior intelligible and thus controllable by humans” [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Miller [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] argues that explanations are social, selective, contrastive, and causal rather than statistical,
all converging on the point that explanations are contexual. He further argues that for making truly
explainable AI systems that ofer explanations, one must conform to these key desiderata. Yet, as
surveyed by Alpsancar et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], it becomes evident that the XAI community tends to abstract away
from the social and political embeddedness of AI systems. Thommes [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] also discusses how failing to
capture the diversity of contextual factors limits the utility of explainable systems for users– e.g., due
to failing to consider the explanatory needs of users by being focused on an one-fits-all explanation.
On this basis, Rohlfing et al. [
        <xref ref-type="bibr" rid="ref1 ref5">1, 5</xref>
        ] have revisited the notion of explainability, presenting explanation as
a social practice where explainers and explainees co-construct an explanation. The explainer provides
an explanation that takes the explainee’s knowledge into account [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The explainee, on the other hand,
attends to the explanation, evaluates whether it is suitable, and provides feedback to the explainer[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Accordingly, we would like to stress that the design of XAI systems can be informed by a better
understanding of how humans explain to each other. We have therefore developed a semantic search
engine which allows to explore human-human explanatory dialogues collected and manually annotated
using annotation schemes compiled by the research projects within the scope of the Collaborative
Research Center “Constructing Explainability” at the Universities of Bielefeld and Paderborn (TRR
318). We integrate these annotation schemes, mapping them into a unified framework, which describes
diverse explanatory behaviors. In the rest of this paper, we provide an overview of the proposed system
(Section 2), present how we integrate the data from multiple projects (Section 3), and report on a
usability evaluation of the system (Section 4). Section 5 follows with concluding remarks.
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>2. System Overview</title>
      <p>Figure 1 shows the search engine interface with numbered sections; for example, number 1 is the
facetted search panel, which allows users to narrow down results by applying multiple filters (or
“facets”). Selecting a category (e.g., INF:assert) retrieves utterances annotated with it, as well as those
utterances annotated with categories mapped to INF:assert—e.g., A01:declarative and A02:assert as close
matches, and A01:additional_info and A02:statement as broader classes. However, if a user selects
categories located lower in the taxonomy, the search results are further narrowed to those more specific
categories. Figure 2 shows the software architecture of the search engine. It consists of two Docker
containers: one for the application and another for the database. The application container includes a
frontend built with Bootstrap and JavaScript, and a backend using an express.js server. Users interact
with the frontend to submit queries, which the backend processes and translates into SPARQL queries.
These SPARQL queries are sent to a Fuseki server housed in the second Docker container, where the
RDF triples are stored. The Fuseki server executes the queries and returns the resulting annotation data.
This data is then passed back through the backend to the frontend, where it is displayed to the user.</p>
      <p>Use case: To illustrate the application of this framework, we present a simple use case from the
perspective of target users. Albert is a first year PhD student in linguistics. He is interested in understanding
displays of understanding in human-human explanatory dialogues. He knows that there is this group of
datasets created by TRR 318 that has some exploration interface. He looks at the tree of categories and
makes an educated guess: completion could be related to display of understanding. He can now conveniently
browse through utterances annotated with the category completion across multiple datasets [even though not
in every dataset the category “completion” was used, but other categories known to be similar to completion].
He checks a couple of these utterances and builds the hypothesis that completion could actually be related to
display of understanding. In such a scenario the interface has helped Albert in several ways: (1) It helped
him find out that these datasets are relevant for his research, and (2) that ‘completion’ is related to
display of understanding. (3) He has looked at a couple of utterances and now understands better what
kind of analysis he could do next. (4) The facetted search in the interface has helped him in building
a hypothesis relevant to his research. An alternative would be the following: Albert has access to a
database containing those datasets. Yet, doing all of this just using a query language would have been
tedious, and would have required him to understand the structure of the data in the databases and how
to write the necessary queries.</p>
      <p>AN ONTOLOGY-BASED SEARCH ENGINE
1
2
3
4</p>
      <p>5</p>
      <p>Fuseki server
RDF Triplestore</p>
    </sec>
    <sec id="sec-3">
      <title>3. Data Integration</title>
      <p>
        What we present here is part of an ongoing larger ontology development process working towards
an ontology of explanations with applications to Social XAI (see Booshehri et al. [
        <xref ref-type="bibr" rid="ref6 ref7">7, 6</xref>
        ]). Here we
discuss how we have provided a unified taxonomy of all existing annotation schemes used in the
TRR 318 projects. First, in the INF project, we initiated with a lexicon of explanatory moves—speech
act-like categories enacted by explainers and explainees in explanatory dialogues, such as assert, confirm ,
acknowledge, etc. The list of explanatory moves was compiled from the literature in fields of tutoring
and computational linguistics, and was subsequently used to annotate various datasets (e.g., see [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]).
Next, we held ontology developer meetings with all projects within TRR 318 that are developing an
annotation schema or categorization (i.e., projects A01, A02, and A04). In these meetings, we discussed
how existing categories from the individual annotation schemes could be mapped to the INF vocabulary.
In this way, the INF vocabulary serves as an interlingua [9], to which the vocabularies of other projects
are mapped. This interlingua also serves as the default entry point for exploration in the faceted search
panel of the search engine. Each annotation scheme was treated as a separate namespace, and SKOS
vocabulary was employed to establish semantic connections between categories, as in:
INF:assert skos:broader A01:additional_info ;
      </p>
      <p>skos:closeMatch a01_rhetorical:declarative</p>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>We evaluated the proposed system using a System Usability Scale (SUS) questionnaire [10]. A total
of 28 participants from diferent backgrounds (see Figure 3c) participated in the usability study. The
SUS scores from 28 participants showed a mean of 66.07 (SD = 15.58), with a median of 70.00, a 25th
percentile of 61.88, and a 75th percentile of 75.00. The SUS scores are skewed towards the higher end of
the scale (see Figure 3a), indicating that users generally perceive the software system as acceptable or of
good quality (see Figure 3 in Bangor et al. [10] for the adjective rating scale). This reflects a positive user
experience and suggests that the system is, overall, well-received. However, a small number of lower
scores point to some usability challenges. These could be attributed to the interdisciplinary context of
current study, which can lead to varying—and at times conflicting—expectations about how the system
should look like or function.</p>
      <p>8
30
40
70
80</p>
      <p>90
(a) Distribution of SUS scores</p>
      <p>Main Area of Research Distribution</p>
      <p>Sociology
Other</p>
      <p>Computer Science
Computer Education</p>
      <p>Media Studies
7.1%</p>
      <p>21.4%
(b) Distribution of SUS scores by Research Area
Q16 3
4</p>
      <p>10
Q15
Q13
Q12
(c) DpaisnttrsibinutuiosanboilfitMyeavinaluAarteioanof Research for Partici- (d) DQi1s6t)ribution of extended Survey Questions
(Q11</p>
      <p>In addition to the original 10 items of the standard SUS, we included six additional questions in the
questionnaire (Q11-Q16; see Table 2), focusing on three specific aspects: the system interface, the
function of the ontology as an interlingua, and the use of the search engine. Regarding the interface of
the search engine (Figure 3d; Q11-Q12), most participants found the organization and display of search
results clear, and about half considered browsing via the taxonomy intuitive. On the ontology’s role as
an interlingua (Figure 3d; Q13-Q14), responses were mixed: roughly one-third found the taxonomy
helpful for understanding cross-project relationships, while others were either unsure or disagreed.
Still, a majority agreed it facilitates searching for data or key phenomena across TRR 318 projects. As
for the system’s potential to support research activities (Figure 3d; Q15-Q16), over 30% believed it
could help form new hypotheses and improve understanding of explanatory interactions at micro-level,
though uncertainty remained high.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>We have developed a semantic search engine for exploring the behavior of interlocutors in
humanhuman explanations, with the first experimental results showing an overall positive user experience.
We hope this framework can support research activities in the realm of XAI by providing insights into
the contextual factors of explanations. As future work, we plan to add more features (e.g., search for
sequences of explanatory moves) to the search engine and enrich it with additional annotated data.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation):
TRR 318/1 2021 – 438445824, Project INF (“Toward a framework for assessing explanation quality”). We
further thank Basil Ell for brainstorming ideas about framing the use-case scenario.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration of Generative AI Use</title>
      <p>During the preparation of this work, GPT-4.5 ChatGPT was used in order to: Paraphrase and reword.
The authors reviewed and edited the content as needed and take full responsibility for the publication’s
content.
[9] E. Hovy, S. Nirenburg, Approximating an interlingua in a principled way, in: Proceedings of the
Workshop on Speech and Natural Language, HLT ’91, Association for Computational Linguistics,
USA, 1992, p. 261–266. URL: https://doi.org/10.3115/1075527.1075588.
[10] A. Bangor, P. Kortum, J. Miller, Determining what individual sus scores mean: adding an adjective
rating scale, J. Usability Studies 4 (2009) 114–123.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Rohlfing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Scharlau</surname>
          </string-name>
          , et al.,
          <article-title>Explanation as a social practice: Toward a conceptual framework for the social design of ai systems</article-title>
          ,
          <source>IEEE Transactions on Cognitive and Developmental Systems</source>
          <volume>13</volume>
          (
          <year>2021</year>
          )
          <fpage>717</fpage>
          -
          <lpage>728</lpage>
          . doi:
          <volume>10</volume>
          .1109/TCDS.
          <year>2020</year>
          .
          <volume>3044366</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>267</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.artint.
          <year>2018</year>
          .
          <volume>07</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Alpsancar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. M.</given-names>
            <surname>Buhl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Matzner</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Scharlau</surname>
          </string-name>
          ,
          <article-title>Explanation needs and ethical demands: unpacking the instrumental value of XAI</article-title>
          ,
          <source>AI and Ethics</source>
          <volume>5</volume>
          (
          <year>2025</year>
          )
          <fpage>3015</fpage>
          -
          <lpage>3033</lpage>
          . URL: https://link.springer.
          <source>com/ 10.1007/s43681-024-00622-3.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Thommes</surname>
          </string-name>
          ,
          <article-title>Evaluation principles</article-title>
          ,
          <source>in: [5]</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Rohlfing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Främling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Thommes</surname>
          </string-name>
          , S. Alpsancar (Eds.),
          <string-name>
            <surname>Social Explainable</surname>
            <given-names>AI</given-names>
          </string-name>
          ,
          <source>Communications of NII Shonan Meeting</source>
          , Springer, Singapore,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Booshehri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Buschmeier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <article-title>A BFO-based ontological analysis of entities in Social XAI</article-title>
          ,
          <source>in: Proc. of the 15th Int. Conf. on Formal Ontology in Information Systems</source>
          , IOS, Catania, Italy,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Booshehri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Buschmeier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <article-title>Towards a BFO-based ontology of understanding in explanatory interactions</article-title>
          ,
          <source>in: Proc. of the 4th Int. Workshop on Data Meets Applied Ontologies in Explainable AI</source>
          (
          <article-title>DAO-XAI), CEUR-EU</article-title>
          , Santiago de Compostela, Spain,
          <year>2024</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3833</volume>
          /paper3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Booshehri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Buschmeier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <article-title>A model of factors contributing to the success of dialogical explanations</article-title>
          , in: H.
          <string-name>
            <surname>Hung</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Oertel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Soleymani</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chaspari</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Dibeklioglu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Shukla</surname>
            ,
            <given-names>K. P.</given-names>
          </string-name>
          Truong (Eds.),
          <source>Proceedings of the 26th International Conference on Multimodal Interaction (ICMI)</source>
          , ACM,
          <year>2024</year>
          , pp.
          <fpage>373</fpage>
          -
          <lpage>381</lpage>
          . URL: https://doi.org/10.1145/3678957.3685744.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>