<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>H. Kim, B.-H. So, W.-S. Han, H. Lee, Natural language to SQL: Where are we today?, Proceedings
of the VLDB Endowment</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.48550/arXiv.2305.02301</article-id>
      <title-group>
        <article-title>Databases: From Data Storage Towards Partners for Information Access and Discovery</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Benjamin Hättasch</string-name>
          <email>benjamin.haettasch@dfki.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carsten Binnig</string-name>
          <email>carsten.binnig@cs.tu-darmstadt.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>German Research Center for Artificial Intelligence (DFKI)</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Technical University of Darmstadt (TU Darmstadt)</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>13</volume>
      <issue>2020</issue>
      <fpage>1</fpage>
      <lpage>7</lpage>
      <abstract>
        <p>Classically, data storage and its usage were separated: There were databases with the task of storing large amounts of data eficiently and allowing fast access. On the other hand, there were experts who built tools to access or manipulate certain parts of that data to solve very specific tasks. With the ever-growing amount of data, users need new semantic sense-making methods that reduce that overhead. Automation may help here-and indeed, in the last years, we have witnessed a strong growth of easily usable approaches for information access, especially through the wide adaptation of ChatGPT in society. However, these approaches often rely on difuse background knowledge and merely use the available structured data. As a result, they take a lot of control from the user and, nevertheless, cannot provide quality guarantees or traceable results. In this position paper, we therefore argue that simple automation, through LLMs or other means, is insuficient. Instead, we propose building systems that leverage structured data and user interaction, and automate some tasks while still leaving the users in control through carefully designed means of interaction. Based on three case studies, we analyze how treating the data system as a partner could not only improve performance, but make relevant information more accessible for all kinds of users, too. In this regard, we identify directions and principles for future research.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;databases</kwd>
        <kwd>data systems</kwd>
        <kwd>information access</kwd>
        <kwd>human automation interaction</kwd>
        <kwd>human data interaction</kwd>
        <kwd>interaction with automation</kwd>
        <kwd>human AI interaction</kwd>
        <kwd>human-centered AI (HCAI)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The need to democratize data systems Accessing, processing, and storing relevant information
is important in many fields, from research, healthcare, and public service to journalism: Researchers
need to find relevant existing work. City administrations want to learn about the needs of their citizens
from large numbers of complaints. Fiscal authorities try to uncover tax evasion and money laundering.
Journalists need facts and statistics to back their articles. Healthcare professionals interpret test results
and patient files to learn about their patients and how to help them.</p>
      <p>
        Meeting these needs at scale requires the automation of knowledge tasks. Computers, with their ability
to quickly store and process enormous amounts of data and the progress in AI (e.g., pattern detection,
automatic translation, language modeling) ofer great new opportunities for that [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Unfortunately,
those approaches are often associated with high efort and overhead, and can only be used by AI experts.
A focus must, therefore, be set on not just providing the general functionality but making it accessible
to a wide range of people. The importance is underlined by the growing demand for data scientists that
currently cannot be met by far [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ], slowing down scientific, industrial, and societal development
and progress.
      </p>
      <p>
        The need for carefully designed automation As a consequence, we witnessed a rise in approaches
that are very easy to use: End users experience tools to directly receive answers to their (knowledge)
questions (e.g., Siri, Google Assistant, ChatGPT) in their day-to-day life. It, therefore, seems only natural
that they expect something similar for their professional tasks. Hence, many tools with conversational
interfaces have emerged recently, e.g., to query data sources in natural language or create code blocks
to access, process, and visualize information. However, from our perspective, these approaches often
overshoot the target by taking away that much control from the users, leading to (unjustified) blind
trust in the results [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and a lack of eficient means to correct errors in the results or resulting behavior.
      </p>
      <p>
        While some automation is needed to solve the relevant tasks at scale, full automation is often
impractical. After all, this is not autonomous driving. Those interacting with the data should still be in
the “driver’s seat”—and hence need to be provided with useful interfaces to interact and control the
process while being relieved of tedious tasks. Thus, it is important to carefully design the interaction
and balance between control and automation to prevent, as described by Bainbridge [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], the irony
that automation might make a user’s tasks more dificult instead of supporting them. Wiberg and
Bergqvist [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] describe how automation requires shifting from direct control towards cooperation with
the computer. This can help ensure that humans identify with their work and feel responsible for the
results [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Designing this interaction to make it feel like working with a team partner seems promising,
but achieving it is not easy.
      </p>
      <p>
        The need for integrated solutions These improved data systems should focus on reducing manual
work, not necessarily the amount of interaction. Combining research on AI and UX to design systems
where AI complements interaction instead of replacing it can be a key to making systems more accessible
and useful [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. We argue that including the data by integrating the semantic functionality directly into
the data system can improve usability and quality even further. This holds, e.g., for conversational
agents, where the number of necessary interactions can be reduced when the computer has access to
data characteristics of the underlying data storage system [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Building systems that adapt to domains
and users based on interaction may also be a good alternative to resolve drawbacks of using LLMs,
such as high computation eforts, legal or privacy problems, lack of guarantees, or missing resources
for fine-tuning. Again, this requires a close integration with relevant data sources.
      </p>
      <p>
        Since the quality of this integration will influence the quality of the results and the answering speed
of the system, it will change how the user perceives the computer it is currently collaborating with
(is it more of an unskilled assistant or an expert?), afecting trust and adaptation. Furthermore, we
argue that interaction in integrated systems can be shaped to achieve better quality compared to other
automated approaches with the same amount of manual work for the users. Thus, this might help to
overcome zero-sum assumptions, as criticized by Shneiderman [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        Finally, it has long been recognized that users need to be involved in processes to avoid boredom and
distraction and thus poorer results overall [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Designing integrated data systems instead of leaving it
up to the user to combine multiple tools can help to achieve such a balanced interaction.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Case Studies</title>
      <p>In this section, we will demonstrate the advantages of integrated interactive data systems mentioned
above. We will collect requirements for these systems and sketch potential interaction patterns. To do
so, we present three case studies of tasks from the data systems community and discuss how interactive
systems could tackle them.</p>
      <p>First, we present an interactive system for ad-hoc information extraction that we built in the last few
years. This system is targeted at people without a background in data or computer science. Through
interaction, this system can outperform learned and few-shot approaches and that at only a fraction of
computing costs. Afterwards, we will describe a tool to increase the productivity of experts dealing with
large data systems and analyze why existing non-interactive approaches cannot solve that problem.
Finally, we present a vision of how data storage and access could radically change and how experts and
non-experts will benefit.</p>
      <sec id="sec-2-1">
        <title>2.1. Table Extraction from Text</title>
        <p>
          Why? Large amounts of information are only available as written text. Users like journalists or
researchers who are confronted with a text collection are often only interested in specific facts from
it, e.g., they need a table of persons mentioned together with their birthplaces or of the prevailing
weather conditions during described events. Such a structured representation allows them to leverage
the contained knowledge without having to read those texts repeatedly. However, many data discovery
techniques require technical background knowledge and are not easily accessible; there is a lack of
easy-to-use tools for information extraction, as information overload can limit the performance of
knowledge workers and not all existing tools provide suficient benefit [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>Information needs difer across users and situations, ranging from facts that are short and often
expressed in similar terms (like the date of an event) to long phrases (like weather conditions), which
can be described in various ways. Therefore, building a one-size-fits-all system that covers all possible
information needs and domains is very dificult—custom extraction pipelines are necessary.
Unfortunately, crafting domain-specific rules, training custom extractors, or prompting an LLM for each source
document is costly and requires time, expertise, and suitable data.</p>
        <p>
          How? Therefore, we built a system that allows users to extract tables interactively and even run
SQL queries over text collections in an ad-hoc manner [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Its core feature is the interaction between
the user and the computer, allowing the computer to adapt to the domain without needing explicit
instructions. The computer will create an initial version of the requested table and then present an
excerpt to request feedback from the user. The user can then confirm cells of that table or point the
computer towards the right solution by selecting the correct span from one of the source texts to fill
that cell. The computer will then update other parts of the table based on that feedback and present
another intermediate result for feedback. By carefully selecting which rows to present for feedback, the
computer can steer for what it receives feedback. On the other hand, we use human abilities to quickly
identify standout values by leaving them the choice of what exactly to feedback from a subset. This
process repeats until the user is satisfied with the extraction quality (the requirements for this can vary
greatly depending on the scenario). Our approach allows non-expert users to automatically extract and
organize relevant content from large text collections using a simple graphical interface without the
need for programming skills.
        </p>
        <p>
          What are the advantages? Our interactive approach allows quickly adapting to a new domain
without requiring extensive and costly training. The users do not have to explain what the names of
potentially very domain-dependent column titles mean, nor do they have to provide additional resources
like manually annotated training data. Our system supports aggregation operations like counting and
summing up. It thus can directly produce tables stating information that is not explicitly mentioned in
the documents and hence not discoverable by pure extraction or search approaches. Our experiments
show that the number of required interactions depends on a well-suited strategy for selecting rows for
feedback.1 Throughout this interaction, the resulting table-filling quality will be much better than for a
non-interactive (e.g., few-shot) system where the user provides the same amount of annotations in an
unguided process [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. By using user-provided samples for more relevant language understanding tasks,
that matching quality can be increased even further [15]. At the same time, the required computation
will be magnitudes lower than for running the complete large text collections through large generative
models. Thus, in summary, an integrated approach with a simple user interface makes the functionality
available to more people, can reduce costs for usage, and might lead to better quality of the results.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. “Sloppy SQL”: Supporting near-correct queries</title>
        <p>
          Why? Accessing or manipulating data in a database traditionally requires writing SQL queries. In the
last decade, natural language interfaces for databases (NLIDBs) emerged as an alternative, treating the
problem either as a single natural language to SQL translation task [16, 17, 18] or building conversational
interfaces that allow multi-turn interaction [
          <xref ref-type="bibr" rid="ref9">19, 9</xref>
          ]. However, even with the usage of large language
1A similar efect can be observed in the training method of LLM distilling [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], where a small model selects examples for
labeling, and the resulting model then outperforms a model trained on a superset.
models like ChatGPT, the translation quality stays substantially below the one of human experts
[20, 21, 22]. Hence, while these NLIDBs are very useful to make data accessible for non-experts, there
are scenarios where the quality reached by automatic translation is not suficient. It seems more
promising to concentrate on making it easier for experts to craft SQL queries.
        </p>
        <p>Real-world database schemas can consist of thousands of wide tables with hundreds of columns,
rendering it nearly impossible to keep the details in mind. Moreover, queries are often crafted by
consultants who might have a general model of the domain they are working on in mind but do not
know the specific design of the organization for which they are creating the reports. An interactive
tool that allows them to enter a sloppy version of the query and then develop the correct one together
with the computer would be very helpful. As sloppy we define syntactically correct queries that use,
e.g., synonyms of correct table or column names, are incomplete, or even assume a difering schema.
How? The interaction could look as follows: The user first inserts a version of the query based on
their mental domain knowledge. The computer then tries to map between the user input and the
database schema, replacing table and column names. At the same time, it selects additional information
for the user, such as displaying possibly relevant table snippets, characteristics and visualizations of
the table contents, or results and error messages from executing the query or parts of it. The user can
then review the proposed changes, choose between options, or manually refine parts of the query. This
process repeats until the user is convinced that their query is correct and works as intended.
What is currently missing? One could argue that there is no need for an integrated, interactive
system here, and that the problem could instead be seen as a simple translation from sloppy to correct
queries. We therefore tested how well diferent versions of OpenAI’s GPT as well as a Llama 3 model
[23] fine-tuned for this task perform when provided with an erroneous query and the real database
schema and are instructed to fix the query. For that, we used modified ( sloppy) versions of queries
from the Spider [24] and BIRD [22] benchmarks, and additionally manually crafted a subset of 100
queries from them for which we ensured that all information necessary to correct the query is present
to the system. We tried diferent prompts and diferent formats for representing the database schema,
included chain-of-thought, and even built a multi-turn version where the corrected query, together
with any error messages from execution, are presented to the LLM again. Nevertheless, throughout all
our experiments, the execution accuracy stayed below 40 % on all models.</p>
        <p>How could it be instead? Therefore, we propose to tackle the problem using an interactive system,
which ofers the following opportunities: The repeated interaction between the user and the computer
should allow it to converge to a good solution for a problem that cannot (yet) be solved by the computer
alone. From the perspective of a programmer, this interaction pattern much more resembles
pairprogramming and, therefore, a process where one works together in a team instead of the instructive
style of giving orders and hoping they lead to the correct result when prompting tools like GitHub
Copilot. Furthermore, such a system could directly incorporate specific database schemas and even
contents to (faster) produce better results. Finally, an interactive system crafted directly for this
task could display headers and parts of the contents, or results for (parts of) the queries (choosing
automatically which of them are relevant based on the interaction with the system). Working with a
computer as a team partner will likely lead to better results and increase the satisfaction of the person
working on the problem.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Self-Organizing Databases</title>
        <p>Why? Governments, small organizations, and individuals often need to handle, store, and analyze
incoming data but lack the knowledge and resources to design suitable storage systems and pipelines
for analysis. For example, the government of a small city might ask its citizens to report any problems
they notice, such as broken streetlights, trash in the park, or unfavorable trafic conditions. These
reports can happen in various formats, some containing a few lines of text or consisting solely of
an image, some with more details in an already semi-structured format. A form with required fields
can help enforce a consistent level of quality and completeness but might drastically reduce users’
willingness to report compared to a completely unstructured transmission method, such as a WhatsApp
number to message. Besides, it might not be obvious upfront which meta information needs to be
requested from the user. As a result, this incoming data will most likely be stored in a way that prevents
direct and convenient analysis and actions. Moreover, the city might already have existing structured
information, e.g., maintenance logs or the schedule of garbage collections, but it is dificult to link it
with the incoming unstructured data.</p>
        <p>How? We, therefore, envision self-organizing data systems. Automatic systems that develop suitable
schemas and even adapt the contents themselves (e.g., to normalize them) could support humans
responsible for making sense of or ensuring the correct storage of data. They would not only reduce
manual workload but also allow for continuous design adjustment and improvement.
What is needed? At the core, such a self-organizing database thus requires a component that
continuously re-evaluates its state based on recent and all previous inputs and queries. This requires
solving a wide range of challenges, which comprise i.a.: (1) Finding the right place to insert data,
which might already require schema updates for new types of data or information pieces that were
not available for previous rows of that semantic type. (2) Deriving structured representations from
the inputs and unifying them across modalities. (3) Inferring relevant columns to answer a query and
deciding whether they should be materialized or computed ad-hoc for answering that query only.
Why interaction plays a crucial role for that There are two central sources from which to draw
the requested information: The data itself on the one hand and the queries of the users working with
the data on the other. However, only concentrating on the data itself seems not suficient: existing
approaches tend to store all incoming data disorganized in data lakes and then focus on retrieving or
producing matching tables for a specific information need posed by the users [ 25, 26, 27, 28]. In contrast,
we aim to directly transform all incoming data into a semantically meaningful representation, starting
with a very rough version that only reflects the basic concepts of the relevant domain and then refining
it based on the needs of those interacting with it. Thus, although we need to preserve raw data for later
database adjustments, the main surface of interaction should be the database automatically created
from the input data. Adapting the data system requires considering the interaction with the system,
adjusting to domain-specific wording, learning which parts of the information have to be combined,
and being able to inform the people administrating the system what information might be missing.</p>
        <p>Users can manually explore databases resulting from such a system without a clear information
need in mind, as they might do with a database designed by a human data engineer. At the same time,
data engineers are supported in their work so they can concentrate on semantically relevant tasks like
adding new data sources instead of routine work like adapting the schema and rewriting queries.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. What should the interaction with future data systems look like?</title>
      <p>Findings from the case studies To summarize the main findings from the case studies, interaction
can be used to provide domain adaptation or to allow computers to choose what is relevant automatically.
By this, users can be supported, and overall, better results can be achieved. Furthermore, this might
reduce the knowledge and skills needed to use such a system, making it accessible to further (groups
of) people and allowing its application in further scenarios. By reducing manual workload, users
can concentrate on semantically relevant tasks instead of routine work, resulting in better scaling
and faster/more frequent results and updates. The interaction pattern needs to balance control and
automation to both gain good results and increase user satisfaction.</p>
      <p>This interaction can be shaped in many diferent ways: Users can enter definitions or prompts.
Alternatively, systems can automatically adapt to the user based on provided examples, feedback, or
their behavior. Furthermore, it will often be relevant to choose what information to present to the user
automatically. This may include explanations, justifications, confidences, visualizations, error messages,
and any additional information relevant for concrete information needs or interesting for exploratory
scenarios.</p>
      <p>Since the scenarios are manifold, there is no one-size-fits-all solution for interaction patterns or
interfaces for these problems; they must always be adapted to the task. However, our case studies hint
that iterative approaches might often be helpful, and that it is often better when the user feels like
working in a team with the computer instead of giving it orders.</p>
      <p>
        Together is better Interactive data systems can, in particular, directly incorporate existing
(structured) data sources for grounded results and optimized paths to them (e.g., by exploiting data
characteristics to narrow down ambiguities eficiently). Integrating these data sources and automated approaches
that react to the user’s behavior can reduce the overall amount of explicit prompts, annotation, and
additional data needed. This might allow all kinds of users to explore data freely (even without a
specific information need) or even make it possible to process substantial amounts of data in a specific
domain in the first place. Thus, we advocate not tackling interaction, semantic interpretation, and data
management separately but considering them together when building new semantic data systems. This,
however, leads to new challenges:
How should interfaces look like? Often, interface design does not play a central role in data systems
research. As part of this integration, this has to evolve. It will not be suficient to simply choose between
a GUI and a conversational interface; researchers must also carefully consider which information to
select and how to display it. They have to ensure that the users have the relevant information but do
not get overwhelmed—which includes adapting to potentially diferent target groups. The resulting
interfaces should relieve users from tedious tasks but simultaneously prevent blind trust and challenge
users to reflect the computer’s suggestions [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. They must ofer transparency and explainability, but
prevent information overload, which might lead to “pseudo-control” and “pseudo-accountability” [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
Here, a close collaboration with the HCI community and their decades of experience with designing
interaction will be key.
      </p>
      <p>How can this be evaluated? The second, probably most prominent question will be how to evaluate
these systems. The data systems community strongly focuses on measuring the accuracy of results and
the performance of their creation; this must be combined with measuring usability and user satisfaction.
Joint measures may then help assess how well a system is suited for real-world applications much better
than inspecting results from performance and usability evaluations individually. However, it will be
necessary to develop techniques (e.g., through simulation and standardized tasks and measures) for
conducting these evaluations without massively increasing the necessary eforts of the involved researchers.
We are confident that combining approaches from these diferent areas of research can lead
to exciting new results. Therefore, we consider it very important to develop answers to the points
above in order to make progress here.</p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <p>Partially funded by the German Federal Ministry of Education and Research within the “The Future of
Value Creation – Research on Production, Services and Work” program (grant 02L19C150) and by the
Hessian Ministry of Higher Education, Research, Science and the Arts.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Grammarly in order to: Grammar and spelling
check, improve writing style. Further, the authors used DeepL in order to: Text Translation. After using
these tools, the authors reviewed and edited the content as needed and take full responsibility for the
publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Coombs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hislop</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Taneva</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. Barnard,</surname>
          </string-name>
          <article-title>The strategic impacts of Intelligent Automation for knowledge and service work: An interdisciplinary review</article-title>
          ,
          <source>The Journal of Strategic Information Systems</source>
          <volume>29</volume>
          (
          <year>2020</year>
          )
          <article-title>101600</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.jsis.
          <year>2020</year>
          .
          <volume>101600</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <article-title>Is Data Scientist Still the Sexiest Job of the 21st Century?</article-title>
          ,
          <source>Harvard Business Review</source>
          (
          <year>2022</year>
          ). URL: https://hbr.org/
          <year>2022</year>
          /07/ is
          <article-title>-data-scientist-still-the-sexiest-job-of-the-21st-century.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3] Bureau of Labor Statistics, U.S. Department of Labor, Data Scientists : Occupational Outlook Handbook: : U.S. Bureau of Labor Statistics,
          <year>2023</year>
          . URL: https://www.bls.gov/ooh/math/data-scientists. htm.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>World</given-names>
            <surname>Data Science Initiative</surname>
          </string-name>
          ,
          <article-title>Why Data Science is the most in-demand skill now and how can you prepare for it?</article-title>
          ,
          <year>2023</year>
          . URL: https://www.worlddatascience.org/blogs/ why
          <article-title>-data-science-is-the-most-indemand-skill-now-and-how-can-you-prepare-for-it.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baldauf</surname>
          </string-name>
          , P. Fröhlich,
          <string-name>
            <surname>AI-Assisted Document</surname>
          </string-name>
          Tagging -
          <article-title>Exploring Adaptation Efects among Domain Experts</article-title>
          , in: P. Fröhlich,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baldauf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Palanque</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Roto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Paternó</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ju</surname>
          </string-name>
          , M. Tscheligi (Eds.),
          <source>Proceedings of the Workshop on Intervening, Teaming, Delegating</source>
          , volume
          <volume>3394</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR, Hamburg, Germany,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Bainbridge</surname>
          </string-name>
          , Ironies of automation,
          <source>Automatica</source>
          <volume>19</volume>
          (
          <year>1983</year>
          )
          <fpage>775</fpage>
          -
          <lpage>779</lpage>
          . doi:
          <volume>10</volume>
          .1016/
          <fpage>0005</fpage>
          -
          <lpage>1098</lpage>
          (
          <issue>83</issue>
          )
          <fpage>90046</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. S.</given-names>
            <surname>Bergqvist</surname>
          </string-name>
          ,
          <source>User Experience (UX) meets Artificial Intelligence (AI</source>
          )
          <article-title>- Designing Engaging User Experiences Through 'Automation of Interaction', AutomationXP22: Engaging with Automation</article-title>
          , CHI'
          <volume>22</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sadeghian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hassenzahl</surname>
          </string-name>
          ,
          <article-title>On Autonomy and Meaning in Human-Automation Interaction</article-title>
          , AutomationXP23: Intervening, Teaming, Delegating - Creating Engag- ing
          <string-name>
            <surname>Automation</surname>
            <given-names>Experiences</given-names>
          </string-name>
          , CHI '23,
          <string-name>
            <surname>April</surname>
            <given-names>23rd</given-names>
          </string-name>
          , Hamburg, Germany (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gassen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hättasch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hilprecht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Geisler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fraser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Binnig</surname>
          </string-name>
          ,
          <string-name>
            <surname>Demonstrating</surname>
            <given-names>CAT</given-names>
          </string-name>
          :
          <article-title>synthesizing data-aware conversational agents for transactional databases</article-title>
          ,
          <source>Proc. of the VLDB Endow</source>
          .
          <volume>15</volume>
          (
          <year>2022</year>
          )
          <fpage>3586</fpage>
          -
          <lpage>3589</lpage>
          . URL: https://www.vldb.org/pvldb/vol15/p3586-h%e4ttasch.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Shneiderman</surname>
          </string-name>
          ,
          <string-name>
            <surname>Human-Centered</surname>
            <given-names>AI</given-names>
          </string-name>
          , Oxford University Press, Oxford, New York,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>W. C.</given-names>
            <surname>Harris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Hancock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Arthur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Caird</surname>
          </string-name>
          , Performance, workload, and
          <article-title>fatigue changes associated with automation</article-title>
          ,
          <source>The International Journal of Aviation Psychology</source>
          <volume>5</volume>
          (
          <year>1995</year>
          )
          <fpage>169</fpage>
          -
          <lpage>185</lpage>
          . doi:
          <volume>10</volume>
          .1207/s15327108ijap0502_
          <fpage>3</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Karr-Wisniewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <article-title>When more is too much: Operationalizing technology overload and exploring its impact on knowledge worker productivity</article-title>
          ,
          <source>Computers in Human Behavior</source>
          <volume>26</volume>
          (
          <year>2010</year>
          )
          <fpage>1061</fpage>
          -
          <lpage>1072</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S0747563210000488. doi:
          <volume>10</volume>
          .1016/j.chb.
          <year>2010</year>
          .
          <volume>03</volume>
          .008, advancing Educational Research on
          <article-title>Computer-supported Collaborative Learning (CSCL) through the use of gStudy CSCL Tools</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hättasch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bodensohn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Vogel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Binnig</surname>
          </string-name>
          , Wannadb:
          <article-title>Ad-hoc SQL queries over text collections</article-title>
          , in: B.
          <string-name>
            <surname>König-Ries</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Scherzinger</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Lehner</surname>
          </string-name>
          , G. Vossen (Eds.),
          <source>Datenbanksysteme für Business</source>
          ,
          <source>Technologie und Web (BTW</source>
          <year>2023</year>
          ),
          <fpage>20</fpage>
          .
          <article-title>Fachtagung des GI-Fachbereichs „Datenbanken und Informationssysteme"</article-title>
          (DBIS),
          <volume>06</volume>
          .-
          <fpage>10</fpage>
          ,
          <year>März 2023</year>
          , Dresden, Germany, Proceedings, volume P-331
          <string-name>
            <surname>of</surname>
            <given-names>LNI</given-names>
          </string-name>
          , Gesellschaft für Informatik e.V.,
          <year>2023</year>
          , pp.
          <fpage>157</fpage>
          -
          <lpage>181</lpage>
          . doi:
          <volume>10</volume>
          .18420/BTW2023-08.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>C.-Y. Hsieh</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-L. Li</surname>
          </string-name>
          , C.
          <article-title>-</article-title>
          K. Yeh,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nakhost</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fujii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ratner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Krishna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-Y.</given-names>
            <surname>Lee</surname>
          </string-name>
          , T. Pfister,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>