<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Exploring the Complexity of AI Applications in the Public Sector: The Interplay of Visibility, Autonomy, and Self- Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zong-Xian Huang</string-name>
          <email>zhuang7@albany.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>J. Ramon Gil-Garcia</string-name>
          <email>jgil-garcia@albany.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mila Gascó-Hernández</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Proceedings EGOV-CeDEM-ePart conference</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidad de las Americas Puebla, Ex-Hacienda Santa Catarina Mártir S/N Ex-Hacienda Santa Catarina Martir Ex-Hacienda Santa Catarina Mártir</institution>
          ,
          <addr-line>72810 San Andrés Cholula, Pue.</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University at Albany, State University of New York</institution>
          ,
          <addr-line>1400 Washington Ave, Albany, NY 12222</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Artificial intelligence (AI) has been deployed in many government contexts and with very different results in countries around the world. There seems to be a distinct transformational power when compared with previous technologies. However, it is not clear how different characteristics of AI systems affect their purpose and outputs. Therefore, by understanding some of the unique characteristics of self-learning systems in the context of a proposed typology consisting of two dimensions-visibility and autonomy-this study explores the interplay of visibility, autonomy, and self-learning in government AI systems. Based on the analysis of four distinct AI cases across diverse U.S. federal agencies, this ongoing research paper aims to uncover some of the opportunities and challenges posed by AI and specifically self-learning as one of its main features. Our preliminary results underscore the necessity of contextual analysis in deploying AI systems, thereby contributing to previous research on different characteristics and types of AI.</p>
      </abstract>
      <kwd-group>
        <kwd>AI typology</kwd>
        <kwd>AI visibility</kwd>
        <kwd>AI autonomy</kwd>
        <kwd>self-learning</kwd>
        <kwd>self-improvement 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The proliferation of artificial intelligence (AI) and machine learning algorithms in
government has significantly transformed the practices of public administration. Some
scholars have begun using the terms ‘algorithmic bureaucracy’ or ‘algorithmic governance’
to describe this new phase of public administration, referring to a new era where AI and
advanced computational algorithms play an integral role in public governance [22, 37]. This
approach of using machine learning techniques in the public sector can not only automate
relatively simple tasks but also augment complex decision-making [36]. Along with
potential benefits for government and society, however, concerns are raised as well, such</p>
      <p>0000-0002-3751-6583 (Z.-X. Huang); 0000-0002-1033-4974 (J. R. Gil-Garcia); 0000-0002-6308-8519 (M.
Gascó-Hernández)</p>
      <p>
        © 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
as issues surrounding privacy intrusions [
        <xref ref-type="bibr" rid="ref16 ref5">5, 16</xref>
        ], transparency and accountability [
        <xref ref-type="bibr" rid="ref15 ref6">6, 15</xref>
        ],
and inequality and discrimination [
        <xref ref-type="bibr" rid="ref16">16, 20, 34</xref>
        ].
      </p>
      <p>
        Depending on the complex contexts in which AI is deployed, researchers have devoted
significant effort to the development of multiple taxonomies to navigate the intricacies of AI
in the public sector. Some studies have sought to describe and compare different AI systems
based on a single dimension, such as technology, application, function, or level of
bureaucratic discretion [
        <xref ref-type="bibr" rid="ref1 ref3">1, 3, 30, 38</xref>
        ]. Others, although relatively rare, have constructed
more complex taxonomies based on multiple dimensions, reflecting intersections of
different theoretical attributes [22, 35]. These existing studies constitute the foundation for
understanding the various core characteristics of AI when applied in the public sector, as
well as underscoring the challenges and risks associated with different AI applications.
      </p>
      <p>
        Nevertheless, we argue that existing typologies do not specifically address one crucial
characteristic of AI － the self-learning attributes. Self-learning algorithms, systems that are
capable of adjusting parameters and weights in machine learning models, have gradually
garnered attention from scholars for their potential to produce not only positive outcomes
but also risky policy solutions to public problems [
        <xref ref-type="bibr" rid="ref15">15, 28, 38</xref>
        ]. As self-learning algorithms
increasingly become the foundation for AI innovations, incorporating this specific
characteristic into current typologies is becoming crucial. Specifically, by considering the
feature of self-learning algorithms, researchers and practitioners can move beyond static
descriptions to compare the effects and risks of different AI applications in a more
sophisticated manner.
      </p>
      <p>Given that current studies have developed several typologies focusing on varying
degrees of discretion and transparency in AI applications, this study aims to incorporate the
self-learning characteristic of AI into existing typologies. By explicitly considering
selflearning attributes in current typologies, we could not only refine our understanding of AI
typologies but also provide nuanced guidance for public managers in tailoring their AI
deployment strategies to fit specific contexts. Accordingly, the research questions in this
paper are (1) how do visibility and autonomy interplay in a typology that represents
different types of AI, and (2) when taking self-learning characteristics into account, what
challenges and opportunities are associated with the deployment of different types of AI
systems?</p>
      <p>From a practical perspective, the interplay among these three dimensions—visibility,
autonomy, and self-learning attributes—can benefit the discussion and understanding of AI
applications in the public sector in different ways. First, a clear taxonomy of AI, supported
by real-world cases, can enable meaningful comparisons of the effects and risks of AI, thus
overcoming the obstacles of overly abstract and general arguments currently available.
Second, with a more detailed and nuanced delineation of different AI systems, this study
offers guidance for public officials in tailoring their AI deployment strategies according to
specific contexts, thereby mitigating potential risks and maximizing the benefits of AI
applications in public administration. In other words, the formation of a more
comprehensive typology can facilitate better-informed decisions and more effective
governance.</p>
      <p>The remainder of this ongoing research study is structured as follows. In section two,
drawing on prior literature related to AI, we develop and define a proposed typology that
categorizes different AI systems based on their visibility and autonomy. Section three
outlines self-learning as one of the most important characteristics of AI systems. Section
four introduces the method use in this study, including the selection of cases. In section five,
we discuss preliminary findings based on the cases including implications for theory and
concrete policy recommendations for public managers when deploying AI systems. Finally,
we provide some final comments and briefly describe the next steps.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Visibility and Autonomy of AI</title>
      <p>Visibility and autonomy form two core attributes of AI systems. Existing studies have used
similar dimensions, such as transparency and alignment of operations, as the foundation for
a conceptual framework to evaluate the impacts of algorithmic systems [22, 35]. For
instance, with this two-dimensional framework, Katzenbach &amp; Ulbricht differentiated four
types of algorithmic governance systems based on low/high degree of transparency and
low/high degree of automation, namely, autonomy-friendly systems, trust-based systems,
licensed systems, and out-of-control systems. These two dimensions, we argue, offer
distinct perspectives on understanding the risks and opportunities of AI. As these two
dimensions are more closely tied to societal elements rather than technological elements,
they provide a basis for assessing how AI applications align with or challenge societal norms
and values.</p>
      <p>
        AI visibility can be defined as the extent to which users are aware of the presence and
operation of AI when interacting with it. This dimension is strongly associated with
elements instrumental in building trustworthy AI, such as transparency, interpretability,
and accountability. Drawing on concepts from information systems integration [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and
egovernment integration [21, 24], we argue that AI visibility will decrease as AI systems
become more seamlessly integrated into other systems. The reasoning is that as these AI
systems blend or become incorporated into larger information systems, the components
unique to AI systems become less visible and more challenging for users to identify. For
example, in AI-powered traffic management systems, where AI technologies are deeply
integrated into larger information systems such as routing and transportation systems, it
may become difficult for most users to recognize or trace the use of AI.
      </p>
      <p>
        AI autonomy, additionally, can be defined as the level of human control and oversight
present when the AI system is executing or operating. This dimension closely relates to the
concepts of 'human-algorithm interactions' [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ] and 'human-in-loops' [32], indicating
the inverse relationship between the degree of human intervention and the degree of AI
autonomy. Given the rising ethical risks and unforeseen impacts associated with machine
learning algorithms, countries and supranational organizations have started to promote
regulatory frameworks and guidance aimed at overseeing algorithmic systems and
initiating intervention when necessary [
        <xref ref-type="bibr" rid="ref13 ref18">13, 18, 28</xref>
        ]. Green, for instance, collected and
summarized various practical guidelines for human oversight of algorithms use in the
public sector, identifying three key elements for human oversight of algorithmic systems
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]: restricting solely automated decisions, emphasizing the importance of human
discretion, and requiring meaningful human input.
      </p>
      <p>Based on the degree of visibility and autonomy of AI systems, Table 1 lists some
prevalent AI-powered applications that exhibit varying levels of these attributes. Chatbots,
which have high visibility and autonomy, usually directly interact with users at the forefront
and make autonomous decisions within their programmed capabilities. In most situations,
users are clearly aware that the agents they interact with are powered by AI technologies,
while the content produced by those agents might not be fully determined by humans. On
the other hand, recommendation systems also exhibit high AI visibility, despite their lower
autonomy. These systems guide users or promote content based on predefined algorithms
and data sets, yet they do not function entirely autonomously. Taking recommendation
systems used by social media platforms as an example, most companies employ human
review teams to ensure sensitive content is detected and, if applicable, rescinded swiftly.</p>
      <p>On the other end of the spectrum, facial recognition systems, while highly autonomous
in their operation, tend to have low visibility in terms of direct user interaction. As these
systems frequently work behind the scenes in security and monitoring applications, users
may not easily notice that they are being screened by an AI application. Predictive policing
systems are also less visible to the public, as they are typically integrated into broader law
enforcement systems to support decision-making for police agencies. Furthermore, since
the results generated by predictive policing systems usually require examination and
approval by police officers, the degree of autonomy in these systems could be considered
low.</p>
      <p>Rooted in previous studies, we build a basic typology of AI applications based on their
visibility and autonomy. This basic typology offers three-fold advantages in making sense
of the complexities of AI systems. First, as mentioned above, these two dimensions are
highly associated with critical issues such as explainability, responsibility, accountability,
and discretion in the decision-making process, linking AI to important theoretical debates.
Second, it provides a coherent classification for understanding a variety of AI use cases in
the public sector, covering a wide range of use cases. Third, rather than merely focusing on
technological elements, these two dimensions offer an intricate societal perspective on
understanding the interaction between users (e.g., citizens) and service providers (e.g.,
government agencies), as these interactions can be shaped by the level of visibility and
autonomy in AI applications.</p>
      <p>In addition, this initial typology can be further developed to provide a deeper
understanding of the complexity of AI. One crucial difference between AI applications and
other information technologies is that most AI systems include learning attributes and the
outputs of AI systems are not purely restricted by rules designed by humans. This means
that the adoption of AI systems in the public sector introduces a new thread of uncertainty
in the decision-making process. To account for this new uncertainty, we argue that the
current typology could be better understood when incorporating the unique self-learning
element of AI technologies. The next section will briefly outline the characteristic of
selflearning attributes of AI, as well as how this self-learning characteristic impacts the
operations of AI systems.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Self-learning as the main characteristic of AI</title>
      <p>
        The discussion about the self-learning characteristic of AI is far from new. Tracing back to
the 1950s, one critical aspect of AI is its self-improvement attribute, which implies that
machines are capable of learning, inferencing, and forming concepts [
        <xref ref-type="bibr" rid="ref17">17, 27</xref>
        ]. While current
AI applications do not meet the acceptable standard of an unboundedly self-improving
machine or artificial general intelligence, the concept of self-learning AI opens up important
possibilities for AI [
        <xref ref-type="bibr" rid="ref17 ref2">2, 17</xref>
        ]. For example, self-adaptive software is designed to modify its
behaviors in response to changes in its operating environment, enabling systems to
autonomously adjust themselves based on feedback from their current performance [26,
31]. This approach allows designers to initially program AI systems and then let the systems
learn the rest for themselves based on the data fed to them.
      </p>
      <p>
        The self-learning characteristic highlights at least two positive opportunities for
deploying AI systems. First, AI systems with self-learning capabilities can enhance their
model performance to respond to evolving environments. For instance, reinforcement
learning models have shown their potential in assisting health professionals with pandemic
control measures [23, 25]. Second, based on large amounts of training data, self-learning
algorithms can modify parameters or weights in their models to make optimal predictions
about input-output relationships. Given this capability, most self-learning algorithms can
leverage these predicted input-output relationships to forecast results on 'as-yet-unseen'
data [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This characteristic has led to the proliferation of AI systems in areas that can utilize
specific input-output relationships to optimize their decisions, such as recommendation
systems or risk assessment systems.
      </p>
      <p>
        These opportunities, however, come with associated challenges. Since the unique
selflearning characteristic has influenced the extent of public discretion within public
organizations, a primary concern when implementing AI in the public sector is the potential
generation of fault scenarios in which bureaucrats lack prior experience [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]. That is, the
probability of encountering unexpected consequences and applying different standards in
the same process increases due to the self-learning characteristic. Furthermore, the
selflearning feature complicates the 'black-box' issue, which is already widely discussed by
scholars and practitioners [e.g., 6, 7, 23, 31]. As Busuioc clearly states, the issue of algorithm
complexity could exacerbate traditional information asymmetry problems and diminish
users' capability to conduct reasonable assessments [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Last, but not least, this self-learning
characteristic could potentially lead to a runaway feedback loop that reinforces and
pronounces existing social injustices [
        <xref ref-type="bibr" rid="ref11 ref9">9, 11</xref>
        ]. For instance, based on the case of predictive
policing systems, research found that the systems would continually direct police to the
same neighborhoods, irrespective of the actual crime rates [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Given this unique and crucial aspect of AI systems, we argue that researchers should play
closer attention to the self-learning characteristic and explore its intersections with other
important features of AI systems. As mentioned before, when public organizations deploy
AI systems, which normally includes the self-learning characteristic, they could face
situations that intersect with profound issues in the public sector, such as discretionary
power, transparency, accountability, or administrative discrimination. Therefore, this study
aims to take the self-learning characteristic into account, integrating it with two other
critical dimensions used to classify different AI systems. In doing so, we expect to produce
a more holistic and dynamic picture of the opportunities and challenges of different AI
systems, based on the illustration of real-world cases. In the next section, we briefly outline
our case selection process.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Case selection process</title>
      <p>
        To identify real-world cases, this study focuses on AI cases within the federal government
of the United States. Based on Executive Order No. 13960 issued in 2020, titled 'Promoting
the Use of Trustworthy Artificial Intelligence in the Federal Government,' the U.S. federal
government has established a website to compile all AI use cases across federal agencies
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The website has documented 710 AI cases deployed in federal agencies, containing
basic information for each case, such as the deploying agency, summary of AI projects,
techniques used, and source code. Given this rich dataset, we identified four cases that
exhibit different degrees of AI visibility and autonomy.
      </p>
      <p>Specifically, we reviewed the summaries of each AI project to identify four cases we used
for further analysis. We conducted three steps to select the appropriate cases. First, as many
AI projects are designed for internal tasks or managerial purposes, such as automatically
scanning barcodes in documents or detecting spam emails, we excluded those cases that are
not related to services or interactions with citizens. Second, we prioritized cases that could
be clearly categorized by our proposed typology based on AI visibility and autonomy. To
select the cases that are more appropriate for further analysis, we used the aforementioned
definitions of AI visibility and autonomy (the level of integration with other systems and the
level of human intervention) to evaluate the description of each AI project. Therefore,
projects whose description is primarily technical are excluded due to the lack of sufficient
information to categorize them into the current typology. Third, while some federal
agencies, such as the Social Security Administration, might have multiple cases that meet
our criteria due to the nature of their tasks, we decided to select cases from different federal
agencies to ensure more variability and broader potential impacts of our results.</p>
      <p>
        Through the case selection process, four cases across different federal agencies were
identified. These include the Aidan Chatbot deployed by the Department of Education [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ],
the Medication Safety Clinical Decision Support system deployed by the Department of
Veterans Affairs [29], the Person-Centric Identity Services system deployed by the
Department of Homeland Security [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and the Quick Disability Determinations system
deployed by the Social Security Administration [33]. In the next section, we illustrate the
attributes of visibility, autonomy, and self-learning for each case, and discuss specific
opportunities and challenges.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Preliminary results</title>
      <p>This section outlines the content of the four selected AI cases. Opportunities and challenges
associated with the self-learning characteristic of AI are also discussed.</p>
      <sec id="sec-5-1">
        <title>5.1. High visibility and high autonomy: Aidan Chatbot</title>
        <p>Deployed by the Department of Education, a virtual assistant named Aidan is designed with
the aim of responding to common questions about federal student aid. Powered by natural
language processing technology, the Aidan Chatbot helps users figure out their current loan
account balance, grants information, or the contact information for loan servicers. Similar
to other chatbot applications, the interface of the Aidan Chatbot is straightforward and
easily recognizable as being powered by AI-related technologies. In fact, on the
StudentAid.gov website, where Aidan services are offered, the Department of Education
explicitly points out that this virtual assistant is powered by AI technologies. Furthermore,
most responses from the Aidan Chatbot are not reviewed by human agents before being
sent to users. Therefore, it can be categorized as a classic application with a high degree of
visibility and autonomy.</p>
        <p>Taking the self-learning feature into account, this type of AI application can generate
customized responses by learning from users' feedback and input. Specifically, Aidan will
keep a record of users' conversations and request logs to improve the quality of future
interactions. Given this, the self-learning characteristic provides greater opportunities for
human-machine collaboration, since it enables the system to adapt and improve over time
based on real-world interactions. On the other hand, the self-learning characteristic might
introduce specific challenges to the Aidan Chatbot. As there is no clear checking point
between AI systems and users, the risk of reproducing large-scale inaccurate and biased
responses might increase if the original data is biased or contains errors. As a result,
establishing certain processes for human intervention in certain situations or at some
regular time intervals could be helpful to prevent these potential issues.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. High visibility and low autonomy: Medication Safety Clinical Decision Support</title>
        <p>The Medication Safety Clinical Decision Support system, deployed by the Department of
Veterans Affairs, features evidence-based recommendations for primary care providers.
Incorporating various electronic clinical data, such as laboratory test results and history of
adverse drug events, the system automatically offers patient-specific recommendations to
care providers. By doing so, this AI application assists veterans and their care providers in
managing chronic and other types of diseases. This AI project is highly visible to its users as
it is not integrated into a complex decision-making structure. Users can easily identify that
the system is making the recommendations. Additionally, the recommendations provided
by the system does not directly make the decision regarding medicine use or other
healthrelated treatments, which reflects a relatively low degree of AI autonomy in this case.</p>
        <p>Considering the self-learning characteristic of this decision support system, one crucial
opportunity is to enhance the precision and personalization of the recommendations over
time. Since users can perceive the iterative learning processes that provide better alignment
between patients' needs and treatments, it creates a sense of trustworthiness and a positive
experience in using this AI application. Nevertheless, the results recommended by the
system might not always lead to final decisions, and algorithms are designed to learn from
interventions made by humans, such as tagging invalid recommendations or offering
alternative responses. Specific challenges might emerge when these learning processes,
intentionally or unintentionally, incorporate human biases, thereby limiting the system's
effectiveness.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Low visibility and high autonomy: Person-Centric Identity Services</title>
        <p>The goal of the Person-Centric Identity Services system is to become a trusted source that
profiles an individual's comprehensive immigration history and status. Deployed by the
Department of Homeland Security, this system aims to establish an identity profile by
compiling and aggregating various biographic and biometric information with the
assistance of machine learning algorithms. Since this system is highly intertwined with
other operational systems in the Department of Homeland Security, the visibility of this AI
application appears to be low. Moreover, the algorithms used to match individuals' different
immigration records are highly autonomous, and explicit human intervention is limited due
to the high complexity of the data structure.</p>
        <p>The characteristic of self-learning presents both opportunities and challenges to this
highly autonomous but less visible AI system. On the upside, the self-learning attribute
could cover more data sources having immigration records and streamline the matching
procedures from various data sources, which can significantly reduce routine tasks and
increase the operational efficiency of agencies. On the downside, however, the self-learning
characteristic in this system might entail higher risks and challenges in rectifying or
detecting potential administrative errors. As public managers or experts might lose
sufficient control during the system's iterative learning processes, the concerns about the
'black-box' effect would be exacerbated.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Low visibility and low autonomy: Quick Disability Determinations</title>
        <p>The Quick Disability Determinations system, deployed by the Social Security Administration,
utilizes an AI-powered model to initially screen applications submitted for disability
benefits. Building on historical data from completed claims, the system identifies claimants
with the most severe disabilities. Subsequently, public employees in the Social Security
Administration prioritize cases where a favorable disability determination is highly likely
and medical evidence is readily available, thereby expediting the application process for
those who are the most vulnerable and in high need. Given that this system is integrated
into the complex structure of the social security benefits system, its visibility could be
considered low. Furthermore, since public employees in social security agencies can control
the final decisions of benefit claims, the autonomy of this system is also deemed low.</p>
        <p>Factoring in the self-learning aspect, the Quick Disability Determinations system could
encounter distinct opportunities and challenges. In terms of opportunities, the self-learning
characteristic can assist human workers to a more manageable decision-making process.
The frequent human intervention could also help to consider very specific characteristics of
certain difficult or unique cases. On the other hand, given the high degree of human
intervention and the high degree of system integrations, the challenges might be centered
on issues related to accountability and evaluating the performance of the models. That is,
the decisions of approving or rejecting disability applications might result from several
factors, such as employees' prejudice or decision structures in social security benefit
applications, rather than from the AI system itself. Given the ambiguous feedback loops
between inputs and outputs, the system might not achieve its full potential from the
selflearning characteristic.</p>
        <p>Based on the cases we illustrated, Table 2 summarizes our preliminary results in terms
of tailored opportunities and challenges of different types of AI applications, when taking
the self-learning characteristic into account. Generally, we underscore that the self-learning
feature can be an important factor contributing to the complexities of benefits and other
consequences. By explicitly recognizing the interplay of visibility, autonomy, and
selflearning in government AI systems, practitioners and researchers can analyze AI
applications under the analytical lens of specific contexts and types. We will conclude our
research findings and outline further steps in the next section.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Final comments and next steps</title>
      <p>With four AI cases deployed in U.S. federal agencies, this ongoing research study contributes
to previous research by exploring the advantages and implications of a more complex
categorization of AI systems [22, 35]. Our preliminary findings suggest that the self-learning
characteristic might introduce different opportunities and challenges for different types of
AI systems. Considering the interplay of visibility, autonomy, and self-learning in
government AI systems, AI systems distinguish themselves from previous technology issues
regarding the complexities and challenges. Thus, it might be fair to argue that the proposed
typology consisting of two dimensions—visibility and autonomy—shows the importance of
understanding complex AI systems in the public sector due to the unique self-learning
nature of AI technologies. Given that these two dimensions highlight essential aspects for
the public sector, future research can utilize these dimensions as key contextual
differentiators among AI systems.</p>
      <p>For future directions, a deeper analysis of the cases would be beneficial. The current
analysis is based on official case documents available on government websites, but the
descriptions of cases in these documents provides limited information. As a result,
conducting interviews with public employees who actively engage with these AI systems,
as well as with citizens and stakeholders affected by the decisions, could yield richer
insights in future studies. Furthermore, incorporating additional cases for comparison
within the same category could be beneficial. The comparison might illuminate the subtle
distinctions among AI cases, teasing out the critical factors that merit closer examination in
AI systems even when they are in the same category.
[20] Heinrichs, B. 2022. Discrimination in the age of artificial intelligence. AI &amp; SOCIETY. 37,
1 (Mar. 2022), 143–154.
[21] Jochen Scholl, H.J. and Klischewski, R. 2007. E-Government Integration and
Interoperability: Framing the Research Agenda. International Journal of Public
Administration. 30, 8–9 (Jul. 2007), 889–920.
[22] Katzenbach, C. and Ulbricht, L. 2019. Algorithmic governance. Internet Policy Review.</p>
      <p>8, 4 (Nov. 2019), 1–18.
[23] Kwak, G.H. et al. 2021. Deep reinforcement learning approaches for global public health
strategies for COVID-19 pandemic. PLoS ONE. 16, 5 (May 2021), e0251550.
[24] Lam, W. 2005. Barriers to e-government integration. Journal of Enterprise Information</p>
      <p>Management. 18, 5/6 (2005), 511–530.
[25] Libin, P.J.K. et al. 2021. Deep Reinforcement Learning for Large-Scale Epidemic Control.</p>
      <p>Machine Learning and Knowledge Discovery in Databases. Applied Data Science and
Demo Track (Cham, 2021), 155–170.
[26] Macías-Escrivá, F.D. et al. 2013. Self-adaptive systems: A survey of current approaches,
research challenges and applications. Expert Systems with Applications. 40, 18 (Dec.
2013), 7267–7279.
[27] McCarthy, J. et al. 2006. A Proposal for the Dartmouth Summer Research Project on</p>
      <p>Artificial Intelligence, August 31, 1955. AI Magazine. 27, 4 (Dec. 2006), 12–12.
[28] Medaglia, R. et al. 2023. Artificial Intelligence in Government: Taking Stock and Moving</p>
      <p>Forward. Social Science Computer Review. 41, 1 (Feb. 2023), 123–140.
[29] Medication Safety (MedSafe) Clinical Decision Support (CDS): 2024.
https://www.actiac.org/et-use-case/medication-safety-medsafe-clinical-decisionsupport-cds. Accessed: 2024-03-07.
[30] OECD 2022. OECD Framework for the classification of AI systems.
[31] Oreizy, P. et al. 1999. An architecture-based approach to self-adaptive software. IEEE</p>
      <p>Intelligent Systems and their Applications. 14, 3 (May 1999), 54–62.
[32] Peeters, R. 2020. The agency of algorithms: Understanding human-algorithm
interaction in administrative decision-making. Information Polity. 25, 4 (Jan. 2020),
507–522.
[33] Quick Disability Determinations (QDD): 2024.</p>
      <p>https://www.ssa.gov/disabilityresearch/qdd.htm. Accessed: 2024-03-07.
[34] Ruijer, E. et al. 2023. Social equity in the data era: A systematic literature review of
datadriven public service research. Public Administration Review. 83, 2 (Mar. 2023), 316–
332.
[35] Straub, V.J. et al. 2023. Artificial intelligence in government: Concepts, standards, and a
unified framework. Government Information Quarterly. 40, 4 (Oct. 2023), 101881.
[36] Veale, M. and Brass, I. 2019. Administration by Algorithm? Public Management Meets</p>
      <p>Public Sector Machine Learning.
[37] Vogl, T.M. et al. 2020. Smart Technology and the Emergence of Algorithmic Bureaucracy:
Artificial Intelligence in UK Local Authorities. Public Administration Review. 80, 6
(2020), 946–961.
[38] Wirtz, B.W. et al. 2019. Artificial Intelligence and the Public Sector—Applications and
Challenges. International Journal of Public Administration. 42, 7 (May 2019), 596–615.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Benbya</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          et al.
          <year>2020</year>
          .
          <article-title>Artificial Intelligence in Organizations: Current State and Future Opportunities</article-title>
          .
          <source>MIS Quarterly Executive</source>
          .
          <volume>19</volume>
          ,
          <issue>4</issue>
          (Dec.
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Brynjolfsson</surname>
            , E. and Mitchell,
            <given-names>T.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>What can machine learning do? Workforce implications</article-title>
          .
          <source>Science</source>
          .
          <volume>358</volume>
          ,
          <issue>6370</issue>
          (Dec.
          <year>2017</year>
          ),
          <fpage>1530</fpage>
          -
          <lpage>1534</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bullock</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          et al.
          <year>2020</year>
          .
          <article-title>Artificial intelligence, bureaucratic form, and discretion in public service</article-title>
          .
          <source>Information Polity</source>
          .
          <volume>25</volume>
          ,
          <issue>4</issue>
          (Jan.
          <year>2020</year>
          ),
          <fpage>491</fpage>
          -
          <lpage>506</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Bullock</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2019</year>
          . Artificial Intelligence, Discretion, and
          <string-name>
            <surname>Bureaucracy</surname>
          </string-name>
          .
          <source>The American Review of Public Administration</source>
          .
          <volume>49</volume>
          ,
          <issue>7</issue>
          (Oct.
          <year>2019</year>
          ),
          <fpage>751</fpage>
          -
          <lpage>761</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Burrell</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>How the machine 'thinks': Understanding opacity in machine learning algorithms</article-title>
          .
          <source>Big Data &amp; Society. 3</source>
          ,
          <issue>1</issue>
          (Jun.
          <year>2016</year>
          ),
          <fpage>205395171562251</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Busuioc</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2021</year>
          . Accountable Artificial Intelligence: Holding Algorithms to Account. Public Administration Review.
          <volume>81</volume>
          ,
          <issue>5</issue>
          (
          <year>2021</year>
          ),
          <fpage>825</fpage>
          -
          <lpage>836</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.-C.</given-names>
          </string-name>
          et al.
          <year>2023</year>
          .
          <article-title>Artificial Intelligence and Public Values: Value Impacts and Governance in the Public Sector</article-title>
          . Sustainability.
          <volume>15</volume>
          ,
          <issue>6</issue>
          (Jan.
          <year>2023</year>
          ),
          <fpage>4796</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8] DHS/USCIS/PIA-087
          <string-name>
            <surname>Person Centric Identity Services Initiative</surname>
          </string-name>
          :
          <year>2022</year>
          . https://www.dhs.gov/publication/dhsuscispia-087
          <string-name>
            <surname>-</surname>
          </string-name>
          person
          <article-title>-centric-identity-servicesinitiative</article-title>
          .
          <source>Accessed: 2024-03-07.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Ensign</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          et al.
          <year>2018</year>
          .
          <article-title>Runaway Feedback Loops in Predictive Policing</article-title>
          .
          <source>Proceedings of the 1st Conference on Fairness, Accountability and Transparency (Jan</source>
          .
          <year>2018</year>
          ),
          <fpage>160</fpage>
          -
          <lpage>171</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Federal</given-names>
            <surname>Student Aid</surname>
          </string-name>
          :
          <year>2024</year>
          . https://studentaid.gov/h/aidan. Accessed:
          <fpage>2024</fpage>
          -03-07.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Ferrara</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2024</year>
          .
          <article-title>The Butterfly Effect in artificial intelligence systems: Implications for AI bias and fairness</article-title>
          .
          <source>Machine Learning with Applications</source>
          .
          <volume>15</volume>
          ,
          <string-name>
            <surname>(Mar</surname>
          </string-name>
          .
          <year>2024</year>
          ),
          <fpage>100525</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Government</surname>
            <given-names>Use of AI</given-names>
          </string-name>
          :
          <year>2024</year>
          . https://ai.gov/ai-use-cases/. Accessed:
          <fpage>2024</fpage>
          -03-07.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Green</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2022</year>
          .
          <article-title>The flaws of policies requiring human oversight of government algorithms</article-title>
          .
          <source>Computer Law &amp; Security Review</source>
          .
          <volume>45</volume>
          ,
          <string-name>
            <surname>(Jul</surname>
          </string-name>
          .
          <year>2022</year>
          ),
          <fpage>105681</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Green</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>The Principles and Limits of Algorithm-in-the-Loop Decision Making</article-title>
          .
          <source>Proceedings of the ACM on Human-Computer Interaction. 3</source>
          ,
          <issue>CSCW</issue>
          (7
          <year>2019</year>
          ),
          <volume>50</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>50</lpage>
          :
          <fpage>24</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Grimmelikhuijsen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2023</year>
          .
          <article-title>Explaining Why the Computer Says No: Algorithmic Transparency Affects the Perceived Trustworthiness of Automated Decision-Making</article-title>
          .
          <source>Public Administration Review</source>
          .
          <volume>83</volume>
          ,
          <issue>2</issue>
          (
          <year>2023</year>
          ),
          <fpage>241</fpage>
          -
          <lpage>262</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Grimmelikhuijsen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Meijer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2022</year>
          .
          <article-title>Legitimacy of Algorithmic Decision-Making: Six Threats and the Need for a Calibrated Institutional Response</article-title>
          .
          <source>Perspectives on Public Management and Governance. 5</source>
          ,
          <issue>3</issue>
          (Sep.
          <year>2022</year>
          ),
          <fpage>232</fpage>
          -
          <lpage>242</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Hall</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>Self-improving AI: an Analysis</article-title>
          .
          <source>Minds and Machines</source>
          .
          <volume>17</volume>
          ,
          <issue>3</issue>
          (Oct.
          <year>2007</year>
          ),
          <fpage>249</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Harrison</surname>
            ,
            <given-names>T.M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Luna-Reyes</surname>
            ,
            <given-names>L.F.</given-names>
          </string-name>
          <year>2022</year>
          .
          <source>Cultivating Trustworthy Artificial Intelligence in Digital Government. Social Science Computer Review</source>
          .
          <volume>40</volume>
          ,
          <issue>2</issue>
          (Apr.
          <year>2022</year>
          ),
          <fpage>494</fpage>
          -
          <lpage>511</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Hasselbring</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>Information system integration</article-title>
          .
          <source>Communications of the ACM. 43</source>
          ,
          <issue>6</issue>
          (Jun.
          <year>2000</year>
          ),
          <fpage>32</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>