<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Heli Kautonen</string-name>
          <email>heli.kautonen@finlit.fi</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Gasparini</string-name>
          <email>a.a.gasparini@ub.uio.no</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Åbo Akademi University, Information Studies</institution>
          ,
          <addr-line>Tuomiokirkontori 3, 20500 Turku</addr-line>
          ,
          <country country="FI">Finland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Finnish Literature Society Library</institution>
          ,
          <addr-line>P.O. Box 259, 00171 Helsinki</addr-line>
          ,
          <country country="FI">Finland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Oslo University Library</institution>
          ,
          <addr-line>Moltke Moes vei 39, 0851 Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Proceedings Tethics, Conference on Technology Ethics</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1910</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>Recent advances in artificial intelligence (AI) applications have raised concerns about the consequences of the uncontrolled development of AI technology for society and humans. Information and knowledge professionals working in research libraries are in professions that have long existed and have globally applied ethical codes that serve as self-regulatory ethical norms. New AI technologies that penetrate throughout libraries' operations cause confusion among librarians and challenge the existing ethics. In this paper, we examine these challenges and present a qualitative study that reveals the ethical considerations that research librarians face when they approach new AI technologies. As there are no established AI ethics norms for research librarians, we compared the international code of conduct for libraries against the European AI guidelines to identify relevant themes for our study. We analyzed the data from two Scandinavian workshops for librarians. Our findings highlight the central role of research libraries in making AI-powered research ethical. Our study also indicates a need to update international codes of conduct for libraries for the AI age by including aspects of AI agency and the interests of future generations. This helps librarians better orient themselves and their patrons towards a trustworthy and existentially sustainable future with AI systems. Research libraries, academic libraries, artificial intelligence, ethics.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Recent and impressive advances in large language models have raised serious concerns about the
consequences of the uncontrolled and business-led race to develop artificial intelligence (AI)
technologies [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Following the initial excitement about the endless-seeming possibilities of these
human-competing digital systems, many of us have woken up to the risks to society and humans [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2–
4</xref>
        ]. In other words, the voices calling for existential sustainability are becoming louder, and the need
for AI ethics is becoming more imperative.
      </p>
      <p>
        Before the current burst of public discussion, information and knowledge professionals working in
research libraries had anticipated the advent of AI in their work contexts. The earliest scenarios for
intelligent machine-operated library tasks were drafted in the mid-1970s. In her article, Smith [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
envisioned automated information retrieval systems and used the term AI. In the past decade, interest
in AI in libraries has grown exponentially (see Figure 1).
      </p>
      <p>2023 Copyright for this paper by its authors.
CEUR</p>
      <p>ceur-ws.org</p>
      <p>
        Figure 1: Publications per year about AI or machine learning in the context of academic or
research libraries. The authors generated the statistics for a literature review [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        The literature indicates that the research library community approaches this new technology with
uncertainty. Many reports demonstrate how AI offers better opportunities than old technologies for
improved library operations and services [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The library community also recognizes the risks. The
power and obscurity of these new tools can be considered a threat to librarians and library patrons
(students, researchers, and citizens) if AI generates biases that exploit users or distort research [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
The ethical implications of AI are slowly alarming the entire scholarly community. The integrity of
the research and the explainability and robustness of the research results gained with the help of AI
tools are also viewed as threatened [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        In addition to ethical norms or regulations set by society's governing institutions, ethics can be
maintained at the individual level through professional codes of conduct. Globally conformed ethical
codes that grow from and build the self-image of librarians have existed and been adhered to for many
years. Since the first formal ethical code for librarians was published in the 1930s [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], the profession
has explicitly expressed its values and principles. The most recent code of conduct that unites
librarians and information workers across countries and organization types was published by the
International Federation of Library Associations and Institutions (IFLA) in 2012 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This document
begins by manifesting the ethical requirements for being a librarian: “Librarianship is, in its very
essence, an ethical activity embodying a value-rich approach to professional work with information.”
      </p>
      <p>
        In this paper, we approach the ethical challenges posed by AI for our society by presenting the
results of a qualitative study of the existing and potentially emerging ethical norms of one
profession—research librarians. The aim is to examine the self-regulative potential of this profession
when AI technologies are emerging forcefully throughout library operations and changing librarians’
work. This study is part of action research [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], where the authors have a dual role: While we serve
our local research and library communities by developing services that explore the potential of AI
technologies, we also research the phenomenon as information science scholars.
      </p>
      <p>Our research question is: What ethical considerations do research librarians obtain when they
explore and approach new AI technologies?</p>
      <p>We limited our examination to Europe because our empirical data is from Scandinavian research
libraries. The European research library community can be considered homogenous because librarians
in Europe have similar educational backgrounds and are well-networked. Academic and other
research libraries are also well connected through national and international library associations, such
as LIBER, the Association of European Research Libraries, and librarians have the means to learn the
ethics of their profession during their education or through library networks.</p>
      <p>In this study, we use the concept of AI in the wide meaning that dominates popular and academic
discourses. In these discussions, the acronym AI includes methods, technologies, applications, and
research approaches—viewpoints from recent advances in machine learning (ML) to futuristic
imaginings. As we focus on a period in which the understanding of this new phenomenon in libraries
is vague, we consider it relevant to include all of these diverse viewpoints despite their heterogeneous
and even contradictory interpretations.</p>
      <p>In the following chapters, we present the conduct of our study, starting with a description of our
methods. We describe our findings in detail, discuss their implications, and summarise our
conclusions.</p>
      <p>1.1.</p>
    </sec>
    <sec id="sec-2">
      <title>Trustworthy and ethical AI</title>
      <p>
        The ethical development and use of AI systems in libraries is grounded on trust. First, the
trustworthiness and integrity of AI systems and their output is an ethical premise that entails AI
technologies working "for the good of humanity, individuals, societies and the environment and
ecosystems” [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Second, the trustworthiness of information has been the ethical mainstay of
libraries for millennia. The digital shift, or, as Ovenden [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] describes it, the “digital deluge”, poses
an existential challenge to the role and mission of libraries.
      </p>
      <p>
        Trust is principally based on the interest of the trusted [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Trust requires a willingness to be
vulnerable to another human and to believe that the trustee will fulfil the agreed-upon commitments
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. As Wheeless and Grotz [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] point out, “trust is antecedent to a willingness to disclose.”
      </p>
      <p>
        To build trust in AI-powered systems, one needs to show theoretical guarantees, such as
algorithmic provenance and dependencies [
        <xref ref-type="bibr" rid="ref15 ref16">15,16</xref>
        ], so that the outcome is interpretable for
nonexperts. Lipton [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] argues that trustworthiness is based on how often the output (that is, the model) is
correct, and in which context and for which examples it is correct. One way to solve this issue is to
present AI-powered outcomes to users with visual or textual artefacts [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>When using library services, patrons need to trust the quality and reliability of the information
provided by a library as an organization and the integrity of its staff [19]. Only then can users disclose
their needs and lack of skills. Luckily, libraries still have a high level of trust in society [20,21].</p>
    </sec>
    <sec id="sec-3">
      <title>2. Methodology</title>
      <p>As there are no established or widely available AI ethics guidelines for research librarians, we
began our study by examining two existing norms provided by organizations with a position of
authority among European research libraries. The first reveals the viewpoint of libraries, and the
second concerns AI technology. Then, to see whether these norms play any role in practice—if
librarians have ethical considerations regarding AI in their work—we studied the data from two
workshops we conducted in 2022. In these workshops, Scandinavian academic librarians explained
and shared their understanding of AI in their work.</p>
      <p>
        In the first part of the study, we compared the contents of two documents that describe current
ethical norms: The first document is the IFLA Code of Ethics for Librarians and Other Information
Workers (full version), published by IFLA in 2012 (hereafter IFLA Codes) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This document
captures and explicates librarians’ ethical norms, some of which have been applied for centuries, such
as the moral imperative to provide access to information [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>The second document that we inspected was the Ethics guidelines for trustworthy AI, created by
the European Commission (EC) and published in 2019 (hereafter AI Guidelines) [22]. Although these
guidelines emphasize the responsibility of AI technology developers, they are also intended to cover
the deployment and use of AI systems. As the document addresses “researchers” and “institutions”
[22], among other stakeholders, it is also intended for research librarians.</p>
      <p>
        Other relevant guidelines, such as the OECD AI principles [23] and UNESCO’s AI ethics [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ],
offer recommendations for the public sector and mention (public) libraries as examples. We limited
our inspection to the IFLA and EC documents, as, based on our professional experience, the OECD
and UNESCO and their documents lack normative power within the European research library
community.
      </p>
      <p>
        We aimed to find and dissect the key themes and, in particular, those ethical norms that are
characteristic of AI technologies. Drawing from recent research [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we assumed that the essence of
contemporary AI applications sets specific requirements for ethical considerations. Our goal was to
identify potential differences that may influence librarians’ abilities to consider ethical questions
specific to AI technology adoption and use.
      </p>
      <p>We compared the contents of the two documents using an ethnographic content analysis approach,
which enables the detection of significance and meaning in a particular context and culture [24]. As
an essential element in our action research methodology, we used visual representations to extract
relevant concepts, themes, and patterns from the texts, as well as to elaborate our findings. We used
the Miro online platform for visual analysis in both parts of the study. As an outcome of the first part
of the study, we generated a list of themes that indicated relevant ethical considerations about AI in
the context of research libraries.</p>
      <p>In the second part of the study, we analyzed the outcome of two workshops conducted in 2022.
Altogether, 45 Scandinavian librarians or professionals working for library services participated in
these on-site workshops. The goal was to help librarians approach the AI phenomenon that has
emerged in their realm and lead to sentiments of a fundamental change. In these workshops, we
facilitated the discussions and future planning exercises using designerly approaches and methods.
The workshops provided qualitative data in the form of handwritten texts and photographs. We
transcribed the participants’ notes on two exercises. The first encouraged participants to think about
AI strategy failures and remedies in their libraries (the so-called sabotage method), and the second
steered them to plan feasible AI activities in their work (the so-called back-casting method). We
analyzed the transcripts and elaborated our analysis visually on the Miro platform.</p>
      <p>The second analysis aimed to reveal the kind of ethical considerations that librarians express when
they are in the process of exploring and approaching new AI technologies. We used the theme list
from the first part of the study as a reference tool and detected equivalent expressions from the
workshop data.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Findings 3.1.</title>
    </sec>
    <sec id="sec-5">
      <title>Comparison of ethics guidelines</title>
      <p>To reveal whether there are differences in ethical principles specifically for the development and
use of AI technologies for librarians, we compared two key documents: the IFLA Codes and the AI
Guidelines. The comparison showed that the documents were conformant in many points that
addressed the importance of basic ethical norms such as privacy, protection of personal data, and
transparency of governance processes. Both documents also emphasize diversity and open access to
information for all (see the grey connectors in Figure 2).</p>
      <p>We identified the following technology-related ethical aspects from the AI Guidelines that had
neither equivalent nor adjacent expressions in the IFLA Codes. Still, we considered that these aspects
could be related to existing ethical principles in librarians’ work (see the blue connectors in Figure 2):
 Robustness and safety, both technical and social, affect the ethical use of information and
service quality. The current IFLA Codes do not mention technical systems and their obvious role
as instruments of access, although the IFLA Code number 2 mentions potential “barriers”. There is
also no indication that technology affects the “highest standards of service quality,” although the
quality depends on the systems’ robustness and safety.
 The highest standards of service quality are also dependent on responsible mechanisms that
ensure data governance, auditability, and accountability “for AI systems and their outcomes.” This
is important because AI systems are inclined toward “unfair bias”.
 In the era of AI, the highest service quality also requires an awareness of AI. This means that
if libraries are using or promoting AI-powered systems and services, they should inform their users
of the existence of AI and take care of traceability mechanisms.
 Informing users of AI-powered systems’ capabilities and limitations and educating them to
become aware of AI systems can be considered a new aspect of information literacy. This is one of
the core services that research libraries provide to their patrons.</p>
      <p>We detected three discrepancies between the two ethics documents, which reflect ambivalent value
propositions regarding research libraries and AI (see the red connectors in Figure 2):
 Acknowledgement of the agency of AI technology: The IFLA document does not
acknowledge the autonomous agency of technology. The IFLA Codes number 2 uses the
expression “autonomous users,” but this refers to individual humans who access library services
without external help. The first requirement in the AI Guidelines concerns human agency, which
can be interpreted as an opponent to AI agency.
 Governance versus citizens’ freedoms: The AI Guidelines imply the importance of
governance by society when it uses expressions such as “data governance mechanisms” and
“legitimate access to data.” On the other hand, the IFLA document emphasizes a citizen’s freedom
through expressions such as “scrutiny of the general public” and “reject… censorship … by states,
governments… or civil society institutions.”.
 Interests of future generations: The AI Guidelines number 6 explicitly expresses the
importance of addressing the needs of future generations, whereas the IFLA Codes seem to
address the needs of current library patrons, employees, and stakeholders. None of the IFLA codes
expresses a concern about the future.</p>
      <p>We also observed that the final IFLA code does not have an equivalent in the AI Guidelines. This
point focuses on collegial relations within a library and a library community, emphasizing fairness
and respect for colleagues. However, we interpreted a close connection between this code and other
IFLA Codes, as well as the AI Guidelines, as it expresses a general ethical principle: “Librarians and
other information workers strive to earn a reputation and status based on their professionalism and
ethical behavior.”
3.2.</p>
    </sec>
    <sec id="sec-6">
      <title>Librarians’ considerations</title>
      <p>From the first part of the study, we extracted 11 themes that we used to analyze the data from the
two workshops. We used these themes to code expressions of ethical considerations from the
workshop transcripts. The first column of Table 1 shows the themes (codes), the two columns in the
middle indicate the occurrences of coded expressions in the data from the workshops, and the
rightmost column provides examples of participants’ notes.</p>
      <sec id="sec-6-1">
        <title>Ambivalent ethics</title>
        <p>A1 Acknowledgement of
AI agency
A2 Governance vs. citizen
freedom
10
10
1
5</p>
      </sec>
      <sec id="sec-6-2">
        <title>Themes</title>
        <p>4</p>
        <p>We identified a few expressions in the thematic category of common ethics, such as a need to
protect users’ privacy or sensitive data, the provision of easy access for users, or a need to understand
users’ or librarians’ diverse interests and skills. Workshop 2 provided notes we interpreted as an
ethical consideration of societal well-being. For example, in its context, the expression “something
great for everyone” reflected the same idea as AI Guideline number 6: “Social and societal impact
should be carefully considered”.</p>
        <p>Most of the notes from both workshops concerned ethical considerations for technology. In the
exercise that involved thinking about failures and their counteractions, the workshop participants
considered aspects related to AI robustness and safety, responsible mechanisms that ensure the
trustworthy deployment of AI technologies, wise distribution of activities between stakeholders, and
accountability of processes and their outcomes. There were several suggestions to establish an AI
strategy for the participants’ own libraries. It is also worth mentioning that only two notes were made
about algorithmic bias in workshop 2.</p>
        <p>Under the ambivalent ethics theme, there was a greater dispersion of notes. We could not identify
any expressions from workshop 1 that reflected the discrepancies we recognized between the IFLA
Codes and the AI Guidelines. However, both workshops resulted in notes with collegial relations with
suggestions for building librarians’ professional identities through AI competence and collaboration
across organizational boundaries. In workshop 2, participants made notes about human agency over
AI, and about maintaining individuals’ freedom in the context of research communities. Again, it is
noteworthy that the notes provided no suggestions for future generations, although all the workshop
participants were from libraries with cultural heritage collections.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>4. Discussion</title>
      <p>Research libraries stand for trustworthiness and reliability in their information and knowledge
services. Sharing and transparency are among the ethical principles of the profession, as manifested in
the IFLA Codes. Exploiting the value of academic libraries as trustworthy institutions in the context
of AI has implications for the identities of these organizations.</p>
      <p>As our workshop participants noted, librarians should aim to build their own AI literacy to
supervise their patrons (see T2 and T3 in Table 1). This competence could then be used to highlight
AI tools’ possibilities and ethical pitfalls. For the time being, few libraries can offer this kind of
service. A set of ethical principles on how to govern the human use of AI in different contexts and the
implications thereof has emerged from several institutions and groups [25]. Auditability, satisfaction,
effectiveness, persuasiveness, efficiency, and trust are values that underpin transparency and
explanations of results in AI-based systems [26].</p>
      <p>
        This is the critical moment for research libraries to build their AI strategies, not least because of
the ethical considerations. If libraries do not take an active role in AI ecosystems and remain passive
users of AI-powered services, they may become obsolete. Academic and other research libraries must
decide whether to keep their role as mediators of trustworthy information based on ethical principles
(see E1 and E2 in Table 1). Various observations from our study pinpoint transparency, open access
to information, and responsible mechanisms as the most important ethical values to consider (see E1,
E2, and T2 in Table 1). These observations conform with voices from the field calling to safeguard
libraries' values [
        <xref ref-type="bibr" rid="ref3">3,27,28</xref>
        ].
      </p>
      <p>Another observation is that librarians must acquire new skills and competencies to cope with
ethical issues when using AI-powered tools and providing AI-enhanced services. These may include
the copyright of the output of a service, possible hidden biases in training data for algorithms, and an
understanding of where training data originates (see T3 and T4 in Table 1). New competencies have
become more critical because the entire knowledge ecosystem, including research communities,
research libraries, and academic publishers, is in the middle of AI transformation. While research
libraries are probing their role with AI technologies, academic publishers are improving their
production processes and services using the power of algorithms [29]. Software companies are also
capitalizing on AI technologies and marketing their innovations to the wide academic community,
from individual researchers to national libraries—a fact also mentioned by participants of workshop
number 2: “[A road to failure:] Forget that most AI tools are commercial”.</p>
      <p>
        While ethics and trust are closely related in libraries, they often remain undiscussed because they
are taken-for-granted components of a librarian’s identity. In general, research libraries need to
consider ethical problems when using AI “so that research libraries will continue to serve as trusted
advisors to our users, and as responsible collectors, disseminators, and preservers of knowledge” [30].
Algorithmic bias is an intensively debated issue since libraries are brokers of a large amount of data,
both their own and that produced by university users [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Finally, we observe that neither the IFLA Codes nor the librarians who attended our workshops
considered the interests of future generations. A growing body of research on societal and cultural
sustainability challenges existing work practices, including in libraries. Research libraries can be
considered fortresses of cultural sustainability if they comply with “development that meets the needs
of the present without compromising the ability of future generations to meet their own needs” [31].
To reach this ethical norm, librarians need to consider their future patrons. It may not be feasible to
anticipate the expectations of several generations ahead, but consideration of the next generation
would be an ethical act.</p>
    </sec>
    <sec id="sec-8">
      <title>5. Conclusion</title>
      <p>In this study, we examined the ethical challenges AI poses in the context of research libraries and
from the perspectives of research librarians. We examined two ethics norms that provide an
understanding of established library ethics, on the one hand, and amended AI ethics, on the other.
From the comparison of these normative documents, we extracted 11 themes for ethical consideration.
Then, we analyzed the empirical data from two Scandinavian workshops and reflected the data
against these themes.</p>
      <p>Our findings indicate that the context of libraries is crucial to maximizing the ethical
trustworthiness of AI-powered services for research communities. Librarians will need to guarantee
the reliability of the knowledge provided by AI systems. Our findings also show that librarians want
to understand the “black box” where AI “magic” happens, and involving them during future projects
is mandatory.</p>
      <p>In the past year, there has been growing concern in the research library profession about the ethical
implications of new AI-based tools. It has been observed that these tools are sometimes utilized in
ways that may compromise research integrity. This notion highlights the need for caution and ethical
considerations. Lack of trust and insecurity in how these new tools work underpin the necessity of
ethical codes for libraries. Existing AI guidelines and regulations need to be revised. International
library institutions, such as the IFLA and the LIBER, should lead their professional communities with
up-to-date norms. However, as it is necessary to consider regional differences, we encourage research
libraries to establish local AI strategies.</p>
      <p>Our findings also indicate a need to update library ethics and codes of conduct to meet the needs of
the age of AI. While the existing main principles align with general AI ethics, and the core mission to
provide open and fair access to information withstands, librarians should also consider independently
acting algorithms as new users of library services. Future generations will be the judges, beneficiaries,
or victims of today’s ethical decisions. Anticipating their needs is also an ethical consideration worth
doing.</p>
    </sec>
    <sec id="sec-9">
      <title>6. Acknowledgements</title>
      <p>We thank all the workshop participants who contributed to this research.
7. References
[19]M. Wojciechowska, Trust as a factor in building cognitive social capital among library workers
and users. Implications for library managers, The Journal of Academic Librarianship 47(1)
(2021) 102300. 10.1016/j.acalib.2020.102300.
[20]C. Wardle,, H. Derakhshan, Information disorder: Toward an interdisciplinary framework for
research and policy making, Council of Europe, Strasbourg, France, 2017, p. 109.
[21]M.C. Sullivan, Leveraging library trust to combat misinformation on social media, Library &amp;</p>
      <p>Information Science Research 41(1) (2019) 2–10. 10.1016/j.lisr.2019.02.004.
[22]High-Level Expert Group on Artificial Intelligence, Ethich Guidelines in Trustworthy AI,</p>
      <p>European Commission, Brussels, Belgium, 2019.
[23]OECD, Recommendation of the Council on OECD Legal Instruments Artificial Intelligence,
2022.
[24]C. Grbich, Qualitative data analysis : an introduction, SAGE Publications Ltd, London, 2013.
[25]A.F.T. Winfield,, M. Jirotka, Ethical governance is essential to building trust in robotics and
artificial intelligence systems, Philosophical Transactions of the Royal Society A: Mathematical,
Physical and Engineering Sciences 376(2133) (2018) 20180085. 10.1098/rsta.2018.0085.
[26]N. Diakopoulos, Accountability in algorithmic decision making, Commun. ACM 59(2) (2016)
56–62. 10.1145/2844110.
[27]A. Head,, B. Fister,, M. MacMillan, Information literacy in the age of algorithms, Project</p>
      <p>Information Literacy, 2020, p. 55.
[28]B. Johnson, Libraries in the Age of Artificial Intelligence, Computers in Libraries 38(1) (2018).
[29]UNSILO AI in Academic Publishing Survey 2019, Unsilo.ai, Aarhus, Denmark, 2019.
[30]M.L. Kennedy, What Do Artificial Intelligence (AI) and Ethics of AI Mean in the Context of</p>
      <p>Research Libraries?, Research Library Issues (299) (2019).
[31]United Nations, Report of the World Commission on Environment and Development: Our
Common Future, World Commission on Environment and Development, 1987.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] Future of Life Institute, Pause Giant AI Experiments: An Open Letter</article-title>
          . Available at: https://futureoflife.org/open-letter/
          <article-title>pause-giant-ai-experiments/</article-title>
          .
          <source>Accessed March</source>
          <volume>31</volume>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hansson</surname>
          </string-name>
          ,
          <article-title>Professional value and ethical self-regulation in the development of modern librarianship: The documentality of library ethics</article-title>
          ,
          <source>JD</source>
          <volume>73</volume>
          (
          <issue>6</issue>
          ) (
          <year>2017</year>
          )
          <fpage>1261</fpage>
          -
          <lpage>80</lpage>
          .
          <fpage>10</fpage>
          .1108/JD-02- 2017-0022.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <source>Technology Innovation and AI Ethics, Research Library Issues</source>
          <volume>299</volume>
          (
          <year>2019</year>
          )
          <fpage>14</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Alexander</surname>
          </string-name>
          ,,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ashford-Rowe</surname>
          </string-name>
          ,, N. Barajas-Murphy,, G. Dobbin,, J. Knott,,
          <string-name>
            <given-names>M.</given-names>
            <surname>McCormack</surname>
          </string-name>
          ,,
          <string-name>
            <surname>J. Pomerantz,</surname>
          </string-name>
          , R. Seilhamer,,
          <string-name>
            <given-names>N.</given-names>
            <surname>Weber</surname>
          </string-name>
          ,
          <source>Educause Horizon report: 2019 Higher Education edition.</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.C.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <source>Artificial intelligence in information retrieval systems</source>
          ,
          <source>Information Processing &amp; Management</source>
          <volume>12</volume>
          (
          <issue>3</issue>
          ) (
          <year>1976</year>
          )
          <fpage>189</fpage>
          -
          <lpage>222</lpage>
          .
          <fpage>10</fpage>
          .1016/
          <fpage>0306</fpage>
          -
          <lpage>4573</lpage>
          (
          <issue>76</issue>
          )
          <fpage>90005</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gasparini</surname>
          </string-name>
          ,, H. Kautonen,
          <source>Understanding Artificial Intelligence in Research Libraries - Extensive Literature Review, LIBER</source>
          <volume>32</volume>
          (
          <issue>1</issue>
          ) (
          <year>2022</year>
          ).
          <volume>10</volume>
          .53377/lq.10934.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Henry</surname>
          </string-name>
          ,
          <article-title>Research Librarians as Guides and Navigators for AI Policies at</article-title>
          Universities,
          <source>Research Library Issues</source>
          <volume>299</volume>
          (
          <year>2019</year>
          )
          <fpage>47</fpage>
          -
          <lpage>66</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.P.</given-names>
            <surname>Nayyer</surname>
          </string-name>
          ,,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rodriguez</surname>
          </string-name>
          ,
          <article-title>Ethical Implications of Implicit Bias in AI: Impact for Academic Libraries</article-title>
          , in: S. Hervieux,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Wheatley (Eds.),
          <article-title>The rise of AI: implications and applications of artificial intelligence in academic libraries</article-title>
          , Association of College and Research Libraries, Chicago,
          <year>2022</year>
          , pp.
          <fpage>165</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>I. International</surname>
          </string-name>
          <article-title>Federation of Library Associations and Institutions</article-title>
          ,
          <source>IFLA Code of Ethics for Librarians and other Information Workers (full version)</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.</given-names>
            <surname>Gummersson</surname>
          </string-name>
          ,
          <article-title>Qualitative methods in management research</article-title>
          , Sage, Newbury Park,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[11]UNESCO, Recommendation on the Ethics of Artificial Intelligence</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ovenden</surname>
          </string-name>
          ,
          <article-title>Burning the books : a history of knowledge under attack</article-title>
          , John Murray, London,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hardin</surname>
          </string-name>
          ,
          <source>The Street-Level Epistemology of Trust, Politics &amp; Society</source>
          <volume>21</volume>
          (
          <issue>4</issue>
          ) (
          <year>1993</year>
          )
          <fpage>505</fpage>
          -
          <lpage>29</lpage>
          .
          <fpage>10</fpage>
          .1177/0032329293021004006.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.R.</given-names>
            <surname>Wheeless</surname>
          </string-name>
          ,,
          <string-name>
            <surname>J. Grotz,</surname>
          </string-name>
          <article-title>The Measurement of Trust and Its Relationship to Self-Disclosure</article-title>
          ,
          <source>Human Communication Research</source>
          <volume>3</volume>
          (
          <issue>3</issue>
          ) (
          <year>1977</year>
          )
          <fpage>250</fpage>
          -
          <lpage>7</lpage>
          .
          <fpage>10</fpage>
          .1111/j.1468-
          <lpage>2958</lpage>
          .
          <year>1977</year>
          .tb00523.x.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.E.</given-names>
            <surname>Dayhoff</surname>
          </string-name>
          ,,
          <string-name>
            <surname>J.M. DeLeo</surname>
          </string-name>
          ,
          <source>Artificial neural networks, Cancer</source>
          <volume>91</volume>
          (
          <issue>S8</issue>
          ) (
          <year>2001</year>
          )
          <fpage>1615</fpage>
          -
          <lpage>35</lpage>
          .
          <fpage>10</fpage>
          .1002/
          <fpage>1097</fpage>
          -
          <lpage>0142</lpage>
          (
          <issue>20010415</issue>
          )91:
          <fpage>8</fpage>
          +&lt;1615:
          <article-title>:AID-CNCR1175&gt;3.0</article-title>
          .CO;2-L.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ridgeway</surname>
          </string-name>
          ,, D. Madigan,, T. Richardson,,
          <string-name>
            <surname>J. O'Kane</surname>
          </string-name>
          ,
          <article-title>Interpretable boosted naïve Bayes classification</article-title>
          ,
          <source>in: Proceedings of the Fourth International Conference on Knowledge Discovery and Data Mining</source>
          , AAAI Press, New York, NY,
          <year>1998</year>
          , pp.
          <fpage>101</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Z.C.</given-names>
            <surname>Lipton</surname>
          </string-name>
          ,
          <article-title>The mythos of model interpretability</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>61</volume>
          (
          <issue>10</issue>
          ) (
          <year>2018</year>
          )
          <fpage>36</fpage>
          -
          <lpage>43</lpage>
          .
          <fpage>10</fpage>
          .1145/3233231.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,, S. Singh,, C. Guestrin, “
          <article-title>Why Should I Trust You?”: Explaining the Predictions of Any Classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , Association for Computing Machinery, New York, NY, USA,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>44</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>