<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Pisa, Italy
* Corresponding author.
" marco.lippi@unifi.it (M. Lippi); erik.longo@unifi.it (E. Longo); simone.marinai@unifi.it (S. Marinai);
giuseppe.mobilio@unifi.it (G. Mobilio)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>LLMs for Democratisation: Risks and Opportunities</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marco Lippi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Erik Longo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Simone Marinai</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giuseppe Mobilio</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Engineering, University of Florence</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Legal Sciences, University of Florence</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Generative artificial intelligence is bringing a deep revolution to many aspects of our society. Public authorities and representative assemblies are not an exception, as they will be profoundly concerned by this transformation. In this paper, we discuss opportunities and risks for the use of this novel technology in representative assemblies, showing how they can enable a process of democratization, while opening to possible threats and perils for the civil society.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large language models</kwd>
        <kwd>Public authorities</kwd>
        <kwd>Democratization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Opportunities</title>
      <p>
        In the last few years, generative Artificial Intelligence (AI) has produced a paradigm shift in the
technological world. The unprecedented success of this novel technology is mostly due to a straightforward
training procedure that allows to train very large neural networks grounding on the basic idea of
predicting the next word in a sentence. The supervision for this task can be automatically constructed
given a text of arbitrary length, and can therefore exploit huge data collections to implement this
selfsupervision technique [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. These pre-trained networks have shown surprising skills (named emergent
abilities) in many heterogeneous tasks: from document summarization to question answering, from
information retrieval to text comprehension and reasoning [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        The spread of this kind of models has enabled the development of many applications across diferent
domains, including the legal one. From a technological perspective, LLMs ofer evident opportunities
with respect to classic AI and machine learning methodologies. First of all, they allow even non-experts
to utilize this technology, since LLMs can be tested in a straightforward way via prompting techniques
formed directly in natural language, without requiring technical skills in computer science. Moreover,
the emergent abilities induced by pre-training permit to develop a wide range of applications without
the need to construct large training data sets – an activity that is usually very expensive in terms of
needed time and resources, and that requires expert knowledge for highly specific domains [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        In the context of democratic assemblies, from the legal point of view, this novel use of LLMs is bound
to bring several benefits. First of all, LLMs can enable greater democratic control: citizens can have easier
and more immediate access to oficial documents that express the political position of their
representatives. Another related key advantage would be that of reducing the distance between representatives and
represented individuals: LLMs can in fact lower the barrier of technical language that too often moves
citizens away from the law [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. By allowing the simplification of a large number of documents, LLMs
could ofer a valuable instrument to counter detachment and disafection from political participation. In
addition, generative AI could promote the right to information, by strengthening the right of citizens to
be informed about facts and issues of general interest, as well as promoting accountability, by fostering
citizens’ control over politicians by giving them access to simplified information.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Risks</title>
      <p>
        Clearly, the proliferation of this novel use of AI will also pose the problem of possible risks and threats to
society. From a legal perspective, a first peril comes from the possibility of undermining parliamentary
debate. In fact, generative AI tools may miss, and therefore reduce, language made up of subtleties,
nuances, and compromises [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        A consequence of this issue could be a weakening of democratic control. In Section 2 we have
highlighted how AI could strengthen democratic control. However, there is also a concrete risk that citizens
without technical expertise or lacking critical thinking, having little knowledge of how AI algorithms
work, cannot exercise control or verify the outcome, except through direct access to the produced
documents [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. As a result, AI algorithms may generate errors and fail to faithfully reproduce political
thinking without citizens being aware of it.
      </p>
      <p>Another potential threat to society is the spread of misinformation or disinformation that
inappropriate use of LLMs could encourage, also amplified by social media platforms, leading to polluting
public debate. In such an arena, traditional media risk seeing their intermediary role weakened. From a
technical point of view, this behaviour of LLMs could be the result of hallucinations, i.e., portions of
generated text where the content is completely invented, without any specific ground.</p>
      <p>This could pose a problem for the stability of democratic institutions, since the relationship of
conifdence between representative assemblies and executives depends on political judgment, which is
subject to the pressure of public opinion. An untruth, if widely and convincingly disseminated, can
have political consequences and undermine this relationship of confidence, the stability of institutions
and their ability to take decisions. This process could even lead to causing legal liability: when LLMs
contribute to the representation of a political actor or attribute an idea to a political force in an untruthful
way, they cause unlawful damage (image damage, defamation, etc.) to individuals and society, and thus
they could depict a scenario for legal liability.</p>
      <p>
        Finally, again from a computer science perspective, there is the potential risk that very large models,
that often show the best performance [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], will be in the hands of a few private companies, without the
possibility, for the civil society, to have any guarantee of transparency or accountability.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <p>The opportunities and risks outlined above can be addressed by applying certain regulatory requirements
and implementing technical solutions to maximise the benefits of using these technologies. At the legal
level, the most recent innovation is the Regulation (EU) 2024/1689 “laying down harmonised rules on
artificial intelligence” (so called AI Act), adopted in June 2024 and not yet fully in force.</p>
      <p>
        One way to increase the reliability and credibility of content produced by LLMs is to use disclaimers
visible on the masks for questioning the LLMs or watermarks in the generated content. The aim is
to make it clear that the information is produced by an LLM and that political statements are not
reproduced exactly. On a legal level, the AI Act imposes conditions in this regard by requiring providers
of AI systems, including general-purpose AI systems, to mark output “in a machine-readable format
and detectable as artificially generated or manipulated” (Art. 50 of AI Act). This is also an active area of
research from a computer science point of view [
        <xref ref-type="bibr" rid="ref11">11, 12</xref>
        ].
      </p>
      <p>Another solution is to cite the sources from which LLMs derive the processed or summarised
information. This allows citizens to check whether a political position is being distorted or whether its
meaning is being faithfully reproduced. Some provisions of the AI Act go in this direction by classifying
LLMs used for this purpose as “high risk”, because they are “intended to be used for influencing the
outcome of an election or referendum or the voting behaviour of natural persons in the exercise of their
vote in elections or referenda” (Annex III of the AI Act). As a result, Article 14 imposes obligations of
“human oversight”, which allows, for example, the output of the AI system to be correctly interpreted
in order to prevent or minimise risks to fundamental rights, such as those related to information and
political participation. From this point of view, we cannot hide the risks of awareness, i.e., of uncritical
reliance on AI systems and the temptation to always trust the results ofered by the technologies.</p>
      <p>The AI Act also imposes general transparency requirements on the functioning of algorithms, so that
their reliability can be assessed. Art. 14 requires AI systems to be “suficiently transparent to enable
deployers to interpret a system’s output and use it appropriately”. In this sense, the instructions for the
use of the LLM should include “the characteristics, capabilities and limitations of performance”, and, for
example, “where applicable, the technical capabilities and characteristics of the high-risk AI system to
provide information that is relevant to explain its output”. This calls for innovative AI methods that are
either interpretable-by-design, or that explain the predictions and decisions of a “black-box” model,
following the paradigm of eXplainable AI (XAI) [13].</p>
      <p>From a technical point of view, there are a number of solutions that can be used to improve the
reliability of the information produced, also with the aim of avoiding hallucinations, citing sources and
enhancing the transparency of LLMs. Retrieval-Augmented Generation (RAG) is one of the most widely
employed techniques developed with this goal [14]. RAG consists in a system architecture designed to
incorporate an external knowledge base (such as a database, or a memory) within the LLM architecture.
By leveraging this additional amount of knowledge, prompts can be enhanced with more information
related to the context. This mechanism is shown to greatly reduce hallucinations, and to improve the
reliability and accountability of LLMs with respect to users, as they become able to cite the sources
that support their claims and arguments [15]. In the case of representative assemblies, the external
knowledge base could contain selected legal documents, such as motions or parliamentary questions,
classified according to certain parameters, such as type, date of approval, context of discussion, which
increase reliability. The RAG approach is often coupled with specific categories of prompts, such as
the Chain-of-Thought [16], where the LLM is specifically asked to produce an answer that contains a
step-by-step illustration of the reasoning path leading to the output. Recent models are also specifically
designed to perform reasoning when answering a given query [17]. This aspect could strongly improve
the interpretability of answers belonging to a legal context for citizens.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>Generative AI tools will afect the way in which citizens perceive the actions of public authorities and
representative assemblies. The proliferation of these tools will also create a new type of intermediary,
diferent from traditional media professionals or outlets. The management and verification of the
information produced by LLMs will require not only information experts, but also IT experts who can
verify the reliability of the machines that process the information. At the same time, there is a growing
need for legislation to clarify the legal responsibilities arising from the use of LLMs. This need arises
both from the need to defend the image of individuals, such as parliamentarians, and from the need to
ensure the reliability of information and the stability of democratic systems. This is the challenge that
the above-mentioned AI Act, which in many parts has yet to come into force, seeks to address.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This research has been partially funded by CAI4DSA actions (Collaborative Explainable neuro-symbolic
AI for Decision Support Assistant), of the FAIR national project on artificial intelligence, PE 1 PNRR
(https://fondazione-fair.it/).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[12] R. Zhang, S. S. Hussain, P. Neekhara, F. Koushanfar, {REMARK-LLM}: A robust and eficient
watermarking framework for generative large language models, in: 33rd USENIX Security Symposium
(USENIX Security 24), 2024, pp. 1813–1830.
[13] D. Minh, H. X. Wang, Y. F. Li, T. N. Nguyen, Explainable artificial intelligence: a comprehensive
review, Artificial Intelligence Review (2022) 1–66.
[14] P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, N. Goyal, H. Küttler, M. Lewis, W.-t. Yih,
T. Rocktäschel, et al., Retrieval-augmented generation for knowledge-intensive nlp tasks, Advances
in neural information processing systems 33 (2020) 9459–9474.
[15] K. Shuster, S. Pof, M. Chen, D. Kiela, J. Weston, Retrieval augmentation reduces hallucination in
conversation, arXiv preprint arXiv:2104.07567 (2021).
[16] J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou, et al., Chain-of-thought
prompting elicits reasoning in large language models, Advances in neural information processing
systems 35 (2022) 24824–24837.
[17] D. Guo, D. Yang, H. Zhang, J. Song, R. Zhang, R. Xu, Q. Zhu, S. Ma, P. Wang, X. Bi, et al.,
Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning, arXiv preprint
arXiv:2501.12948 (2025).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Inter-Parliamentary</surname>
            <given-names>Union</given-names>
          </string-name>
          ,
          <source>World e-parliament report</source>
          ,
          <year>2024</year>
          . URL: https://www.ipu.org/resources/ publications/reports/2024-10/world-e-parliament
          <source>-report-2024.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Pesch</surname>
          </string-name>
          ,
          <article-title>Potentials and challenges of large language models (LLMs) in the context of administrative decision-making</article-title>
          ,
          <source>European Journal of Risk Regulation</source>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Weidinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mellor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rauh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Grifin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uesato</surname>
          </string-name>
          , P.-S. Huang, M. Cheng, M. Glaese,
          <string-name>
            <given-names>B.</given-names>
            <surname>Balle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kasirzadeh</surname>
          </string-name>
          , et al.,
          <article-title>Ethical and social risks of harm from language models</article-title>
          ,
          <source>arXiv preprint arXiv:2112.04359</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Weidinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uesato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rauh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Grifin</surname>
          </string-name>
          , P.-S. Huang,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mellor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Glaese</surname>
          </string-name>
          , M. Cheng, B.
          <string-name>
            <surname>Balle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Kasirzadeh</surname>
          </string-name>
          , et al.,
          <article-title>Taxonomy of risks posed by language models</article-title>
          ,
          <source>in: Proceedings of the 2022 ACM conference on fairness, accountability, and transparency</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>214</fpage>
          -
          <lpage>229</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Min</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dong</surname>
          </string-name>
          , et al.,
          <article-title>A survey of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2303.18223 1</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bommasani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zoph</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Borgeaud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yogatama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Metzler</surname>
          </string-name>
          , et al.,
          <article-title>Emergent abilities of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2206.07682</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Breton</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Billami</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chevalier</surname>
            ,
            <given-names>H. T.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Satoh</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Trojahn</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Zin</surname>
          </string-name>
          ,
          <article-title>Leveraging LLMs for legal terms extraction with limited annotated data</article-title>
          ,
          <source>Artificial Intelligence and Law</source>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pałka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lagioia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Liepina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lippi</surname>
          </string-name>
          , G. Sartor,
          <article-title>Make privacy policies longer and appoint LLM readers</article-title>
          ,
          <source>Artificial Intelligence and Law</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Palonen</surname>
          </string-name>
          ,
          <article-title>Making sense of parliamentary jargon: The need for parliamentary literacy</article-title>
          , Rhetoric, Politics and Conceptual
          <string-name>
            <surname>Change</surname>
          </string-name>
          (
          <year>2011</year>
          )
          <fpage>13</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gerlich</surname>
          </string-name>
          ,
          <article-title>AI tools in society: Impacts on cognitive ofloading and the future of critical thinking</article-title>
          ,
          <source>Societies</source>
          <volume>15</volume>
          (
          <year>2025</year>
          )
          <article-title>6</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <surname>Attacking</surname>
          </string-name>
          <article-title>LLM watermarks by exploiting their strengths</article-title>
          ,
          <source>in: ICLR 2024 Workshop on Secure and Trustworthy Large Language Models</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>