=Paper= {{Paper |id=Vol-3828/ISWC2024_paper_38 |storemode=property |title=Here’s Charlie! Realising the Semantic Web vision of Agents in the age of LLMs |pdfUrl=https://ceur-ws.org/Vol-3828/paper38.pdf |volume=Vol-3828 |authors=Jesse Wright |dblpUrl=https://dblp.org/rec/conf/semweb/Wright24a }} ==Here’s Charlie! Realising the Semantic Web vision of Agents in the age of LLMs== https://ceur-ws.org/Vol-3828/paper38.pdf
                                Here’s Charlie! Realising the Semantic Web vision of
                                Agents in the age of LLMs
                                Wright, Jesse1
                                1
                                    Computer Science Department, University of Oxford, UK


                                              Abstract
                                              This paper presents our research towards a near-term future in which legal entities, such as individuals
                                              and organisations can entrust semi-autonomous AI-driven agents to carry out online interactions on
                                              their behalf. The author’s research concerns the development of semi-autonomous Web agents, which
                                              consult users if and only if the system does not have sufficient context or confidence to proceed working
                                              autonomously. This creates a user-agent dialogue that allows the user to teach the agent about the
                                              information sources they trust, their data-sharing preferences, and their decision-making preferences.
                                              Ultimately, this enables the user to maximise control over their data and decisions while retaining the
                                              convenience of using agents, including those driven by LLMs.
                                                  In view of developing near-term solutions, the research seeks to answer the question: “How do we
                                              build a trustworthy and reliable network of semi-autonomous agents which represent individuals and
                                              organisations on the Web?”. After identifying key requirements, the paper presents a demo for a sample
                                              use case of a generic personal assistant. This is implemented using (Notation3) rules to enforce safety
                                              guarantees around belief, data sharing and data usage and LLMs to allow natural language interaction
                                              with users and serendipitous dialogues between software agents.

                                              Keywords
                                              Agent, Dialogue, LLM, Data Privacy, Trust, Semantic Web, Solid Reasoner, Inference, RDF, N3, Notation3,
                                              RDF Surfaces, Semantic Web, Proof, Proof Engine, Solid




                                1. Introduction
                                There exists a substantial body of research on communication protocols for multi-agent systems,
                                and it is reflected in the vision of the Semantic Web itself [1, 2, 3] as shown by Charlie, the “AI that
                                works for you”. Yet, the 2006 lamentation that “[b]ecause we haven’t yet delivered large-scale,
                                agent-based mediation, some commentators argue that the Semantic Web has failed” [4] still
                                rings true today. The growing use of LLMs raises a key challenge in building Trustworthy and
                                Reliable Web Agents [5, 6]. This is heightened by growing interest among LLM researchers in
                                building dialogues between multiple LLMs [7, 8]. Moreover, recent research indicates the strong
                                potential of the Semantic Web to complement emerging LLM technologies [9]. For example,
                                the use of Retrieval Augmented Generation (RAG) with Knowledge Graphs has shown to be
                                effective in grounding LLM queries [10]. The universal semantics and proof mechanisms of the
                                Semantic Web stack are therefore pertinent to the successful development of semi-autonomous
                                Web agents using LLMs.
                                Posters, Demos, and Industry Tracks at ISWC 2024, November 13–15, 2024, Baltimore, USA
                                Envelope-Open jesse.wright@cs.ox.ac.uk (W. Jesse)
                                GLOBE https://www.cs.ox.ac.uk/people/jesse.wright/ (W. Jesse)
                                Orcid 0000-0002-5771-988X (W. Jesse)
                                            © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).




CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
2. Design Requirements
 We identify the following non-functional requirements for an agent communication protocol.
It must be possible for semi-autonomous agents to:
   1. Identify legal entities, such as individuals or organisations, on the Web [11] so they can
      be referenced.
   2. Deterministically discover other agents representing an entity from their Web identity [11].
      This does not require all agents to be publicly advertised; some may be discovered from
      links to protected documents.
   3. Describe, and agree to, any usage controls [12, 13, 14] associated with data they exchange.
      This allows sharing of protected data while articulating the recipient’s legal or moral
      obligations [15].
   4. Describe the origin and provenance of data they exchange. In an open world of agents
      that can “say anything about anything,” systems can identify which external claims to
      believe for a given task, based on the agent’s internal trust model.
   5. Unambiguously describe ground truths they send, and agreements they make, using a
      formal representation. Consider the case where an individual’s agent purchases a flight
      from an airline’s agent. Structured ground truths eliminate an LLM’s risk of hallucination
      or misinterpretation of key information, such as the flight time (“10 o’clock” could be 22:00
      or 10:00). As agents represent entities in binding agreements, this approach also reduces
      the risk of legal disputes by limiting the subjectivity of agreed terms and thus the ability to
      reinterpret or rescind them [16]. Furthermore, agents can implement rule-based internal
      safeguards, such as user-defined daily spending limits. Truly generic agents may generate
      and communicate structured ontologies when encountering new tasks. In many cases
      we expect LLM-supported ontology construction [17] to facilitate generation; however,
      research is required to understand how (1) agents can align on conceptual models for use
      and (2) how human oversight can be maintained without disrupting user experience.
   6. Contextualise a task which may be ambiguous or poorly defined, such that interacting
      agents can introduce new solution spaces or negotiating actors in a serendipitous manner.


3. Sample Use-Case and Implementation

  We implemented the following flow where agents act as personal assistants
for individual users:
   1. Jun types into a chat “Schedule a meeting with Nigel next week”;
   2. Jun’s agent identifies data to be shared with Nigel and requests relevant
      sharing permissions from Jun (where not already obtained);
   3. Nigel’s agent receives a request from Jun;
   4. Nigel is prompted to confirm that he believes Jun is an authoritative source
      of truth for her calendar (where not already obtained);
   5. Nigels agent proposes a meeting time to Nigel; and
   6. the meeting is proposed to Jun’s agent and automatically confirmed.
Figure 1: Flow diagram for the scheduling use case. Alice is Jun’s agent and Bob is Nigel’s agent.


  We have created a running demo with a video, flow-diagrams (including Figure 1) and other
resources for our codebase1 . The implementation corresponds to the above use-case steps:

      1. Given the user prompt and a set of known WebID profiles [11], an LLM called by Jun’s
         agent identifies the relevant entities for the agent to negotiate with (Nigel), and the
         WebIDs of those entities. Given the user prompt, and the user’s personal knowledge
         graph, an LLM called by Jun’s agent identifies which subset (as a list of named graphs) of
         the user data are needed to fulfil the user’s request.
      2. Notation3 [18] reasoning is used to identify the policies applicable to the data subset. In
         the available demo recording, policies are encoded in ACP [12]; we are currently migrating
         to use ODRL [13] and DPV [14]. If these policies do not yet permit read access to Nigel,
         Jun is prompted to modify them. Jun’s agent then dereferences Nigel’s WebID [11] to
         discover information about his agent.
      3. Jun’s agent uses an LLM to construct a message for Nigel’s agent, explaining the context
         of Jun’s task: “Jun seeks to schedule a meeting for next week. Propose a time for Jun and
         Nigel to meet using their calendars.” Jun’s agent sends Nigel’s agent this message along
         with the RDF description of Jun’s calendar and any associated policies and provenance.
         With ACL, Nigel’s agent does not need to agree to any policy obligations; this changes
         with ODRL. The provenance in this case is simply a signature of the canonicalised calendar
         dataset [19] using Jun’s public key.
1
    https://github.com/jeswr/phd-language-dialogue-experiment
   4. As Nigel has instructed his agent that Jun is an authoritative source of information on
      all topics, his agent believes (takes as ground truth) the signed RDF dataset sent by her
      agent. We are developing conceptual models for agentic trust; these extend existing trust
      vocabularies [20, 21, 22, 23] with a range of features including (1) qualifying whether
      sources are trusted for particular types of claims; for instance, most agents should trust
      certified airlines to present flight times and prices, but not medical data (2) qualifying
      the forms of provenance secure enough for a given task; for instance, an insurance
      provider may require provenance demonstrating a user was signed in with two-factor
      authentication when entering financial details to their knowledge base.
   5. Nigel’s agent proposes a meeting time, using the natural language context (not a ground
      truth) and the calendar dataset (ground truth). The LLM proposes a meeting time, then
      the N3 reasoner applies rules to (1) ensure no calendar conflicts and (2) check for user
      confirmation, before adding the proposed time to the knowledge base. In a future iteration,
      we plan to use the LLM to generate an N3 query that proposes a meeting time based on
      Nigel’s Personal Data Store and Jun’s calendar.
   6. Upon meeting the above requirements, the reasoner sends to Jun’s agent a meeting
      proposal, in the form of an RDF dataset with attached usage policies and provenance.
      Jun’s agent confirms this dataset can be believed based on the internal trust model. The
      rules within Jun’s agent validate that there are no conflicting events. Jun’s personal
      knowledge base is updated with the event, and a confirmation is sent to Nigel’s agent.


4. Conclusion and Future Research
We have implemented a generic personal assistant that communicates using a protocol satisfying
the requirements of Section 2. Future work will make the design requirements more rigorous
by (1) gathering requirements for personal agents through user studies, and (2) engaging with
industry to develop specialised agents, including product sales agents. Concurrently, we shall
formalise the vocabularies for exchanging provenance and terms of use between agents and
modelling trust and data policies within agents, extending those vocabularies discussed in
Section 3. Once these vocabularies mature, we will develop reasoning specifications to mediate
between the internal representations and exchanged metadata. This enables agents to negotiate
to obtain sufficient provenance to believe claims, and find agreeable data terms of use between
agents - whilst concurrently updating their internal models via user interaction.


Acknowledgements
Jesse Wright is funded by the Department of Computer Science, University of Oxford.
References
 [1] O. Lassila, J. Hendler, T. Berners-Lee, The semantic web, Scientific American 284 (2001)
     34–43.
 [2] S. Luke, L. Spector, D. Rager, J. Hendler, Ontology-based web agents, in: Proceedings of
     the first international conference on Autonomous agents, 1997, pp. 59–66.
 [3] S. Poslad, Specifying protocols for multi-agent systems interaction, ACM Transactions on
     Autonomous and Adaptive Systems (TAAS) 2 (2007) 15–es.
 [4] N. Shadbolt, T. Berners-Lee, W. Hall, The semantic web revisited, IEEE Intelligent Systems
     21 (2006) 96–101. doi:10.1109/MIS.2006.62 .
 [5] Y. Deng, A. Zhang, Y. Lin, X. Chen, J.-R. Wen, T.-S. Chua, Large language model powered
     agents in the web, learning 2 (2024) 20.
 [6] L. Sun, Y. Huang, H. Wang, S. Wu, Q. Zhang, C. Gao, Y. Huang, W. Lyu, Y. Zhang, X. Li,
     et al., Trustllm: Trustworthiness in large language models, arXiv preprint arXiv:2401.05561
     (2024).
 [7] Q. Wu, G. Bansal, J. Zhang, Y. Wu, S. Zhang, E. Zhu, B. Li, L. Jiang, X. Zhang, C. Wang,
     Autogen: Enabling next-gen llm applications via multi-agent conversation framework,
     arXiv preprint arXiv:2308.08155 (2023).
 [8] Y. Deng, W. Zhang, W. Lam, S.-K. Ng, T.-S. Chua, Plug-and-play policy planner for large
     language model powered dialogue agents, in: The Twelfth International Conference on
     Learning Representations, 2023.
 [9] J. Wright, The old and the new - using semantic web technologies to build better AI, 2024.
     URL: https://blog.jeswr.org/2024/04/18/better-ai.
[10] M. Kang, J. M. Kwak, J. Baek, S. J. Hwang, Knowledge graph-augmented language models
     for knowledge-grounded dialogue generation, arXiv preprint arXiv:2305.18846 (2023).
[11] H. Story, T. Berners-Lee, A. Sambra, R. Taelman, J. Scazzosi, Web Identity (WebID) 1.0, W3C
     Community Group Final Report, W3C, 2024. https://w3c.github.io/WebID/spec/identity/.
[12] M. Bosquet, Access Control Policy (ACP), Solid Editor’s Draft, W3C, 2022. https://w3c.
     github.io/WebID/spec/identity/.
[13] R. Iannella, S. Villata, Odrl information model 2.2, 2023. URL: https://www.w3.org/TR/
     2018/REC-odrl-model-20180215/.
[14] H. J. Pandit, A. Polleres, B. Bos, R. Brennan, B. Bruegger, F. J. Ekaputra, J. D. Fernández,
     R. G. Hamed, E. Kiesling, M. Lizar, et al., Creating a vocabulary for data privacy: The
     first-year report of data privacy vocabularies and controls community group (dpvcg), in:
     On the Move to Meaningful Internet Systems: OTM 2019 Conferences: Confederated
     International Conferences: CoopIS, ODBASE, C&TC 2019, Rhodes, Greece, October 21–25,
     2019, Proceedings, Springer, 2019, pp. 714–730.
[15] J. Wright, B. Esteves, R. Zhao, Me want cookie! towards automated and transparent data
     governance on the web, 2024. URL: https://arxiv.org/abs/2408.09071. arXiv:2408.09071 .
[16] M. Garcia, What Air Canada Lost In ‘Remarkable’ Lying AI Chat-
     bot             Case,              https://www.forbes.com/sites/marisagarcia/2024/02/19/
     what-air-canada-lost-in-remarkable-lying-ai-chatbot-case/, 2024. [Accessed 05-07-
     2024].
[17] V. K. Kommineni, B. König-Ries, S. Samuel, From human experts to machines: An llm
     supported approach to ontology and knowledge graph construction, 2024. URL: https:
     //arxiv.org/abs/2403.08345. arXiv:2403.08345 .
[18] T. Berners-Lee, Notation3, http://www.w3.org/DesignIssues/Notation3.html (1998).
[19] D. Longley, G. Kellogg, D. Yamamoto, M. Sporny, Access Control Policy (ACP), Solid
     Editor’s Draft, W3C, 2022. https://w3c.github.io/WebID/spec/identity/.
[20] M. Richardson, R. Agrawal, P. Domingos, Trust management for the semantic web, in:
     International semantic Web conference, Springer, 2003, pp. 351–368.
[21] S. Galizia, Wsto: A classification-based ontology for managing trust in semantic web
     services, in: European semantic web conference, Springer, 2006, pp. 697–711.
[22] W. Sherchan, S. Nepal, J. Hunklinger, A. Bouguettaya, A trust ontology for semantic
     services, in: 2010 IEEE International Conference on Services Computing, IEEE, 2010, pp.
     313–320.
[23] G. Amaral, T. P. Sales, G. Guizzardi, D. Porello, Towards a reference ontology of trust,
     in: On the Move to Meaningful Internet Systems: OTM 2019 Conferences: Confederated
     International Conferences: CoopIS, ODBASE, C&TC 2019, Rhodes, Greece, October 21–25,
     2019, Proceedings, Springer, 2019, pp. 3–21.