<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>User Interface for Advanced Multimodal Lifelog Querying</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Florian Ruosch</string-name>
          <email>ruosch@ifi.uzh.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luca Rossetto</string-name>
          <email>luca.rossetto@dcu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Multimodal Knowledge Graph, Graph-based Retrieval, Interactive Retrieval, Query Builder, User Interface</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dublin City University</institution>
          ,
          <addr-line>Dublin</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>FPR Consulting</institution>
          ,
          <addr-line>Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <fpage>2</fpage>
      <lpage>6</lpage>
      <abstract>
        <p>Lifelogging, the practice of recording parts of the subjective daily life, generates rich, multimodal data, but poses significant challenges for eficient retrieval. Building upon the based lifelog retrieval systems, this paper presents the fith iteration, which introduces a novel user interface to facilitate intuitive and powerful querying of lifelogs from multimodal knowledge graphs. We showcase how this frontend, powered by our custom MediaGraph store MeGraS, seamlessly exposes and leverages SPARQL capabilities. Through interactive demonstration scenarios, we illustrate how users can easily construct complex and expressive queries that also include advanced features such as similarity-based search, near-duplicate detection, and dynamic content extraction, all the while using native SPARQL syntax. This work highlights LifeGraph 5's user-centric design and MeGraS's role in bridging gaps between complex knowledge graph operations and accessible multimodal lifelog exploration.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Lifelogging [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is the practice of continuously capturing an individual’s subjective daily experiences
through various means of recording. Prominent data sources are wearable cameras to collect
firstperson-view images, to track the GPS location, sensors to record information such as heart rate, but also
lists of consumed media and their associated metadata. This generates vast and inherently multimodal
datasets, holding immense potential for memory augmentation and data-driven insights. However,
the sheer volume, diversity, and semi-structured nature of lifelog data pose significant challenges
for eficient and efective retrieval. Multimodal knowledge graphs have emerged as a powerful tool
for structuring such complex and interconnected information, enabling sophisticated querying and
semantic understanding: the LifeGraph series [
        <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2, 3, 4, 5</xref>
        ].
      </p>
      <p>
        We developed LifeGraph 5 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] for the 8th Lifelog Search Challenge (LSC’25) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], an annual competition
for docuemnt retrieval and question answering on a large multimodal dataset of lifelogs. It represents
the latest iteration of our system, which pushes for enhanced query capabilities by extending SPARQL
with concepts like implicit and derived relations. Having come a long way since the first demonstration
of LifeGraph [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], a core component of LifeGraph 5’s advancements is the newly designed user interface
and the underlying custom MediaGraph Store (MeGraS [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]), which enables access to the contents of the
documents in the multimodal knowledge graph.
      </p>
      <p>In this demo, we focus on the interactive aspects and the user experience. We aim to showcase how
our new frontend transforms complex querying using extended SPARQL into an intuitive and accessible
process for users. Through this, we will highlight how LifeGraph 5 enables the formulation of advanced
queries that leverage multimodal features and non-materialized relations, thereby unlocking deeper
insights into the lifelog data.</p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>2. LifeGraph 5</title>
      <p>This section first describes our custom multimodal knowledge graph store, MeGraS, then explains our
multimodal knowledge graph LifeGraph and, finally, its user interface.</p>
      <p>
        LifeGraph 5 represents a significant evolution from the foundational work of LifeGraph 1 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Our
development has been guided by the requirements of the Lifelog Search Challenges [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] (LSC), an annual
competition wherein participants aim to eficiently retrieve documents and answer questions from a
multimodal, large-scale lifelog dataset. It is comprised of over 700, 000 images taken from a first-person
point of view with a wearable camera, additional data from sensors, and rich metadata, presenting
a significant challenge for traditional retrieval systems. The specific tasks in the LSC range from
known-item search, where participants must find a specific item, to question answering about events or
objects, and complex ad-hoc searches, where as many images as possible should be retrieved that fit the
given description.
      </p>
      <p>While LifeGraph 1 established the foundation and the principle of using multimodal knowledge
graphs for lifelog retrieval, LifeGraph 5 introduces key architectural and user-facing advancements to
handle the multimodal nature of the LSC more efectively. The most significant innovations can be
summarized as follows:
• Custom Multimodal Knowledge Graph Store: Unlike its predecessors, which treated
multimedia as external, LifeGraph 5 is built upon a novel, custom-designed backend called the MediaGraph
Store (MeGraS). It elevates multimedia documents to “first-class citizens” within the knowledge
graph, allowing the query engine to directly access and process their content. This is a
fundamental architectural shift that is critical for meeting the demands of the LSC, where content-based
retrieval, such as finding specific objects or actions within an image, is paramount.
• Extended SPARQL Capabilities: MeGraS directly extends the SPARQL query language with
native support for advanced operations. This is vital for the LSC, as it enables complex functions
like k-nearest neighbor searches, which can find semantically similar images based on their vector
embeddings, and the ability to detect near-duplicates. These capabilities provide a powerful toolset
for participants to tackle the challenge’s retrieval tasks with greater precision and eficiency.
• Dynamic and Unmaterialized Relations: The system can handle derived and implicit relations
that are not necessarily materialized in the graph but computed at query time and persisted, if
necessary. This provides a more flexible data model for rapid exploration.
• Intuitive User Interface: LifeGraph 5 features a completely newly designed frontend that
transforms complex, SPARQL-based querying into an intuitive process for users. This
usercentric design is particularly beneficial in the time-constrained environment of the LSC, as it
allows participants to quickly build and refine queries using a series of selectors and filters. The
interface renders the SPARQL query in real-time, providing transparency into the underlying
logic without requiring the user to manually write code.</p>
      <sec id="sec-2-1">
        <title>2.1. MediaGraph Store</title>
        <p>
          MeGraS [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], our MediaGraph Store, serves as both the RDF graph store as well as the query engine
and aims to elevate multimedia documents to first-class citizens in knowledge graphs. Contrary to
traditional graph stores, it makes the linked documents available through assigned URIs and gives its
query engine access to their contents, allowing for advanced manipulation such as segmentation or
feature extraction. It also supports the Unified Multimedia Segmentation scheme [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] which allows the
direct addressing of arbitrary parts of these documents.
        </p>
        <p>Interactions with the graph happen through the RESTful API, but MeGraS also ofers a SPARQL
endpoint leveraging Apache Jena. Furthermore, it natively supports vector operations such as k-nearest
neighbor or similarity search, which is also available through SPARQL. MeGraS also handles derived
and implicit information, which is not necessarily materialized and can be computed at query time.</p>
        <p>Derived relations may be available in the graph ex ante, but are not required to be, and always have
a graph node as the subject and a literal as the object. They result from predefined functions, such as
computing the embedding of an image or extracting features such as visible text. If a requested derived
relation is not available at query time, MeGraS runs the functions and persists its output, avoiding the
need for repeated calculations.</p>
        <p>Implicit relations are defined to be always between two graph nodes, but never materialized, and are
inferred from other information that is available to the query engine. They are not persisted as they
may depend on other nodes in the graph and can change based on additions or removals. Examples
include k-nearest neighbor, spatial (e.g., behind, above), and temporal (e.g., during, after) relations.</p>
        <p>All these mechanisms allow for advanced and more expressive querying, purely through SPARQL.
Combined with the capability of handling tens of millions of triples through the integrated PostgreSQL
database, this makes MeGraS an optimal backend for our multimodal lifelog retrieval.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Ontology of LifeGraph</title>
        <p>
          The structure of the multimodal knowledge graph that contains the data of LifeGraph 5 is described
in a dedicated ontology.1 It serves as the formal schema for the image-related information within the
LSC dataset, and its core is the Image class representing individual images. For the object properties,
the ontology links images to instances of days for temporal facts and to tags to associate them with
descriptive keywords that do not carry any semantics. The data properties capture literal attributes. They
consist of the automatically extracted OCR (Optical Character Recognition), the running numbering,
the identifier, and additional annotations like the VLM-generated caption or the manually curated
category. Furthermore, they also include spatial references such as location name, city, and country.
Finally, we also employ a custom float vector datatype to precisely represent numerical vector data
(e.g., CLIP [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] image embeddings). Adherence to this ontological structure ensures that LifeGraph 5’s
data is consistent, well-defined, and semantically rich.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. User Interface</title>
        <p>This subsection explores LifeGraph 5’s frontend in detail. First, we focus on how the SPARQL query is
constructed. Then, we describe the results display and its advanced functionalities.
1https://github.com/MediaGraphOrg/LifeGraph5/blob/main/LifeGraph5.owl</p>
        <sec id="sec-2-3-1">
          <title>2.3.1. Query Builder</title>
          <p>Figure 1 shows an overview of the user interface and its components. The SPARQL query, created based
on the filters and selectors on the left, is rendered in real time in the dedicated area on the right. The
criteria represented in the query and their order can be configured for the user interface. By default,
the entire list is enabled, which can be divided into two classes: selectors (for tag, category, country,
city, and location) and filters (date, time, caption, OCR, and CLIP).</p>
          <p>The selectors allow for searching for and selecting a value, which is then used as the object for the
associated predicate in the query. Their design can be seen on the left in Figure 2: the search bar, the
scrollable list, and the selected objects. Tags are manually annotated labels for the images, based on
their content (e.g., bedroom, building). Categories, meanwhile, are associated with the type of location
of the image (e.g., airport terminal, bakery). Country, city, and location are all inferred from the location
metadata of the image and then mapped to entities in Wikidata based on the smallest distance.</p>
          <p>
            The filters can be distinguished into two types: simple data-based search and matching possibly
unmaterialized relations. The dates are all extracted from the metadata of the images and can be
searched either with a range, by year, month, or weekday. The time filter also uses a range of start and
end. The caption was generated using Vision-Language Models for LifeGraph 4 [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] and can be full-text
searched.
          </p>
          <p>Most of the filtering can be done by constructing sequences of simple triple patterns, based on
the selected criteria, of the following form: “?s lsc:predicate lsc:object .” This is shown in
the SPARQL query area in Figure 1, and the active criteria are indicated on the left with a green
dot. Furthermore, the query can also be specified to retrieve the dates for all results, allowing for
hierarchically grouping them by year, month, and day, for easier access.</p>
          <p>
            A more sophisticated query can be seen in Listing 1. Thereby, the CLIP [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ] embedding of a textual
description is computed and compared to the embedding of images using a custom SPARQL function.
The resulting set is then limited to the eight nearest images in the embedding space.
          </p>
          <p>This showcases two important aspects leveraging MeGraS: the derived relation clipEmbedding,
which may be unmaterialized and can be computed and persisted at query time, and the two custom
functions CLIP_TEXT (computing the CLIP embedding of text) and COSINE_SIM (calculating the cosine
distance of two vectors). Hence, we can achieve finding images matching a textual description based on
their distance in the CLIP embedding space using pure SPARQL. Similarly, the OCR relation does not
need to be precomputed but can be extracted at query time, if not yet available.</p>
          <p>
            Listing 1: Finding the eight images for which the CLIP [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ] embeddings are the most similar to the
embedding of the textual description.
          </p>
          <p>SELECT ?img
WHERE {</p>
          <p>BIND (megras:CLIP_TEXT("A man walking his dog on a rainy day.") as ?textVec)
?img derived:clipEmbedding ?clipVec .</p>
          <p>BIND (megras:COSINE_SIM(?textVec, ?clipVec) as ?cosSim)
}
ORDER BY DESC(?cosSim)
LIMIT 8</p>
        </sec>
        <sec id="sec-2-3-2">
          <title>2.3.2. Results Display and Interaction</title>
          <p>In Figure 1, the grid of results is visible, which allows for responsive and eficient exploration of the
query results. Clicking on the preview of an image brings up the result overlay, as shown in Figure 2.
The result overlay ofers functionality for the exploration of the results set by navigating with either
buttons or the arrow keys. Also, all triple relevant for the displayed image can be retrieved with the
“Show Infos” button in the bottom right. Clicking on a line in the table of predicates and objects refines
the query by adding the corresponding criterion, allowing for user feedback on the result set.</p>
          <p>Furthermore, the buttons in the bottom left can be used for similarity search. It ofers a
CLIPembedding-based k-nearest neighbor filter, whereby the k can be set dynamically (see Listing 2 for the
pure SPARQL query). Again, the CLIP embedding does not need to have been materialized ex ante but
can be computed and persisted at runtime with MeGraS. Likewise, a chosen number of near duplicates
can be retrieved, whereby the relation is implicit and not materialized in the graph.</p>
          <p>Listing 2: Detecting the implicit relation of k-nearest neighbors. In this example, k is equal to 5.
SELECT ?img
WHERE {</p>
          <p>lsc:imgID implicit:clip5nn ?img .
}</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Demonstration</title>
      <p>
        LifeGraph 5 is open source, free to download,2 and intended to be used together with MeGraS [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].3 The
latter can be compiled from source or downloaded as a Docker container.4 Furthermore, a demo video
is available that showcases LifeGraph 5’s capabilities.5
      </p>
      <p>
        During the demonstration, participants will have the opportunity to interact with LifeGraph 5. The
setting will be analogous to that during the evaluation of the Lifelog Search Challenge [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]: participants
will have to find images based on textual descriptions and answer questions about events in the lifelog
data. The user interface can be used to explore, browse, and retrieve the relevant images.
      </p>
      <p>Participants will engage in an interactive session demonstrating the system’s core features. They will
use the various selectors and filters on the left-hand side to construct a query. The real-time generation
of the corresponding SPARQL query will be visible throughout the process, providing a clear link
between the user’s actions and the underlying query logic. After execution, the results are displayed in
2https://github.com/MediaGraphOrg/LifeGraph5
3https://megras.org
4https://megras.org/docker
5https://megras.org/2025iswcdemo
a responsive grid. Participants can then select a specific item to bring up a detailed overlay showing
the image and its node neighborhood. This also provides the ability to refine the query with new
criteria based on the displayed information. The interactive process, combined with advanced features
like similarity search and near-duplicate detection, will highlight how LifeGraph 5 simplifies complex
multimodal lifelog exploration.</p>
      <sec id="sec-3-1">
        <title>Illustrative Case Study: Who did I go to dinner with? Imagine a user wants to find images from</title>
        <p>a specific event, recalling only a few details: “I went out to dinner in France in October 2019, but I
cannot remember who I was with. I think we had seafood and wine.” Now, how can the user find the
picture taken at that event to identify the person he was with?
Step 1 The user interacts with the user interface’s selectors and filters. They set the date filter to Year:
2019 and Month: October. They also select Country: France, and use the selectors for Tags
and Category, specifying wine and seafood restaurant, respectively.</p>
        <p>Step 2 As these criteria are selected, the LifeGraph 5 frontend constructs the corresponding SPARQL
query in real-time, rendering it in the dedicated query area. This live generation provides
transparency into the underlying query logic.</p>
        <p>Step 3 After executing the query, the system returns a grid of relevant images. The user can select a
specific image to look at to find instances of the event they were searching for. Selecting an item
brings it up in its full resolution and will allow the user to identify the person on it, answering
their question.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>
        In this paper, we presented LifeGraph 5 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a novel user interface designed to simplify and enhance the
retrieval of lifelog data from multimodal knowledge graphs. By seamlessly integrating with our powerful
MediaGraph Store (MeGraS [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]), it empowers users to perform sophisticated queries leveraging extended
SPARQL capabilities. We demonstrated how this intuitive interface allows for complex knowledge graph
operations such as similarity-based search, near-duplicate detection, and dynamic content extraction
from documents.
      </p>
      <p>
        LifeGraph 5 combined with MeGraS represents a step forward in making knowledge-graph-based
lifelog exploration more accessible and user-friendly. While this and previous iterations of LifeGraph
highlighted the potential of our approach, it also exhibited areas for further optimization, particularly
concerning query performance in extremely large datasets. Future work will focus on improving these
backend limitations and further enhancing the user interface, potentially exploring
natural-language-toSPARQL interfaces such as NLQxform [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. LifeGraph 5 and MeGraS aim to serve as foundational tools
for advancing research and practical applications in the complex and evolving landscape of multimodal
knowledge management and retrieval.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work was partially funded by the Swiss National Science Foundation through Project “MediaGraph”
(Grant Number 202125).</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Gemini and Grammarly for drafting content,
improving writing style, as well as grammar and spell check. After using these tools, the authors
reviewed and edited the content as needed, and they take full responsibility for the publication’s
content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Gurrin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Smeaton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Doherty</surname>
          </string-name>
          , Lifelogging:
          <article-title>Personal big data</article-title>
          ,
          <source>Found. Trends Inf. Retr</source>
          .
          <volume>8</volume>
          (
          <issue>2014</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>125</lpage>
          . doi:
          <volume>10</volume>
          .1561/1500000033.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baumgartner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ashena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pernischová</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <article-title>Lifegraph: A knowledge graph for lifelogs</article-title>
          ,
          <source>in: Proceedings of the Third Annual Workshop on Lifelog Search Challenge</source>
          , LSC '20,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2020</year>
          , p.
          <fpage>13</fpage>
          -
          <lpage>17</lpage>
          . doi:
          <volume>10</volume>
          .1145/3379172.3391717.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baumgartner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gasser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Heitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <article-title>Exploring graph-querying approaches in lifegraph</article-title>
          ,
          <source>in: Proceedings of the 4th Annual on Lifelog Search Challenge</source>
          , LSC '21,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , p.
          <fpage>7</fpage>
          -
          <lpage>10</lpage>
          . URL: https://doi.org/10. 1145/3463948.3469068.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Inel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          <article-title>, Multi-mode clustering for graphbased lifelog retrieval</article-title>
          ,
          <source>in: Proceedings of the 6th Annual ACM Lifelog Search Challenge, LSC</source>
          <year>2023</year>
          , Thessaloniki, Greece, June 12-15,
          <year>2023</year>
          , ACM,
          <year>2023</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>40</lpage>
          . doi:
          <volume>10</volume>
          .1145/3592573.3593102.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kyriakou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wardatzky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          , Lifegraph 4
          <article-title>- lifelog retrieval using multimodal knowledge graphs and vision-language models</article-title>
          ,
          <source>in: Proceedings of the 7th Annual ACM Workshop on the Lifelog Search Challenge</source>
          , LSC '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>88</fpage>
          -
          <lpage>92</lpage>
          . doi:
          <volume>10</volume>
          .1145/3643489.3661127.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          <article-title>SPARQL in the dark: Shining a light on multimodal lifelogs with LifeGraph 5</article-title>
          , in:
          <source>Proceedings of the 8th Annual ACM Workshop on the Lifelog Search Challenge (LSC '25)</source>
          ,
          <source>LSC '25</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .1145/3729459.3748694.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C.</given-names>
            <surname>Gurrin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Healy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Bailer</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.-T.</surname>
            Dang-Nguyen,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Hodges</surname>
            ,
            <given-names>B. Þór</given-names>
          </string-name>
          <string-name>
            <surname>Jónsson</surname>
            , M.-
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Tran</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Schöfmann</surname>
          </string-name>
          ,
          <article-title>Introduction to the 8th annual lifelog search challenge</article-title>
          ,
          <source>lsc'25, in: Proceedings of the 2025 International Conference on Multimedia Retrieval</source>
          , ICMR '25,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2025</year>
          , p.
          <fpage>2143</fpage>
          -
          <lpage>2144</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 3731715.3734579.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baumgartner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ashena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pernischová</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <article-title>A knowledge graph-based system for retrieval of lifelog data</article-title>
          , in: K. L.
          <string-name>
            <surname>Taylor</surname>
            , R. S. Gonçalves,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Lécué</surname>
          </string-name>
          , J. Yan (Eds.),
          <source>Proceedings of the ISWC</source>
          <year>2020</year>
          <article-title>Demos and Industry Tracks: From Novel Ideas to Industrial Practice co-located with 19th International Semantic Web Conference (ISWC</article-title>
          <year>2020</year>
          ),
          <article-title>Globally online</article-title>
          ,
          <source>November 1-6</source>
          ,
          <year>2020</year>
          (UTC), volume
          <volume>2721</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>223</fpage>
          -
          <lpage>228</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2721</volume>
          /paper557.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <article-title>MeGraS: An Open-Source Store for Multimodal Knowledge Graphs</article-title>
          ,
          <source>in: Proceedings of the 33rd ACM International Conference on Multimedia (MM '25)</source>
          , MM '25,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .1145/3746027.3756872.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Willi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          , L. Rossetto,
          <article-title>Unified multimedia segmentation - A comprehensive model for uri-based media segment representation</article-title>
          ,
          <source>TGDK</source>
          <volume>2</volume>
          (
          <year>2024</year>
          ) 1:
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          :
          <fpage>34</fpage>
          . doi:
          <volume>10</volume>
          .4230/TGDK.2.
          <issue>3</issue>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Krueger</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Learning transferable visual models from natural language supervision</article-title>
          , in: M.
          <string-name>
            <surname>Meila</surname>
          </string-name>
          , T. Zhang (Eds.),
          <source>Proceedings of the 38th International Conference on Machine Learning</source>
          ,
          <string-name>
            <surname>ICML</surname>
          </string-name>
          <year>2021</year>
          ,
          <volume>18</volume>
          -
          <issue>24</issue>
          <year>July 2021</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          , volume
          <volume>139</volume>
          <source>of Proceedings of Machine Learning Research, PMLR</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>8748</fpage>
          -
          <lpage>8763</lpage>
          . URL: http://proceedings.mlr.press/v139/ radford21a.html.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <article-title>Nlqxform: A language model-based question to SPARQL transformer</article-title>
          , in: D.
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Usbeck</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Mutharaju</surname>
          </string-name>
          , P. Kapanipathi (Eds.),
          <source>Joint Proceedings of Scholarly QALD 2023</source>
          and
          <article-title>SemREC 2023 co-located with 22nd</article-title>
          <source>International Semantic Web Conference ISWC</source>
          <year>2023</year>
          , Athens, Greece, November 6-
          <issue>10</issue>
          ,
          <year>2023</year>
          , volume
          <volume>3592</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2023</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3592</volume>
          /paper2.pdf.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>