<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Knowledge Graph-based System for Retrieval of Lifelog Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Luca Rossetto</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matthias Baumgartner</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Narges Ashena</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Florian Ruosch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Romana Pernisch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abraham Bernstein</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Informatics, University of Zurich</institution>
          ,
          <addr-line>Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Lifelogging is a phenomenon where practitioners record an increasing part of their subjective daily experience with the aim of later being able to use these recordings as a memory aid or basis for datadriven self improvement. The resulting lifelogs are, therefore, only useful if the lifeloggers have efficient ways to search through them. The logs are inherently multi-modal and semi structured, combining data from several sources, such as cameras and other wearable physical as well as virtual sensors, so representing the data in a graph structure can effectively capture all produced interrelations. Since annotating each entry with a sufficiently large semantic context is infeasible, either manually or automatically, alternatives must be found to capture the higher level semantics. In this paper, we demonstrate LifeGraph, a first approach of creating a Knowledge Graph-based lifelog representation and retrieval solution, able of capturing a lifelog in a graph structure and augmenting it with external information to aid with the association of higher-level semantic information.</p>
      </abstract>
      <kwd-group>
        <kwd>Lifelogging</kwd>
        <kwd>Lifelog Retrieval</kwd>
        <kwd>Multi-modal Retrieval</kwd>
        <kwd>Multimodal Graphs</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Various technological advances over the last few decades have led to a rapid
increase in possibilities for mobile sensing, processing, and storing of various forms
of data. One of the many consequences of these possibilities was the emergence of
the quantified-self movement as well as the Lifelogging phenomenon. Lifelogging
describes the act of capturing aspects of one’s personal experience using
cameras to continuously record one’s point of view, sensors to quantify bio-feedback,
or properties of one’s immediate surroundings, purely software-based means to
analyze one’s interaction with the digital environment as well as more
traditional methods, such as entries in a personal diary. Thus, the resulting lifelog is
Copyright c 2020 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
inherently multi-modal and can exhibit a high degree of interconnectivity. An
appropriate, but currently under-explored approach, is, therefore, to represent a
lifelog as a knowledge graph.</p>
      <p>
        Lifelogs are commonly used as a memory aid or to monitor one’s own life
in order to improve or maintain health. In order for these logs to be useful,
it is paramount that lifeloggers have the possibility to efficiently search within
the logs which becomes increasingly difficult due to the growing volume and
diversity of available data. To catalyze research in the area of lifelog retrieval,
a competitive benchmark called the Lifelog Search Challenge (LSC) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] was
established in 2018 as an annual workshop to evaluate interactive lifelog retrieval
systems. In this paper, we demonstrate an interactive knowledge graph-based
lifelog retrieval system called LifeGraph [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which was first introduced in the
context of the 2020 edition of the LSC. The following provides an overview of
how the graph was constructed, how the system uses it to answer queries, and
how users can expect to be able to interact with the system itself.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Constructing LifeGraph</title>
      <p>
        The basis for LifeGraph is formed by the lifelog dataset made available in the
context of LSC 2020, which consists of 114 days of lifelogging data made up of
roughly 200k images taken by a wearable camera with accompanying metadata
as well as various sensor data. Ideally, each entry in a lifelog, independent of
its type and method of representation, would have rich semantic annotations
describing not only its content but also placing it into a larger context. Since
generating such annotations manually is infeasible and doing so automatically
is beyond what is currently possible, alternatives need to be used to make the
collected data searchable and, hence, usable. To overcome this limitation, we
construct a graph structure which captures the relations between higher level
semantic concepts and lower level labels, such as the type of environment or
the presence of everyday objects within an image, both of which can be reliably
detected with currently available technology [
        <xref ref-type="bibr" rid="ref4 ref9">9,4</xref>
        ].
      </p>
      <p>
        A challenge of constructing LifeGraph is to link high-level search entities
with low-level image features. We use COEL (Classification of Everyday
Living) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] as a starting point which provides a taxonomy of everyday actions,
such as housework, sports, or food related activities. In order to combine COEL
with detectable entities, we defined a mapping to Wikidata [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. It was created
semi-automatically, based on the similarity to Wikidata entity labels, whereas
ambiguities were resolved manually.
      </p>
      <p>To connect the mapped concepts to lower-level semantic image labels we
combined them with additional data from Wikidata. For this, we created a graph
initially consisting of the Wikidata entities mapped to COEL as well as all
detectable entities. We then successively retrieved all entities from wikidata that
can be connected to at least two entities in the present graph, and added them
to the graph. This procedure is repeated three times, and two additional times
considering only class relations (:instance-of, :subclassOf). Finally, we
discarded disconnected components of the graph since they failed to link image
labels to high-level concepts. We further removed entities intended for
navigational or internal purpose, such as disambiguation or template pages.</p>
      <p>
        We modeled the metadata which was already part of the dataset provided
for LSC in a simple and flat way following the schema explained in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Each
metadata entry is associated with an image based on time. Using either an object
or datatype property, we link all the metadata information with the image in
the graph. Contrary to what was stated in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], each image is linked to a day
resource. This enables us to link day-specific information directly to the day
resource instead of repeating this information for all ca. 2000 images of that day.
Before populating the schema, we interpolated some aspects of the metadata such
as the provided semantic location labels (“work” vs. “home”) but also detected
objects [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and places [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. If the same concept is interrupted by a sufficiently
small gap, we add associations to this concept to all entries in the gap.
      </p>
      <p>To associate images with objects and places, we then directly establish a link
between images and Wikidata entries using either the :detected or
:detected_place property.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Exploring LifeGraph</title>
      <p>
        The browser-based LifeGraph user interface is a modified version of vitrivr-ng [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
– the UI of the vitrivr [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] content-based multimedia retrieval stack – which has
already been successfully used for effective retrieval of lifelog data [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Apart from
the user interface, LifeGraph does not share any logic with the vitrivr stack. For
the sake of simplicity, a user can specify queries by selecting nodes from the
graph, representing the concepts in which they are interested, using a simple
auto-completing text box. This query formulation scheme offers a reasonable
compromise between specificity and efficiency while not requiring any technical
expertise on the user side. Once specified, a query is sent to a back-end which
performs its execution and the subsequent scoring of the results. The query is
evaluated by traversing the graph, searching for paths from each of the specified
start nodes to all log entries which could be part of the result. The maximum
length of this path is continuously increased until an empirically determined
threshold is reached, either in path length or number of potential results. Each
log entry found in this way is scored by the sum of the inverse of the path lengths
leading to it, normalized by the number of start points in the query. This means
the shorter the distance to the highest number of start points, the higher the score
of the retrieved item. These aggregated results are then returned to the UI, where
they are presented as either a list of log entries sorted by score or a list of entries
grouped by day, sorted by time within each. Since the user is supposed to be able
to browse through the retrieved results in order to identify the desired elements,
the inevitable false-positives produced by some of the underlying detectors are
less of a problem than missing detections or false-negatives, since a large number
of false-positives will cause users to spend an unnecessarily large amount of
time browsing, while too many false-negatives could prevent them from finding
the desired result at all. All log entries are accompanied by some metadata,
including technical information, like its originating time and date. Additionally,
entries include lifelog specific semantic information, such as manual semantic
annotations or bio feedback properties, e.g., heart-rate, number of steps taken
within a specific interval, etc. The UI uses these metadata to provide additional
filtering options to a user, excluding entries from the result if they do not match
Boolean criteria relevant to the query. Since these filters can only reduce the size
of the result set, they can be evaluated directly in the UI without the need for
further back-end communication or interaction with the underlying graph.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Demonstration</title>
      <p>During the demonstration, participants will have the opportunity to use the
system in a setting analogous to the evaluation during LSC. They will be provided
with a textual description of a point in the life of the lifelogger whose
experiences serve as foundation for the dataset. They can then try to find the described
life events using LifeGraphs retrieval and browsing functionality. A screenshot
illustrating the system in action is shown in Figure 1.
In this paper, we demonstrated LifeGraph, a first step into the direction of
using a Knowledge Graph in the context of the emerging phenomenon which is
Lifelogging. Due to the inherent multi-modality and semi-structured nature of
lifelogs, we see great potential in the interaction of the Semantic Web and the
Lifelogging communities and are looking forward to future developments in this
area.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bruton</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Langford</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reed</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Snelling</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Classification of everyday living version 1</article-title>
          .0 (
          <issue>2019</issue>
          ), https://docs.oasis-open.org/coel/COEL/v1.0/os/COEL-v1.
          <article-title>0-os</article-title>
          .html,
          <source>last updated 23 January 2019</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Gasser</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rossetto</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuldt</surname>
          </string-name>
          , H.:
          <article-title>Towards an all-purpose content-based multimedia information retrieval system</article-title>
          .
          <source>arXiv preprint arXiv:1902</source>
          .
          <volume>03878</volume>
          (
          <year>2019</year>
          ), https: //arxiv.org/abs/
          <year>1902</year>
          .03878
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>T.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ninh</surname>
            ,
            <given-names>V.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dang-Nguyen</surname>
            ,
            <given-names>D.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jónsson</surname>
            ,
            <given-names>B.Þ.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lokoš</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hürst</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tran</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schoeffmann</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Introduction to the Third Annual Lifelog Search Challenge (LSC'20)</article-title>
          .
          <source>In: Proceedings of the 2020 International Conference on Multimedia Retrieval</source>
          . pp.
          <fpage>584</fpage>
          -
          <lpage>585</lpage>
          (
          <year>2020</year>
          ). https://doi.org/10.1145/3372278.3388043
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Redmon</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farhadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>YOLO9000: better, faster, stronger</article-title>
          .
          <source>In: 2017 IEEE Conference on Computer Vision</source>
          and Pattern Recognition,
          <string-name>
            <surname>CVPR</surname>
          </string-name>
          <year>2017</year>
          ,
          <article-title>Honolulu</article-title>
          ,
          <string-name>
            <surname>HI</surname>
          </string-name>
          , USA, July
          <volume>21</volume>
          -
          <issue>26</issue>
          ,
          <year>2017</year>
          . pp.
          <fpage>6517</fpage>
          -
          <lpage>6525</lpage>
          . IEEE Computer Society (
          <year>2017</year>
          ). https://doi.org/10.1109/CVPR.
          <year>2017</year>
          .690
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Rossetto</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baumgartner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ashena</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruosch</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pernischová</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bernstein</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Lifegraph: A knowledge graph for lifelogs</article-title>
          .
          <source>In: Proceedings of the Third Annual Workshop on Lifelog Search Challenge</source>
          . pp.
          <fpage>13</fpage>
          -
          <lpage>17</lpage>
          (
          <year>2020</year>
          ). https://doi.org/10.1145/3379172.3391717
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Rossetto</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gasser</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heller</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amiri</surname>
            <given-names>Parian</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Schuldt</surname>
          </string-name>
          , H.:
          <article-title>Retrieval of structured and unstructured data with vitrivr</article-title>
          .
          <source>In: Proceedings of the ACM Workshop on Lifelog Search Challenge</source>
          . pp.
          <fpage>27</fpage>
          -
          <lpage>31</lpage>
          (
          <year>2019</year>
          ). https://doi.org/10.1145/3326460.3329160
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Rossetto</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giangreco</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tanase</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuldt</surname>
          </string-name>
          , H.:
          <article-title>vitrivr: A flexible retrieval stack supporting multiple query modes for searching in multimedia collections</article-title>
          .
          <source>In: Proceedings of the 24th ACM international conference on Multimedia</source>
          . pp.
          <fpage>1183</fpage>
          -
          <lpage>1186</lpage>
          (
          <year>2016</year>
          ). https://doi.org/doi.org/10.1145/2964284.2973797
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Vrandečić</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krötzsch</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Wikidata: a free collaborative knowledgebase</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>57</volume>
          (
          <issue>10</issue>
          ),
          <fpage>78</fpage>
          -
          <lpage>85</lpage>
          (
          <year>2014</year>
          ). https://doi.org/10.1145/2629489
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lapedriza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khosla</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oliva</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Torralba</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Places: A 10 million image database for scene recognition</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          (
          <year>2017</year>
          ). https://doi.org/10.1109/TPAMI.
          <year>2017</year>
          .2723009
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>