=Paper=
{{Paper
|id=Vol-2721/paper557
|storemode=property
|title=A Knowledge Graph-based System for Retrieval of Lifelog Data
|pdfUrl=https://ceur-ws.org/Vol-2721/paper557.pdf
|volume=Vol-2721
|authors=Luca Rossetto,Matthias Baumgartner,Narges Ashena,Florian Ruosch,Romana Pernisch,Abraham Bernstein
|dblpUrl=https://dblp.org/rec/conf/semweb/RossettoBARPB20
}}
==A Knowledge Graph-based System for Retrieval of Lifelog Data==
A Knowledge Graph-based System for Retrieval of Lifelog Data Luca Rossetto, Matthias Baumgartner, Narges Ashena, Florian Ruosch, Romana Pernisch, and Abraham Bernstein Department of Informatics, University of Zurich, Zurich, Switzerland {lastname}@ifi.uzh.ch Abstract. Lifelogging is a phenomenon where practitioners record an increasing part of their subjective daily experience with the aim of later being able to use these recordings as a memory aid or basis for data- driven self improvement. The resulting lifelogs are, therefore, only useful if the lifeloggers have efficient ways to search through them. The logs are inherently multi-modal and semi structured, combining data from several sources, such as cameras and other wearable physical as well as virtual sensors, so representing the data in a graph structure can effec- tively capture all produced interrelations. Since annotating each entry with a sufficiently large semantic context is infeasible, either manually or automatically, alternatives must be found to capture the higher level semantics. In this paper, we demonstrate LifeGraph, a first approach of creating a Knowledge Graph-based lifelog representation and retrieval solution, able of capturing a lifelog in a graph structure and augmenting it with external information to aid with the association of higher-level semantic information. Keywords: Lifelogging, Lifelog Retrieval, Multi-modal Retrieval, Multi- modal Graphs 1 Introduction Various technological advances over the last few decades have led to a rapid in- crease in possibilities for mobile sensing, processing, and storing of various forms of data. One of the many consequences of these possibilities was the emergence of the quantified-self movement as well as the Lifelogging phenomenon. Lifelogging describes the act of capturing aspects of one’s personal experience using cam- eras to continuously record one’s point of view, sensors to quantify bio-feedback, or properties of one’s immediate surroundings, purely software-based means to analyze one’s interaction with the digital environment as well as more tradi- tional methods, such as entries in a personal diary. Thus, the resulting lifelog is Copyright c 2020 for this paper by its authors. Use permitted under Creative Com- mons License Attribution 4.0 International (CC BY 4.0). 2 Rossetto et al. inherently multi-modal and can exhibit a high degree of interconnectivity. An appropriate, but currently under-explored approach, is, therefore, to represent a lifelog as a knowledge graph. Lifelogs are commonly used as a memory aid or to monitor one’s own life in order to improve or maintain health. In order for these logs to be useful, it is paramount that lifeloggers have the possibility to efficiently search within the logs which becomes increasingly difficult due to the growing volume and diversity of available data. To catalyze research in the area of lifelog retrieval, a competitive benchmark called the Lifelog Search Challenge (LSC) [3] was es- tablished in 2018 as an annual workshop to evaluate interactive lifelog retrieval systems. In this paper, we demonstrate an interactive knowledge graph-based lifelog retrieval system called LifeGraph [5], which was first introduced in the context of the 2020 edition of the LSC. The following provides an overview of how the graph was constructed, how the system uses it to answer queries, and how users can expect to be able to interact with the system itself. 2 Constructing LifeGraph The basis for LifeGraph is formed by the lifelog dataset made available in the context of LSC 2020, which consists of 114 days of lifelogging data made up of roughly 200k images taken by a wearable camera with accompanying metadata as well as various sensor data. Ideally, each entry in a lifelog, independent of its type and method of representation, would have rich semantic annotations describing not only its content but also placing it into a larger context. Since generating such annotations manually is infeasible and doing so automatically is beyond what is currently possible, alternatives need to be used to make the collected data searchable and, hence, usable. To overcome this limitation, we construct a graph structure which captures the relations between higher level semantic concepts and lower level labels, such as the type of environment or the presence of everyday objects within an image, both of which can be reliably detected with currently available technology [9,4]. A challenge of constructing LifeGraph is to link high-level search entities with low-level image features. We use COEL (Classification of Everyday Liv- ing) [1] as a starting point which provides a taxonomy of everyday actions, such as housework, sports, or food related activities. In order to combine COEL with detectable entities, we defined a mapping to Wikidata [8]. It was created semi-automatically, based on the similarity to Wikidata entity labels, whereas ambiguities were resolved manually. To connect the mapped concepts to lower-level semantic image labels we com- bined them with additional data from Wikidata. For this, we created a graph initially consisting of the Wikidata entities mapped to COEL as well as all de- tectable entities. We then successively retrieved all entities from wikidata that can be connected to at least two entities in the present graph, and added them to the graph. This procedure is repeated three times, and two additional times considering only class relations (:instance-of, :subclassOf). Finally, we dis- A Knowledge Graph-based System for Retrieval of Lifelog Data 3 carded disconnected components of the graph since they failed to link image labels to high-level concepts. We further removed entities intended for naviga- tional or internal purpose, such as disambiguation or template pages. We modeled the metadata which was already part of the dataset provided for LSC in a simple and flat way following the schema explained in [5]. Each metadata entry is associated with an image based on time. Using either an object or datatype property, we link all the metadata information with the image in the graph. Contrary to what was stated in [5], each image is linked to a day resource. This enables us to link day-specific information directly to the day resource instead of repeating this information for all ca. 2000 images of that day. Before populating the schema, we interpolated some aspects of the metadata such as the provided semantic location labels (“work” vs. “home”) but also detected objects [4] and places [9]. If the same concept is interrupted by a sufficiently small gap, we add associations to this concept to all entries in the gap. To associate images with objects and places, we then directly establish a link between images and Wikidata entries using either the :detected or :detected_place property. 3 Exploring LifeGraph The browser-based LifeGraph user interface is a modified version of vitrivr-ng [2] – the UI of the vitrivr [7] content-based multimedia retrieval stack – which has already been successfully used for effective retrieval of lifelog data [6]. Apart from the user interface, LifeGraph does not share any logic with the vitrivr stack. For the sake of simplicity, a user can specify queries by selecting nodes from the graph, representing the concepts in which they are interested, using a simple auto-completing text box. This query formulation scheme offers a reasonable compromise between specificity and efficiency while not requiring any technical expertise on the user side. Once specified, a query is sent to a back-end which performs its execution and the subsequent scoring of the results. The query is evaluated by traversing the graph, searching for paths from each of the specified start nodes to all log entries which could be part of the result. The maximum length of this path is continuously increased until an empirically determined threshold is reached, either in path length or number of potential results. Each log entry found in this way is scored by the sum of the inverse of the path lengths leading to it, normalized by the number of start points in the query. This means the shorter the distance to the highest number of start points, the higher the score of the retrieved item. These aggregated results are then returned to the UI, where they are presented as either a list of log entries sorted by score or a list of entries grouped by day, sorted by time within each. Since the user is supposed to be able to browse through the retrieved results in order to identify the desired elements, the inevitable false-positives produced by some of the underlying detectors are less of a problem than missing detections or false-negatives, since a large number of false-positives will cause users to spend an unnecessarily large amount of time browsing, while too many false-negatives could prevent them from finding 4 Rossetto et al. the desired result at all. All log entries are accompanied by some metadata, including technical information, like its originating time and date. Additionally, entries include lifelog specific semantic information, such as manual semantic annotations or bio feedback properties, e.g., heart-rate, number of steps taken within a specific interval, etc. The UI uses these metadata to provide additional filtering options to a user, excluding entries from the result if they do not match Boolean criteria relevant to the query. Since these filters can only reduce the size of the result set, they can be evaluated directly in the UI without the need for further back-end communication or interaction with the underlying graph. 4 Demonstration During the demonstration, participants will have the opportunity to use the sys- tem in a setting analogous to the evaluation during LSC. They will be provided with a textual description of a point in the life of the lifelogger whose experi- ences serve as foundation for the dataset. They can then try to find the described life events using LifeGraphs retrieval and browsing functionality. A screenshot illustrating the system in action is shown in Figure 1. Fig. 1. Screenshot of the LifeGraph retrieval system in action. The shown user in- terface is a modified version of vitrivr-ng [2]. The retrieved results have been made unrecognizable to comply with the privacy protection requirements of the dataset. 5 Conclusion and Outlook In this paper, we demonstrated LifeGraph, a first step into the direction of using a Knowledge Graph in the context of the emerging phenomenon which is A Knowledge Graph-based System for Retrieval of Lifelog Data 5 Lifelogging. Due to the inherent multi-modality and semi-structured nature of lifelogs, we see great potential in the interaction of the Semantic Web and the Lifelogging communities and are looking forward to future developments in this area. References 1. Bruton, P., Langford, J., Reed, M., Snelling, D.: Classification of everyday living version 1.0 (2019), https://docs.oasis-open.org/coel/COEL/v1.0/os/COEL-v1. 0-os.html, last updated 23 January 2019 2. Gasser, R., Rossetto, L., Schuldt, H.: Towards an all-purpose content-based multi- media information retrieval system. arXiv preprint arXiv:1902.03878 (2019), https: //arxiv.org/abs/1902.03878 3. Gurrin, C., Le, T.K., Ninh, V.T., Dang-Nguyen, D.T., Jónsson, B.Þ., Lokoš, J., Hürst, W., Tran, M.T., Schoeffmann, K.: Introduction to the Third Annual Lifelog Search Challenge (LSC’20). In: Proceedings of the 2020 International Conference on Multimedia Retrieval. pp. 584–585 (2020). https://doi.org/10.1145/3372278.3388043 4. Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Hon- olulu, HI, USA, July 21-26, 2017. pp. 6517–6525. IEEE Computer Society (2017). https://doi.org/10.1109/CVPR.2017.690 5. Rossetto, L., Baumgartner, M., Ashena, N., Ruosch, F., Pernischová, R., Bernstein, A.: Lifegraph: A knowledge graph for lifelogs. In: Proceedings of the Third Annual Workshop on Lifelog Search Challenge. pp. 13–17 (2020). https://doi.org/10.1145/3379172.3391717 6. Rossetto, L., Gasser, R., Heller, S., Amiri Parian, M., Schuldt, H.: Re- trieval of structured and unstructured data with vitrivr. In: Proceedings of the ACM Workshop on Lifelog Search Challenge. pp. 27–31 (2019). https://doi.org/10.1145/3326460.3329160 7. Rossetto, L., Giangreco, I., Tanase, C., Schuldt, H.: vitrivr: A flexible retrieval stack supporting multiple query modes for searching in multimedia collections. In: Proceedings of the 24th ACM international conference on Multimedia. pp. 1183– 1186 (2016). https://doi.org/doi.org/10.1145/2964284.2973797 8. Vrandečić, D., Krötzsch, M.: Wikidata: a free collaborative knowledgebase. Com- munications of the ACM 57(10), 78–85 (2014). https://doi.org/10.1145/2629489 9. Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence (2017). https://doi.org/10.1109/TPAMI.2017.2723009