<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Organizer Team at ImageCLEFlifelog 2017: Baseline Approaches for Lifelog Retrieval and Summarization</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Liting</forename><surname>Zhou</surname></persName>
							<email>zhou.liting2@mail.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">Insight Centre for Data Analytics</orgName>
								<orgName type="institution">Dublin City University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Luca</forename><surname>Piras</surname></persName>
							<email>luca.piras@diee.unica.it</email>
							<affiliation key="aff1">
								<orgName type="department">DIEE</orgName>
								<orgName type="institution">University of Cagliari</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Michael</forename><surname>Riegler</surname></persName>
							<email>michael@simula.no</email>
							<affiliation key="aff2">
								<orgName type="laboratory">Simula Research Laboratory</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giulia</forename><surname>Boato</surname></persName>
							<email>boato@disi.unitn.it</email>
							<affiliation key="aff3">
								<orgName type="department">DISI</orgName>
								<orgName type="institution">University of Trento</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Duc-Tien</forename><surname>Dang-Nguyen</surname></persName>
							<email>duc-tien.dang-nguyen@dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">Insight Centre for Data Analytics</orgName>
								<orgName type="institution">Dublin City University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cathal</forename><surname>Gurrin</surname></persName>
							<email>cathal.gurrin@dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">Insight Centre for Data Analytics</orgName>
								<orgName type="institution">Dublin City University</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Organizer Team at ImageCLEFlifelog 2017: Baseline Approaches for Lifelog Retrieval and Summarization</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">87F2E6D509066018BE0737A0AAE40675</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes the participation of Organizer Team in the ImageCLEFlifelog 2017 Retrieval and Summarization subtasks. In this paper, we propose some baseline approaches, using only the provided information, which require different involvement levels from the users. With these baselines we target at providing references for other approaches that aim to solve the problems of lifelog retrieval and summarization.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Personalized multimedia archives that contain a large amount of data collected using various personal devices, such as smart phones, cameras, wearable devices and so on are getting more and more common nowadays. In these archives, every moment and aspect of our lives are stored. They can contain information about our daily routines, consumed food but also about our health status, etc. These data logs of a human lives, also commonly referred to as lifelogs, are more and more interesting for the research community but also companies. Collecting and storing the data is one challenge but getting insights from the collected data and find new information by connecting different types of data requires a lot of researches for analyzing, categorizing and querying these huge amounts of data in a efficient way.</p><p>In this paper we present our approach to tackle the Image CLEF 2017 <ref type="bibr" target="#b10">[11]</ref> Lifelog Task <ref type="bibr" target="#b5">[6]</ref>, which aims at solving the problems of lifelog retrieval and summarization. Lifelogs are usually chronologically organized and moments that belong to the same activity or the same event are normally very similar. This can be exploited to reduce processing time by grouping moments that are similar based on the time when they happened and the belonging concepts. This transforms the image retrieval challenge into a image segments retrieval challenge. This has the advantage that boundaries between moments or activities are automatically segmented based on time and concepts <ref type="bibr" target="#b6">[7]</ref>. To remove non-relevant images filtering is recommended. In our case, we remove images that seem to be sparse on information (blurry, only big objects, etc.) Retrieved images then can be diversified into clusters which then can be further used for summarization, which can be done automatically or via relevance feedback by follow the methods described in <ref type="bibr" target="#b4">[5]</ref>.</p><p>The remainder of this paper is organized as follows, first we present related work in the field. This is followed by a detailed description of our approach. After that we present the experimental results which is followed by a discussion and conclusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>In this section we discuss briefly recent studies on lifelog segmentation and the retrieval problem in term of relevant and diversity. In addition, many novel techniques are proposed and evaluated, to accurately retrieve the similar events from lifelog dataset using contextual data.</p><p>Typically, chronological images segmentation is done by heuristic split based on long interval with no capture <ref type="bibr" target="#b13">[14]</ref> or by thresholding the distances between the frames (or images) based on the content <ref type="bibr" target="#b2">[3]</ref>. Doherty et al. in <ref type="bibr" target="#b7">[8]</ref>, to determine the similarity between adjacent block of images, proposed to use Hearst's Text Tiling Algorithm <ref type="bibr" target="#b8">[9]</ref> on edge histogram, which is extracted by using Canny edge detection. For egocentric photo streams (from wearable cameras), a typical segmentation is based on unsupervised hierarchical agglomerative clustering to extract the key-frame summary <ref type="bibr" target="#b0">[1]</ref>.</p><p>Current works in multimedia retrieval have considered relevance and diversity as two core criteria. Relevance was commonly estimated based on textual information, e.g., from the photo tags, and many of current search engines are still mainly based on this information. Diversity is usually improved by applying clustering algorithms which rely on textual or/and visual properties <ref type="bibr" target="#b11">[12]</ref>. Recently, in social image retrieval, some methods have exploited the participation of humans by collecting the feedbacks of the results to improve the diversification <ref type="bibr" target="#b1">[2]</ref>. To reduce the number of images to be returned to the user, some papers in the past years proposed the use of image clustering techniques <ref type="bibr" target="#b14">[15]</ref>. These approaches exploit the hierarchical indexing structure of the clusters to refine the number of images to consider <ref type="bibr" target="#b12">[13]</ref>. More recently different type of relevance feedback has been used to expand the query to improve both relevant and diversity as well as reduce the number of iterations <ref type="bibr" target="#b4">[5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">The proposed approaches</head><p>The proposed approaches follow the schema as illustrated in Figure <ref type="figure">1</ref>. Since lifelogs are chronologically organized and moments in the same activity or the same -NLP -Manual</p><p>Fig. <ref type="figure">1</ref>. Schema of the proposed methods.</p><p>event are normally very similar to each other, in order to reduce the processing time, we group similar moments together based on time and concepts. By applying this chronological-based segmentation, we turn the problem of images retrieval into image segments retrieval, in which the boundary between activities such as having breakfast, working in front of a computer, and so on <ref type="bibr" target="#b6">[7]</ref>, are automatically decided based on the time and concepts. Starting from a topic query, it is transformed into small inquiries, where each of them is asking for a single piece of information of concepts, location, activity, and time. The moments that matched all of those requirements are returned as the retrieval results. In order to remove the non-relevant images, a filtering step is applied on the retrieved images, by removing blurred and images that covered mainly by huge object or by the arms of the user. Finally, the images are diversified into clusters and the top images that close to center are selected for the summarization, which can be done automatically or using relevance feedback by follow the methods in <ref type="bibr" target="#b4">[5]</ref>. These steps are described as follows:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Segmentation</head><p>For the segmentation we applied a simple chronological-based segmentation as follow: For each pair of two continuous images I t and I t+1 at the time t, the distance d(I t , I t+1 ) between them is computed as:</p><formula xml:id="formula_0">d(I t , I t+1 ) = ||C t − C t+1 ||</formula><p>where || • || is the normalized Euclidean distance, and C is the concept vector of each image provided from the task. If d(I t , I t+1 ) &lt; τ , where τ is a threshold, the two images are set belong to the same segment, otherwise they are set in different segments. If τ is too small, an activity should be split into small activities, while larger value of tau should grouped different activities into the same one. Since || • || is normalized, when τ = 0, the images are grouped into different segments, and when τ = 1, all images are belong to a single segment.</p><p>Segmenting the activities is not simply an incident of identifying the exact event boundaries; it also concerns with keeping track of the fine-grained group of events together into extended meaningful units, and thus deciding the right value of τ is not trivial. In the proposed approaches, we try different values of tau for different runs, which will be explained in Section 4.</p><p>After this step, each segment is represented by the first image (of that segment) with these basic information: location, activity, time segment, number of people and the list of the concepts. If any of these information is missing from the first image, we take it from the second image and so on.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Parsing the Query and Retrieval</head><p>Converting a topic into precise criteria for retrieval is the key question for both sub-tasks. It can be automatically done by considering any word in the topic as the queried concepts and then searching for all segments that contain those concepts; by applying natural language processing techniques; or by fine-tunning by a human in the loop, i.e., the user will read the topic and "translate" it into the search criteria. For example, with the topic:</p><p>-Topic: Using laptop out of office -Query: Find the moment(s) in which user u1 was using his laptop outside the working places. -Description: To be consider to relevant, the user should use his laptop, for work or for entertainment out of his working place.</p><p>Can be "translated into":</p><p>-User: +u1 -Concepts: +laptop -Activities: +working -Time: ---Location: -work</p><p>where +/-means the retrieval images has to contain/not contain this information, respectively, and --means any.</p><p>In the proposed approaches, we use the automatic and the human-in-the-loop methods. Our "translation" will be shown in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Filtering</head><p>An image can be considered as blurred based on it focus level. In the proposed approaches, we estimate the focus by computing the absolute sum of the wavelet coefficients and comparing it to a threshold, by exploiting the method in <ref type="bibr" target="#b9">[10]</ref>. The return of this method is a scalar number in [0, 99] which the bigger value the sharper image. From our observation, for values below 30, most of the images are blurred, and thus we set this threshold to 30.</p><p>In order to remove images that covered by large objects, we apply an heuristic method as follows:</p><p>Step 1 Convert the image to binary images by applying thresholding with several thresholds.</p><p>Step 2 Extract connected components and calculate their centers.</p><p>Step 3 Group centers based on their coordinates, and then close them to form the corresponds blob.</p><p>Step 4 Take the biggest blob and its size (in pixels).</p><p>If the size is over 50% of the whole area, the image is considered as covered. This whole method is implemented by calling the function SimpleBlobDetector from OpenCV 5 .</p><p>After this step, all remain images are considered as relevant to the topic. Please notice that the images still kept inside the segment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Diversification</head><p>In this step, for automatic approach, we use a hierarchical agglomerative clustering algorithm (see in <ref type="bibr" target="#b3">[4]</ref>) to group similar segments into the same cluster based on the concepts. The clusters are then sorted based on the number of segments, decreasingly. Finally, we produce the summary for the queried by selecting representative images from the clusters based by selecting the images closest to the center of each cluster.</p><p>We also propose a human-in-the-loop approach in this step by using the usual dichotomous Relevance Feedback paradigm (more details can be seen in <ref type="bibr" target="#b4">[5]</ref>), that asks the user to assign the labels Relevant \ Non-relevant to the retrieved images. The system asks the user to label the representative images of the top N results returned by the automatic diversification procedure (as mentioned above), and the number of images that have been labeled as being Relevant \ Non-relevant for each cluster is computed. Then, the clusters are sorted as follows:</p><p>-Clusters that have a large number of relevant counts are sorted higher.</p><p>-Clusters that have the same number of relevant counts are sorted based on the number of non-relevant counts (i.e., a cluster that contains a larger number of 'non-relevant' images should be selected later). -Clusters that have the same number of Relevant \ Non-relevant counts are sorted on the basis of the number of segments.</p><p>For each cluster, the images that are selected to represent the topic are chosen in the same way as in the automatic diversification.</p><p>4 Experimental results</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Submitted Runs</head><p>We submitted 3 runs on the Retrieval sub-task and 5 runs on the Summarization sub-task, summarized in Table <ref type="table" target="#tab_0">1</ref>.</p><p>As for the retrieval task, the first run is exploiting only time and the concepts information. We consider every single image as the basic unit and the retrieval just returns all images that contains the concepts extracted from the topics. We named this run is the 'baseline' with the purpose that any other approaches should obtain better performance than this. With the second run, we applied the optimized value for τ (optimized from the devset) to do the segmentation. So in this run, the only difference is the basic unit of retrieval now is the segment, not image.</p><p>For the Fine-tunning runs, the "translation" is applied as in Tables2, and 3. The same strategy is applied on the summarization subtask, in which the first three runs were ran to test the automatic approach with the increasing level of the 'criteria', while the last two runs are used to test the fine tuning and the relevance feedback approaches. For the relevance feedback approach, we ran a simulation by exploiting the ground-truth annotated data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Results</head><p>Shown in Tables <ref type="table" target="#tab_4">4 and 5</ref> are the results of the runs on the retrieval and summarization sub-tasks, respectively. The results confirm that applying segmentation improved both retrieval and summarization performance. It is quite clear that applying fine-tunning significantly improved the performance. The big gaps in results between the automatic approach with the fine-tunning and between the fine-tunning with the human-in-theloop (relevance feedback) approaches, shown that we need better natural language processing as well as machine learning studies for these problems. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Discussions and Conclusions</head><p>In this paper we introduced different baseline approaches, that came from fully automatic to fully manual paradigm, proposed by the Organizer Team of the ImageCLE-Flifelog 2017 task as participant of the Retrieval and Summarization subtasks. These approaches, that require different level of involvement of the users, exploit only the information provided by the organizers along with the collection of images, i.e., the description of the semantic locations and the physical activities. From the obtained results it appears clear that deeper analysis of the methods should be considered as well as the use of extra information. </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Submitted Runs.</figDesc><table><row><cell>RunID Name</cell><cell>τ</cell><cell>Parsing</cell><cell>Filtering</cell><cell>Diversification</cell></row><row><cell>LRT Run 1 Baseline</cell><cell>0</cell><cell>Automatic</cell><cell>-</cell><cell>-</cell></row><row><cell>LRT Run 2 Segmentation</cell><cell cols="2">0.05 Automatic</cell><cell>-</cell><cell>-</cell></row><row><cell>LRT Run 3 Fine-tunning</cell><cell cols="2">0.05 Fine-tunning</cell><cell>-</cell><cell>-</cell></row><row><cell>LST Run 1 Baseline</cell><cell>0</cell><cell cols="2">Automatic Not apply</cell><cell>Automatic</cell></row><row><cell>LST Run 2 Segmentation</cell><cell cols="3">0.05 Automatic Not apply</cell><cell>Automatic</cell></row><row><cell>LST Run 3 Filtering</cell><cell cols="2">0.05 Automatic</cell><cell>Apply</cell><cell>Automatic</cell></row><row><cell>LST Run 4 Fine-tunning</cell><cell cols="3">0.05 Fine-tunning Apply</cell><cell>Automatic</cell></row><row><cell cols="5">LST Run 5 Relevance Feedback 0.05 Fine-tunning Apply Relevance Feedback</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Parsing as a fine-tuning on testset, LRT subtask. + means selection and -means exception.</figDesc><table><row><cell cols="2">Topic User</cell><cell>Activities</cell><cell>Times</cell><cell>Locations</cell><cell>Concepts</cell></row><row><cell>T001</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID:</cell><cell>-Work, +Home,</cell><cell>+Laptop</cell></row><row><cell></cell><cell></cell><cell>-Running</cell><cell>400-1400</cell><cell>+Science Gallery</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Caf, + Helix</cell><cell></cell></row><row><cell>T002</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID: 720-</cell><cell>-Work, -Home</cell><cell>+Microphone</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>1080(workday)</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell></cell><cell></cell><cell></cell></row><row><cell>T003</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID:</cell><cell>-Work, -Home,</cell><cell>+Hard disc,</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>540-1080</cell><cell>+Dublin Airport</cell><cell>+Knee pad,</cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell></cell><cell>(DUB)</cell><cell>+Mouse, +CD</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>player</cell></row><row><cell>T004</cell><cell>u1</cell><cell>+Running,</cell><cell>+MinuteID: 400-</cell><cell>-Work, -Home,</cell><cell>+Park bench</cell></row><row><cell></cell><cell></cell><cell>-Walking</cell><cell>660(weekend)</cell><cell>+Place in Saint</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Anne's Park,</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>+Hampstead</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Park</cell><cell></cell></row><row><cell>T005</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID: 540-</cell><cell cols="2">+Work, -Home +Table, +Laptop</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>1140(workday)</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell></cell><cell></cell><cell></cell></row><row><cell>T006</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID: 400-</cell><cell>-Work, +Home</cell><cell>+TV</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>540(workday),</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell>1140-</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>1400(workday),</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>400-</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>1400(weekend)</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Parsing as a fine-tuning on testset, LST subtask. + means selection and -means exception.</figDesc><table><row><cell cols="2">Topic User</cell><cell>Activities</cell><cell>Times</cell><cell>Locations</cell><cell>Concepts</cell></row><row><cell>T001</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID: 540-</cell><cell cols="2">+Work, -Home +Table, +Laptop</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>1140(workday)</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell></cell><cell></cell><cell></cell></row><row><cell>T002</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID: 400-</cell><cell>-Work, +Home</cell><cell>+TV</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>540(workday),</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell>1140-</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>1400(workday),</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>400-</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>1400(weekend)</cell><cell></cell><cell></cell></row><row><cell>T003</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID:</cell><cell>-Work, +Home,</cell><cell>+Laptop</cell></row><row><cell></cell><cell></cell><cell>-Running</cell><cell>400-1400</cell><cell>+Science Gallery</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Caf, + Helix</cell><cell></cell></row><row><cell>T004</cell><cell>u1</cell><cell>-Walking,</cell><cell>+MinuteID:</cell><cell>+Home</cell><cell>+Laptop,</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>400-540,</cell><cell></cell><cell>+Notebook</cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell>1140-1400</cell><cell></cell><cell></cell></row><row><cell>T005</cell><cell>u2</cell><cell>-Waking,</cell><cell>+MinuteID:</cell><cell>+DCU, +Home,</cell><cell>+Banana,</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>400-1400</cell><cell>+Starbucks,</cell><cell>+Apple, +Peach,</cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell></cell><cell>+Costa Coffee</cell><cell>+Broccoli,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>+Spaghetti</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>squash,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>+Cheeseburger,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>+Hotdog, +</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>Mashed potato</cell></row><row><cell>T006</cell><cell>u2</cell><cell>-Waking,</cell><cell>+MinuteID: 720-</cell><cell>-Work,-Home</cell><cell>+Wine bottle,</cell></row><row><cell></cell><cell></cell><cell>-Running,</cell><cell>1400(weekend),</cell><cell></cell><cell>+Beer bottle,</cell></row><row><cell></cell><cell></cell><cell>-Transport</cell><cell>960-</cell><cell></cell><cell>+Beer glass</cell></row><row><cell></cell><cell></cell><cell></cell><cell>1400(workday)</cell><cell></cell><cell></cell></row><row><cell>T007</cell><cell>u2</cell><cell>-Transport,</cell><cell>+MinuteID:</cell><cell>-Work, -Home,</cell><cell>N/A</cell></row><row><cell></cell><cell></cell><cell>+Walking,</cell><cell>400-1140</cell><cell>+Place in</cell><cell></cell></row><row><cell></cell><cell></cell><cell>-Running</cell><cell></cell><cell>Beijing, +Place</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>in Yong He Gong</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Lama Temple,</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>+Place in</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Chaoyang</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Lifelog Retrieval Results.</figDesc><table><row><cell>Run</cell><cell>Name</cell><cell>Average NDCG</cell></row><row><cell cols="2">LRT Run 1 Baseline</cell><cell>0.09</cell></row><row><cell cols="2">LRT Run 2 Segmentation</cell><cell>0.14</cell></row><row><cell cols="2">LRT Run 3 Fine Tuning</cell><cell>0.39</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>Lifelog Summarization Results.</figDesc><table><row><cell>Run</cell><cell>Name</cell><cell>Average F1@10</cell></row><row><cell cols="2">LST Run 1 Baseline</cell><cell>0.10</cell></row><row><cell cols="2">LST Run 2 Segmentation</cell><cell>0.17</cell></row><row><cell cols="2">LST Run 3 Filtering</cell><cell>0.18</cell></row><row><cell cols="2">LST Run 4 Fine Tuning</cell><cell>0.32</cell></row><row><cell cols="2">LST Run 5 Relevance Feedback</cell><cell>0.77</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_0">http://opencv.org</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Senseseer mobilecloud-based lifelogging framework. Visual summary of egocentric photostreams by representative keyframes</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bolanos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mestre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Talavera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">G</forename><surname>Nieto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Radeva</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015-07">July 2015</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A relevance feedback perspective to image search result diversification</title>
		<author>
			<persName><forename type="first">B</forename><surname>Boteanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Mironic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ionescu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 10th International Conference on Intelligent Computer Communication and Processing (ICCP)</title>
				<imprint>
			<date type="published" when="2014-09">2014. Sept 2014</date>
			<biblScope unit="page" from="47" to="54" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Using bluetooth and gps metadata to measure event similarity in sensecam images</title>
		<author>
			<persName><forename type="first">D</forename><surname>Byrne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lavelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Doherty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">J</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Smeaton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th International Conference on Intelligent Multimedia and Ambient Intelligence</title>
				<imprint>
			<date type="published" when="2007-07">July 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A hybrid approach for retrieving diverse social images of landmarks</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Dang-Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Piras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Giacinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Boato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">G</forename><surname>De Natale</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Multimedia and Expo (ICME)</title>
				<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Multimodal retrieval with diversification and relevance feedback for tourist attraction images</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Dang-Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Piras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Giacinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Boato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">G</forename><surname>De Natale</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Multimedia Computing</title>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Communications, and Applications</publisher>
		</imprint>
	</monogr>
	<note>accepted</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Overview of ImageCLEFlifelog 2017: Lifelog Retrieval and Summarization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Dang-Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Piras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Boato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</author>
		<ptr target="WS.org&lt;http://ceur-ws.org&gt;" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2017 Labs Working Notes. CEUR Workshop Proceedings</title>
				<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">September 11-14 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Automatically segmenting lifelog data into events</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Doherty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Smeaton</surname></persName>
		</author>
		<ptr target="http://doras.dcu.ie/4651/" />
	</analytic>
	<monogr>
		<title level="m">9th International Workshop on Image Analysis for Multimedia Interactive Services</title>
				<imprint>
			<date type="published" when="2008-06-30">30 June 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Multimodal segmentation of lifelog data. Large Scale Semantic Access to Content</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Doherty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Smeaton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Ellis</surname></persName>
		</author>
		<ptr target="http://dl.acm.org/citation.cfm?id=1931393" />
	</analytic>
	<monogr>
		<title level="m">Text, Image, Video, and Sound</title>
				<imprint>
			<date type="published" when="2007-06">June 2007</date>
			<biblScope unit="page" from="21" to="38" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Texttiling: A quantitative approach to discourse segmentation</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Hearst</surname></persName>
		</author>
		<idno>UCB:S2K-93-24</idno>
		<imprint>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Robust measure of image focus in the wavelet domain</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">T</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Phoong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Intelligent Signal Processing and Communication Systems</title>
				<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="157" to="160" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Overview of ImageCLEF 2017: Information extraction from images</title>
		<author>
			<persName><forename type="first">B</forename><surname>Ionescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Villegas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Arenas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Boato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Dang-Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dicente Cid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Eickhoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garcia Seco De Herrera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Islam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kovalev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Liauchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mothe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Piras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Schwall</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction 8th International Conference of the CLEF Association, CLEF 2017</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">September 11-14 2017</date>
			<biblScope unit="volume">10456</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Visual diversification of image search results</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">H</forename><surname>Van Leuken</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Garcia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Olivares</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th International Conference on World Wide Web</title>
				<meeting>the 18th International Conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="341" to="350" />
		</imprint>
	</monogr>
	<note>WWW &apos;09</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Hierarchical clustering relevance feedback for content-based image retrieval</title>
		<author>
			<persName><forename type="first">I</forename><surname>Mironica</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ionescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vertan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Workshop on Content-Based Multimedia Indexing</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Peitgen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jürgens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Saupe</surname></persName>
		</author>
		<title level="m">Chaos and fractals -new frontiers of science</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
	<note>2</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Interactive search in image retrieval: a survey</title>
		<author>
			<persName><forename type="first">B</forename><surname>Thomee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Lew</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Multimedia Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="71" to="86" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
