<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">UAM&apos;s participation at CLEF eRisk 2017 task: Towards modelling depressed bloggers</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Esaú</forename><surname>Villatoro-Tello</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Information Technologies Department</orgName>
								<orgName type="laboratory">Language and Reasoning Research Group</orgName>
								<orgName type="institution">Universidad Autónoma Metropolitana (UAM) Unidad Cuajimalpa</orgName>
								<address>
									<settlement>México City</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gabriela</forename><surname>Ramírez-De-La-Rosa</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Information Technologies Department</orgName>
								<orgName type="laboratory">Language and Reasoning Research Group</orgName>
								<orgName type="institution">Universidad Autónoma Metropolitana (UAM) Unidad Cuajimalpa</orgName>
								<address>
									<settlement>México City</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Héctor</forename><surname>Jiménez-Salazar</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Information Technologies Department</orgName>
								<orgName type="laboratory">Language and Reasoning Research Group</orgName>
								<orgName type="institution">Universidad Autónoma Metropolitana (UAM) Unidad Cuajimalpa</orgName>
								<address>
									<settlement>México City</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">UAM&apos;s participation at CLEF eRisk 2017 task: Towards modelling depressed bloggers</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">100DD3FC78A4C485286F71EEE679A2D2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>N-gram graphs</term>
					<term>Graph similarities</term>
					<term>Early text classification</term>
					<term>Natural Language Processing</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we describe the participation of the Language and Reasoning Research Group of UAM Cuajimalpa at eRisk 2017 pilot task: Early Risk Prediction on the Internet. The goal of the eRisk task consists in detecting with enough anticipation cases of depression on texts. For evaluating this task, organizers provided a dataset containing comments from a set of Social Media users. All comments are chronologically ordered and represent writings from depressed and non-depressed users. Our proposed approach addressed this problem by means of graph models. This type of representation allows to capture some inherent characteristics from documents that can be determined though traditional graph measurements, and then, employed as features in a supervised classification system. Obtained results indicate that more experiments, as well as a more thorough analysis is required to conclude regarding the pertinence (or not) of our proposed strategy.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social media is an excellent tool for anyone to express their opinion about any topic in any context; furthermore, social media is perhaps the most used communication channel nowadays. In spite of this easiness of communication, this type of tools comprise a major threat to users, who are exposed to a number of risks and potential attacks. Consider, for instance, the problem of detecting sexual predators approaching minors or the identification of aggressive users <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5]</ref>. These threats pose a challenge to the research community, that has to develop protective and preventive tools for avoiding potential risks. As mentioned in <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7]</ref>, more recently, special attention has been paid to another type of threats, such as those menaces coming from the individuals themselves, for instance depression, mental situation that may lead to more complicated situations such as suicide.</p><p>Although considerable research has been devoted to detect different types of threats, most of the current solutions work in a forensic scenario, i.e., they are applied once the treat or the attack has been accomplished. Even though these forensic methods are very useful in some particular scenarios, it is known that preventive mechanisms would have a greater and immediate impact into users own security.</p><p>Accordingly, the eRisk <ref type="bibr" target="#b6">[7]</ref> at CLEF-2017 <ref type="foot" target="#foot_0">1</ref> proposes an exploratory task on early risk detection of depression. The challenge consists of sequentially processing pieces of evidence and detect early traces of depression as soon as possible. The task is mainly concerned about evaluating Text Mining solutions and, thus, it concentrates on texts written in Social Media<ref type="foot" target="#foot_1">2</ref> .</p><p>We address the eRiks problem as a Text Classification (TC) problem. However, contrary to most TC techniques, we go beyond the Bag-of-Terms (BoT) models, and instead we represent documents and categories as graph models. This type of representation is able to incorporate contextual information by means of considering terms<ref type="foot" target="#foot_2">3</ref> co-occurrence values, which is an important limitation of the traditional BoT representation <ref type="bibr" target="#b7">[8]</ref>. Thus, our main hypothesis establishes that writing patterns, both from content and style, can be encapsulated by means of this type of representation and would provide to a classifier discriminatory information for learning to distinguish depressed bloggers.</p><p>The remainder of the paper is organized as follows. Section 2 describes the applied methodology; Section 3 shows our experimental setup and obtained results; finally, in Section 4 some conclusions of this work are presented.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Methodology</head><p>Our work was mainly inspired by the ideas proposed in <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>. Contrary to the traditional Bag-of-Terms representation model, a graph model considers the order of terms' appearances in the original text, thus incorporating valuable contextual information to the representation. Generally speaking, the graph model associates neighboring pairs of tokens with edges that denote their frequency of co-occurrence. As a result, documents with different term sequences end up having identical or at least highly similar representations. It is worth mentioning that authors from <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>, performed and exhaustive evaluation of this approach on thematic text classification, but for the best of our knowledge, this approach has not been previously evaluated in a non-thematic text classification task. Accordingly, we consider as an important contribution of this paper the pertinence evaluation of this approach on the posed task.</p><p>Thus, for applying the method described in <ref type="bibr" target="#b7">[8]</ref> we need to perform the following two steps: i) build the classes' graph, i.e., a prototype graph model for each category; and, ii) extract similarity features to train a supervised classifier, each instance is represented by 4 × N features (where N is equal to the number of categories, |C|).</p><p>As we mentioned, during the first step we create the prototype graph for each category c j ∈ C. Thus, we represent each document d i from the training set, belonging to category c j as the graph G ij . Notice that each node from G ij is a term (or token) w within the document d i , and the edges of G ij account for the number of co-occurrences of every pair of nodes (terms) in the same contextual window within d i . For the experiments reported in this paper we employed as terms character n-grams of size 3, and single words; and the size of the contextual window was defined as a symmetric window of 4 terms, meaning two terms to the left and two terms to the right from token w.</p><p>Once the graph for every document d has been generated, its necessary to merge all the graphs from documents of the same class c j , resulting in the prototype graph of the class. As established in <ref type="bibr" target="#b7">[8]</ref>, the prototype graph has the following properties: its edges (nodes) comprise the union of the edges (nodes) of the individual document graphs, and its weights are adjusted so that they converge to the mean value of the respective weights. At the end, the resulting prototype graph captures patterns common in the content and style of the entire category, such as recurring and neighboring character and word sequences.</p><p>For the second stage, the training phase, we need to compute the features for every new document in relation to the prototype graph models. Basically, we compute the similarity between documents and prototype graphs through the closeness of their respective graph representations. Similarly as the work described in <ref type="bibr" target="#b7">[8]</ref> we employed the three measures proposed by the author and we incorporated a content based metric. Next we describe the intuitive ideas behind each metric, however for further reference please follow <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>.</p><p>-Containment Similarity (CS). This metric expresses the proportion of edges of a graph G k that are shared with a second graph G l , i.e., sequences of shared nodes and edges.</p><p>-Value Similarity (VS). This measure indicates how many of the edges contained in graph G k are contained in graph G l , as well, considering also the weights of the matching edges. Notice that VS converges to its maximum value for graphs that share both the edges and the corresponding weights.</p><p>-Normalized Value Similarity (NVS). This is a derived measure from the previous metric, but without considering the relative size of the graphs being compared.</p><p>-Dice Similarity (DS). It it worth mentioning that this metric was not originally considered in <ref type="bibr" target="#b7">[8]</ref>. This metric accounts for the number of shared nodes between graphs.</p><p>As we mentioned before, every instance will be represented by a feature vector of size 4 × N ; where N is the number of categories in the classification problem and 4 similarity measures, namely: CS, VS, NVS and DS. This section is organized as follows: first we describe the provided dataset for performing our experiments; next we provide details on every system's configuration; then a brief description of the evaluation metrics, and finally, we discuss the obtained results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Dataset</head><p>The test collection for the pilot task was initially described in <ref type="bibr" target="#b5">[6]</ref>. Consist in a collection of posts from a set of social media users. There are two categories of users, depressed and non-depressed, and, for each user, the collection contains a sequence of writings in chronological order. In order to simulate an early detection scenario, the dataset was divided into 10 chunks. The first chunk contains the oldest 10% of the messages, the second chunk contains the second oldest 10%, and so forth. In summary, the training set contained 486 users (83 depressed, 403 non-depressed) and the test set contained 401 users (52 depressed, 349 non-depressed). Further details can be found in <ref type="bibr" target="#b6">[7]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Submitted runs</head><p>As we mentioned before, for generating the prototype graphs for each category (depressed and non-depressed ), we followed the methodology described in section 2 using two different forms for defining the terms w: character 3-grams and single words. It is important to mention that our generated graph models were constructed using the 100% of the provided training data, i.e., the prototype models do not consider the chronology of the writings. At this stage, we only wanted to evaluate if this type of representation was capable of modeling depressed blogs at any time. Next we briefly describe the general configuration of each of our competing system.</p><p>-LyRA: For this configuration, the prototype graph models are build using as w terms single words. The idea behind this approach was to evaluate if thematic correspondences (sequences of words) might be helpful in distinguishing the writing patterns of depressed users.</p><p>-LyRB: Contrary to the previous configuration, this systems employs character 3-grams as terms for the construction of the prototype graph models. The rationale idea behind using character n-grams is to represent content and style patterns in the graph model, characteristics that might result helpful when distinguishing among depressed users.</p><p>-LyRC: We refer to this configuration as an hybrid system given that each instance is represented by a feature vector of size 2 × 4 × N , i.e., 16 features extracted from LyRA and LyRB systems. With this setting we wanted to evaluate how complementary are both graph models (based on single words and character n-grams), in solving the posed task.</p><p>-LyRD: This is referred as a conservative ensemble method. This configuration considers the output of the LyRA, LyRB and LyRC classification systems, and assigns the class depressed if and only if the three systems agree on assigning this category. -LyRE: Similarly to the previous system, this is also an ensemble method but contrary to LyRD, this is a majority vote scheme among the decissions from LyRA, LyRB and LyRC. Another characteristic of this system is that it began to work until chunk 9, once enough information was accumulated in order to emit a more confident decision.</p><p>As described in <ref type="bibr" target="#b6">[7]</ref>, the test stage consisted of 10 sequential releases of data, and each participating system has to choose, for each user in the collection, between two options: (a) emitting a decision on the user (i.e. depressed or nondepressed), or (b) waiting to see more chunks. For all our proposed systems, if the classifier assigns the class "depressed" to an user in any chunk, we emitted the final decision, otherwise we choose the option "seeing more chunks". It is important to mention that all our experiments waited until the last chunk for emitting the "non-depressed" decision; this is, a user that is difficult to classify across chunks is at the very end (i.e., chunk 10) acknowledged as non-depressed users. Finally, as our classification algorithm we employed a lazy method, namely k-nearest neighbors algorithm, particularly we employed the provided implementation by the Weka toolkit <ref type="bibr" target="#b9">[10]</ref>, with k = 1 using an Euclidean distance<ref type="foot" target="#foot_3">4</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Evaluation</head><p>For reporting our obtained results we employ traditional set-based metrics such as F-measure, Precision and Recall. However, as described in <ref type="bibr" target="#b5">[6]</ref>, these metrics are time unaware, and for this reason we also report our results using the official ERDE metric. Intuitively, the ERDE metric considers the correctness of the (binary) decision and the delay taken by the system to make the decision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Results</head><p>Figure <ref type="figure" target="#fig_0">1</ref> shows the behaviour of our proposed systems across different chunks. For this analysis we do not consider as final decision the classification results of preceding chunks, i.e., we evaluate our classifiers performance at every chunk.</p><p>Several important aspects can be observed across systems, for instance, all configurations, except for LyRE, obtained good precision results between chunk 3 and chunk 6. In general, after chunk 5 (chunk 4 and 6 for the LyRC configuration), the performance decays very rapidly; however as more evidence (text from chunks) is obtained, the performance begins to recover up to chunk 10, which is an expected behaviour since our prototype graphs were built using the  entire set of posts from depressed/non-depressed users. Although more experiments need to be done, our explanation for the peaks reached between chunks 3 and 6 is due to highly distinctive characteristics of depressed users at particular points of their writings. For proving this hypothesis, we need to test our proposed method considering the chronological order of the texts, in other words, to generate prototype graphs for different depression stages (chunks).</p><p>Regarding the terms considered by the graphs, our experiments indicate that both models, single words and character n-grams, are to some extent complementary to each other, e.g., observe that the LyRC is able to obtain in a couple of times high precision results. Similarly, the conservative ensemble (LyRD) shows a more stable performance (in terms of precision) in comparison to LyRA and LyRB configurations.</p><p>In Table <ref type="table" target="#tab_0">1</ref> we observe the official results of our proposed systems during the eRisk competition. Notice that the last three rows from Table <ref type="table" target="#tab_0">1</ref> report the maximum, minimum and average values for each metric obtained from all the participating systems<ref type="foot" target="#foot_4">5</ref> . According to the ERDE metrics our best configuration was the LyRE system, however, as we mentioned before, this system began to emit decisions very late on the process, in chunk 9 (see Figure <ref type="figure" target="#fig_0">1</ref>). For this reason, we assume as our best configuration the LyRD system, which obtains minimum error rates and better precision values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions</head><p>In this paper, we have described the experiments performed by the Language and Reasoning Research Group from UAM Cuajimalpa in the context of the eRisk 2017 pilot task. Our proposed system was designed for addressing the posed problem by means of using as main form of representation graphs models; a graph model considers the order of terms' appearances in the original text, thus incorporating valuable contextual information to the representation.</p><p>Even though this type of representation has been evaluated on thematic text classification, our main goal was to determine its pertinence on non-thematic classification tasks. Thus, the main hypothesis is that through this representation we can capture patterns common in the content and style from depressed and non-depressed users, such as recurring and neighboring character and word sequences.</p><p>From this exercise we have learned that our proposed method performs better when character n-grams are employed to build the graph models. However, combining this representation with graph models based on single words allows to identify more depressed users in early stages, but not in general. In addition, our results indicate that depressed users seem to reach a climax point during the first chunks of its writing, and as the times passes, they return to more "traditional" writing. Nonetheless, we need to perform more experiments in order to validate this hypothesis, for instance, to evaluate our proposal at modeling different depression stages.</p><p>As future work we plan to incorporate more features to our representation, for instance POS tags and word n-grams, which could help to identify more elaborated regularities among users. We also plan to model the writing characteristics from users at different chunks individually. And finally we want to perform experiments using different classification methods.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Results for the positive class (depressed ) of our proposed systems across the 10 chunks. We report Recall, Precision and F-measure obtained individually for every chunk.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Official results during the eRisk competition. Last three rows represent the maximum (MAX), minimum (MIN) and Average values obtained across all participating systems.</figDesc><table><row><cell>System</cell><cell>ERDE5</cell><cell cols="4">ERDE50 F-measure Precision Recall</cell></row><row><cell>LyRA</cell><cell>15.65%</cell><cell>15.15%</cell><cell>14.00%</cell><cell>11.00%</cell><cell>19.00%</cell></row><row><cell>LyRB</cell><cell>16.75%</cell><cell>15.76%</cell><cell>16.00%</cell><cell>11.00%</cell><cell>29.00%</cell></row><row><cell>LyRC</cell><cell>16.14%</cell><cell>15.51%</cell><cell>16.00%</cell><cell>12.00%</cell><cell>25.00%</cell></row><row><cell>LyRD</cell><cell>14.97%</cell><cell>14.47%</cell><cell>15.00%</cell><cell>13.00%</cell><cell>17.00%</cell></row><row><cell>LyRE</cell><cell>13.74%</cell><cell>13.74%</cell><cell>8.00%</cell><cell>11.00%</cell><cell>6.00%</cell></row><row><cell>MAX</cell><cell>19.14%</cell><cell>17.15%</cell><cell>64.00%</cell><cell>69.00%</cell><cell>92.00%</cell></row><row><cell>MIN</cell><cell>12.70%</cell><cell>9.68%</cell><cell>8.00%</cell><cell>11.00%</cell><cell>6.00%</cell></row><row><cell>Average</cell><cell>14.67%</cell><cell>12.76%</cell><cell>39.77%</cell><cell>36.50%</cell><cell>51.17%</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Conference Labs of the Evaluation Forum: http://clef2017.clef-initiative.eu</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://early.irlab.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Unless explicitly mention, a term can be either a character n-gram or word n-gram.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">http://www.cs.waikato.ac.nz/ml/weka/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">Further details on each participating system can be found in<ref type="bibr" target="#b6">[7]</ref>.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgments. This work was partially funded by CONACYT under project grant number 258588. We also thank to UAM Cuajimalpa for the provided support.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of the international sexual predator identification competition at pan-2012</title>
		<author>
			<persName><forename type="first">G</forename><surname>Inches</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF (Online working notes/labs/workshop)</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">30</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A two-step approach for effective detection of misbehaving users in chats</title>
		<author>
			<persName><forename type="first">E</forename><surname>Villatoro-Tello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Juarez-Gonzalez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Escalante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes-Y-Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Villaseñor-Pineda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF (Online Working Notes/Labs/Workshop</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Forner</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Karlgren</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Womser-Hacker</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Sexual predator detection in chats with chained classifiers</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Escalante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Villatoro-Tello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Juarez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes-Y-Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Villaseñor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis</title>
				<meeting>the 4th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis<address><addrLine>Atlanta, Georgia</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="46" to="54" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">-de-la Rosa</title>
		<author>
			<persName><forename type="first">Á</forename><surname>Callejas-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Villatoro-Tello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Meza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ramírez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">From Dialogue Corpora to Dialogue Systems: Generating a Chatbot with Teenager Personality for Preventing Cyber-Pedophilia</title>
				<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="531" to="539" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Follower behavior analysis via influential transmitters on social issues in twitter</title>
		<author>
			<persName><forename type="first">K.-Y</forename><surname>Jeong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-S</forename><surname>Le</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computacion y Sistemas</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A test collection for research on depression and language use</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Losada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference Labs of the Evaluation Forum</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">eRISK 2017: CLEF Lab on Early Risk Prediction on the Internet: Experimental foundations</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Losada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Parapar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings Conference and Labs of the Evaluation Forum CLEF 2017</title>
				<meeting>Conference and Labs of the Evaluation Forum CLEF 2017<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Representation models for text classification: A comparative analysis over three web document types</title>
		<author>
			<persName><forename type="first">G</forename><surname>Giannakopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mavridi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Paliouras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Papadakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tserpes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2Nd International Conference on Web Intelligence, Mining and Semantics, WIMS &apos;12</title>
				<meeting>the 2Nd International Conference on Web Intelligence, Mining and Semantics, WIMS &apos;12<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">12</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Summarization system evaluation revisited: N-gram graphs</title>
		<author>
			<persName><forename type="first">G</forename><surname>Giannakopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Karkaletsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Vouros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stamatopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Speech Lang. Process</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page">39</biblScope>
			<date type="published" when="2008-10">Oct. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Weka: The waikato environment for knowledge analysis</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R</forename><surname>Garner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the New Zealand Computer Science Research Students Conference</title>
				<meeting>of the New Zealand Computer Science Research Students Conference</meeting>
		<imprint>
			<date type="published" when="1995">1995</date>
			<biblScope unit="page" from="57" to="64" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
