<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">An Automated Annotation Process for the SciDocAnnot Scientific Document Model</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Hélène</forename><surname>De Ribaupierre</surname></persName>
							<email>helene.deribaupierre@unige.ch</email>
							<affiliation key="aff0">
								<orgName type="laboratory">CUI</orgName>
								<orgName type="institution">University of Geneva</orgName>
								<address>
									<addrLine>7, route de Drize</addrLine>
									<postCode>CH, 1227</postCode>
									<settlement>Carouge</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Oxford</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gilles</forename><surname>Falquet</surname></persName>
							<email>gilles.falquet@unige.ch</email>
							<affiliation key="aff0">
								<orgName type="laboratory">CUI</orgName>
								<orgName type="institution">University of Geneva</orgName>
								<address>
									<addrLine>7, route de Drize</addrLine>
									<postCode>CH, 1227</postCode>
									<settlement>Carouge</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">An Automated Annotation Process for the SciDocAnnot Scientific Document Model</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">4711FC645B29B4A0634529FA38D6D513</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T22:25+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Answering precise and complex queries on a corpus of scientific documents requires a precise modelling of the document contents. In particular, each document element must be characterised by its discourse type (hypothesis, definition, result, method, etc.). In this paper we present a scientific document model (SciAnnotDoc) that takes into account the discourse types. Then we show that an automated process can effectively analyse documents to determine the discourse type of each element. The process, based on syntactic rules (patterns), has been evaluated in terms of precision and recall on a representative corpus of more than 1000 articles in Gender studies. It has been used to create a SciDocAnnot representation of the corpus on top of which we built a faceted search interface. Experiments with users show that searching with this interface clearly outperforms standard keyword search for complex queries.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>One of the challenges, today, for Information Retrieval System for Scientific Document is to fulfil the information needs of scientists. For scientists, being aware of others' work and publications in the world is a crucial task, not only to stay competitive but also to build their work upon already proven knowledge. In 1945, Bush <ref type="bibr" target="#b1">[2]</ref> already argued that too many publications can be a problem because the information contained in these publications cannot reach other scientists. Bush expounded his argument using the example of Mendel's laws of genetics. These laws were lost to the world for a generation because the Mendel's publication did not reach the few people who were capable of understanding and extending this concept. Today this problem is even more important with the exponential growth of literature in all domains (e.g., Medline has a growth rate of 0.5 million items per year <ref type="bibr" target="#b7">[8]</ref>). Today's IR systems are not able to answer precisely to queries such as "find all the definition of the term X" or "find all the findings that analyse why the number of women in academics falls more sharply than the number of men after their first child, using qualitative and quantitative methodologies". These systems are in general using only the metadata of the documents to index them (title, author(s), keywords, abstract, etc.), but to obtain systems that answer to such precise queries, we need to have very precise semantic annotation of the entire documents. In <ref type="bibr" target="#b12">[13]</ref>, <ref type="bibr" target="#b14">[15]</ref>, we have proposed a new annotation model for scientific document (SciAnnotDoc annotation model). The SciAnnotDoc (see Figure <ref type="figure" target="#fig_0">1</ref>) annotation model is a generic model for scientific documents. This model can be decomposed in four different dimensions or facets:</p><p>1. Conceptual dimension: Ontologies or controlled vocabularies that describe scientific terms (the SciDeo ontology) or concepts used in the document (conceptual indexing) 2. Meta-data dimension: description of meta-data document information (bibliographic notice) 3. Rhetorical or discursive dimension: description of the discursive role played by each document element 4. Relationships dimension: description of the citations and relationships between documents</p><p>The third facet is extremely important when considering precise scientific queries and is decomposed into five discourse elements types that are: findings, hypothesis, definition, methodology and related work. We retained these five discourses elements after analysing the result of a survey and interviews made with scientists in different field of research to determine what and how scientists are searching and reading scientific documents <ref type="bibr" target="#b11">[12]</ref>, <ref type="bibr" target="#b13">[14]</ref>. The SciAnnotDoc model is implemented in OWL. The ontology contains 69 classes, 137 object properties and 13 datatype properties (counting those imported from CiTO<ref type="foot" target="#foot_1">3</ref>  <ref type="bibr" target="#b17">[18]</ref>). The model also integrate ontologies that help in the annotation process (the violet ontologies), and that are given more information about the content such as domain concept, scientific object or methods names contained in the different discourse element. In this paper, we present the automatic annotation processes we used to annotate scientific documents using the SciAnnotDoc model. The process is based on natural language processing (NLP) techniques.</p><p>To evaluate the annotation process, we used a corpus in gender studies. We chose this domain because it consists of very heterogeneous written documents ranging from highly empirical studies to "philosophical" texts, and these documents are less structured than in other fields of research (i.e. medicine, biomedicine, physic, etc.) and rarely use the IMRaD model (introduction, methods, results and discussion). This corpus is therefore more difficult to annotate than a corpus of medical documents, which is precisely the kind of challenge we were looking for. We argue that if our annotation process can be applied to such a heterogeneous corpus, it should also apply to other, more homogeneous types of papers. Therefore, the annotation process should be generalisable to other domains.</p><p>In the literature, there are three types of methods for the automatic annotation or classification of scientific text documents. The first type is rule based. This type of system is based on the detection of general patterns in sentences <ref type="bibr" target="#b3">[4]</ref>, <ref type="bibr" target="#b19">[20]</ref>, <ref type="bibr" target="#b16">[17]</ref>. Several systems are freely available, such as XIP<ref type="foot" target="#foot_2">4</ref> , EXCOM <ref type="foot" target="#foot_3">5</ref> , and GATE <ref type="foot" target="#foot_4">6</ref> . The second method is based on machine learning and requires a training corpus, such as the systems described by <ref type="bibr" target="#b9">[10]</ref>, <ref type="bibr" target="#b18">[19]</ref>, <ref type="bibr" target="#b15">[16]</ref>, <ref type="bibr" target="#b5">[6]</ref>. Several classifiers are available, such as the Stanford Classifier<ref type="foot" target="#foot_5">7</ref> , Weka<ref type="foot" target="#foot_6">8</ref> , and Mahout<ref type="foot" target="#foot_7">9</ref> , based on different algorithms (Decision Trees, Neural Networks, Naïve Bayes, etc., <ref type="foot" target="#foot_8">10</ref> ). The third method is a hybrid between the two aforementioned systems <ref type="bibr" target="#b8">[9]</ref>.</p><p>In this work, we opted for a rule-based system because we did not have a training corpus and because documents in the human sciences are generally less formalised than in other domains, it may be difficult to have sufficient features with which to distinguish the different categories. Between the several free semantic annotation tools existing, we choose to use GATE, because it is used by a very large community and several plug-ins are available.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Annotation Implementation</head><p>The annotation processes transforms each sentence into a discourse element (or if it is not one of the five discourse element into a non defined discourse element) and a paragraph into a fragment. Each fragment contains one to many discourses elements and each sentence can be attributed to one or many discourse elements (e.g. a sentence that describe a definition can be also a sentence that describe a finding). The following sentence will be annotated as a definition and a finding. "We find, for example, that when we use a definition of time poverty that relies in part on the fact that an individual belongs to a household that is consumption poor, time poverty affects women even more, and is especially prevalent in rural areas, where infrastructure needs are highest." <ref type="bibr" target="#b0">[1]</ref> The discourse element related work is a special case, it will always be defined at first as one of the four other discourses elements and defined as a related work thereafter. The reason for this kind of annotation is the results of the analyses we made from the interviews. Scientists sometimes are looking for a finding, a definition, a methodology or a hypothesis but the attribution of the document to an author is not in their priority, it is only later that they might be interested to know who the author of the document are or the referenced sentences. For example, the following sentence is a finding and in a second time a related work, as this sentence refer to another works.</p><p>"The results of the companion study <ref type="bibr">(Correll 2001)</ref>, and less directly the results of Eccles (1994; Eccles et al. 1999), provide evidence that is consistent with the main causal hypothesis that cultural beliefs about gender differentially bias men and women's self-assessments of task competence". <ref type="bibr" target="#b2">[3]</ref> In a first step, to discover and analyse the syntactic patterns of the discourse elements, we manually extracted sentences that corresponded to the different discourse elements from scientific documents in two areas of research: computer science and gender studies. In a second step, we uploaded these sentences in GATE and ran a pipeline composed of components included in ANNIE (ANNIE Tokeniser component, the ANNIE sentence splitter component and the ANNIE part-of-speech component) and obtained the syntactic structures of these sentences. The aim of this analyse was to build the JAPE rules for detecting the different discourses elements. The methodology used to create the rules was the following. First, we started to look at the syntactic structure produced by the ANNIE output for each of the different sentences. The following example (see table <ref type="table" target="#tab_0">1</ref>) describes the entire tag's sequence obtained by ANNIE on the following definition of the term "gender" (for space reason, we didn't write down all the sequence 11 ).</p><p>"On this usage, gender is typically thought to refer to personality traits and behavior in distinction from the body". <ref type="bibr" target="#b6">[7]</ref> For each tag's sequence, we simplified those rules, reduced them and merged some of them, to obtain more generic rules able, not only to catch the very specific syntactic pattern, but also to catch the variation of the pattern. We also 11  relaxed the rules and used some unspecified token (see table <ref type="table" target="#tab_1">2</ref>). To increase the precision, we added typical terms that appear in each type of discourse element.</p><p>For the definition example, instead of using the tag VBZ <ref type="foot" target="#foot_9">12</ref> that could be too generic, we used a macro that was the inflection of the verb be and have in the singular and plural form. We also used a macro that defined the different inflection of the verb refer to. With this simplification and relaxation, we can now annotate sentences such as those shown in the table 2. We uploaded the domain concept ontology to help to define more precise rules. For example, to detect a definition such as "It follows then that gender is the social organization of sexual difference" <ref type="bibr" target="#b6">[7]</ref>; we created a rule that was searching for a concept defined in the domain ontology followed at a short distance by the declension of the verb be. To be able to use the different ontologies, we used the ontologies plug-in contained in GATE <ref type="foot" target="#foot_10">13</ref> . We imported the different ontologies that we created to help the annotation process: the gender studies ontology (GenStud), the scientific ontology (SciObj) and the methodology ontology (SciMeth). The ontologies were used not only for the JAPE rules but also to annotate the concepts in the text. With this methodology we defined 20 rules for findings, 34 rules for definitions, 11 rules for hypothesis and 19 rules for methodologies and 10 rules for the referenced sentences.</p><p>We automatically annotated 1,400 documents in English from various journals in gender and sociological studies. The first step consisted of transforming a PDF file into a raw text. PDF is the most frequently used format to publish scientific documents, but it is not the most convenient one to transform into raw text. The Java program implemented to transform the PDF into raw text used the PDFbox<ref type="foot" target="#foot_11">14</ref> API and regular expressions to clean the raw text. Second, we applied the GATE pipeline (see figure <ref type="figure" target="#fig_2">2</ref>). The output given by GATE is a XML file. Third, we implemented a Java application (see figure <ref type="figure" target="#fig_3">3</ref>) using the OWL API to transform the GATE's XML files into an RDF representation of the text. Each XML tag corresponding to concept or object properties in the ontologies were transformed. The sentences that did not contain one of the four discourse elements (definition, hypothesis, finding or methodology) were annotated with the tag &lt;NonDefinedDE&gt;, allowing for the annotation of each sentence of the entire document, even those not assigned to discourse elements. Each discourse element that contained a tag with &lt;AuthorRefer&gt; were defined as a related work. The different RDF representations created by the Java application were loaded into an RDF triple store. We chose Allegrograph<ref type="foot" target="#foot_12">15</ref> because it supports RDF S + + reasoning in addition to SPARQL query execution.</p><p>The table <ref type="table" target="#tab_2">3</ref> presents the distribution of the discourse elements by journal. We can observe that the number of referenced documents is greater than the number of related works. This result is observed because authors generally refer to several documents simultaneously rather than to a single one. We can also observe that the most fund discourse element is the finding, followed by methodology, followed by hypothesis and definition. This distribution seems to follow the hypothesis that scientist communicate more their finding than everything else, even in field of research such as sociology or gender Studies. And it is also the most researched discourse element that from the survey and the interviews scientists are looking for <ref type="bibr" target="#b11">[12]</ref>, <ref type="bibr" target="#b13">[14]</ref>. To test the quality of the patterns, we uploaded 555 manually annotated sentences that constitute our gold standard into GATE and processed through the same pipeline (see Figure <ref type="figure" target="#fig_2">2</ref>). We did not use any of the sentences analysed to create the JAPE rules to construct the gold standard to avoid bias. We performed measurements of precision and recall on these sentences (see Table <ref type="table" target="#tab_3">4</ref>). The results indicated good precision but a lower recall. One of the reasons of the lower recall could be that the JAPE rules are very conservative. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">User evaluation on complex queries</head><p>We conducted user evaluations to check how the annotation system and the SciDocAnnot model compare to standard keyword search. We implemented two interactive search interfaces: a classic keywords based search (with a TF*IDF based weighting scheme) and a faceted interface (FSAD) based on our model (facets correspond to the types of discourse elements). Both systems are indexing and querying at the sentence level (instead of the usual document level). The first tests we conducted with 8 users (scientists, 4 in gender studies and 50% women with an average age of 38 years old). Scientists had to perform 3 tasks with only one of the system (see below). The design of the experiment was based on a Latin square rotation of tasks to control for a possible learning effect of the interface on the participants.</p><p>task 1 Find all the definitions of the term "feminism". task 2 Show all findings of studies that have addressed the issue of gender inequality in academia. task 3 Show all findings of studies that have addressed the issue of gender equality in terms of salary.</p><p>We gave them a small tutorial on how the system works, but didn't give more exact instruction on how to search. The participants, who decided whether they had obtained enough information on the given subject, determined the end of a task. They have to perform the task and complete 4 different questionnaire (1 socio-demographic, 1 after each task and a final at the end of the evaluation, for more precision about the questionnaires see <ref type="bibr" target="#b10">[11]</ref>). The questionnaire after each task contained 10 questions and the last questionnaire 11 questions; most of the questions were using a Likert scale. The evaluation was performed in French. The questionnaires were conducted on LimeSurvey. We computed the average response for the three tasks and we tested the difference between the participants who had to evaluate the FSAD versus the keyword search, using an analysis of the variance (Anova) tests. Because of the lack of space, in this paper, we will only present a part of the evaluation.</p><p>The first question (Do you think the set of results was relevant to the task? 1=not useful, 5=useful) was about the relevance of the set of results,. We didn't observe any significant difference between the two groups of user, both fund that the set of answer was useful (FSAD M=4.0; keywords M=3.75). But in the second question (Do you think the number of results was too large to be useful? 1 = totally unusable, 5 = usable), about the irrelevance of the question, the keywords-search group (M=2.75) find the irrelevance of the set of answer more important than the FSAD group (M=4.5), and a significant difference was observed (p&lt;0.05) between the two groups. We also asked the same question, but instead to have to answer with a Likert scale, users have to answer with a scale in percent (How many elements correspond to your request? 1 = 0-5%, 2= 6-15%, 3= 16-30%, 4=31-50%, 5=51-75%, 6=76-90%7,+90%). Again we didn't find a significant difference in this question between the group (FSAD M=5.0; keywords M=4.25), but for the second question, we again find a significant difference between the two group (FSAD M=1.5; keywords M=3.75; p&lt;0.05). When we asked the users for the level of satisfaction they experimented with the set of results (Did you obtain satisfactory results for each query you made? 1 = not at all satisfied, 5 = very satisfied), the difference between the two groups is not significant (FSAD M=3.83; Keywords M=3.41). The next question was about the level of satisfaction overall the set of results for the whole task (Are you satisfied with the overall results provided? 1 = not at all satisfied, 5 = completely satisfied). Most of the users did more than one query by task. The participants who used the keyword search interface seemed to be less satisfied with the overall results than the participants who used the FSAD, but the difference was not significant (FSAD M=4.16, Keywords M=3.41). We also ask the user about their level of frustration to the set of results (Are you frustrated by the set(s) of results provided? 1 = totally frustrated, 5 = not at all frustrated). The FSAD group seemed to be less frustrated with the set of results than the keyword search interface group, but the difference was not significant (FSAD M=3.16; keywords M=4.25).</p><p>Aside from the user evaluation, we also performed a precision and recall evaluation for the first task. For the FSAD system, when the user choose the facet definition and type the keyword feminism, the system send a set of 148 answers and 90 were relevant, the precision was 0.61. For the keywords-search system, the set of result was the combination of the term "define AND feminism" and "definition AND feminism" (this combination of term was the one the most used by users for the task 1), the system sent a set of 29 answers of which 24 where relevant), the precision was 0.82.</p><p>For the recall, as we didn't know the number of definitions contained in the corpus, we simply observed that the ratio between FSAD and the keyword search is 3.77. In other words, the FSAD system was able to find 3.77 times more definitions of the term "gender" than the keywords-search. However even if the precision is slightly lower in the FSAD system than in the keywords-search system, the FSAD system has a considerably higher recall than the keyword search system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>The aim of this work is to propose an approach to help scientists to find the documents they need for their works. As presented in the introduction, it is important for scientists to have a search engine that is able to answer to precise questions such as "retrieve all the findings that women have a tendency to drop their academic carrier after their first child more than men, using qualitative and quantitative methodologies". In this case, knowing or indexing only the metadata is not enough, and annotation about the content of the full text such as the discourse element, the references to other documents and the concept is crucial. In this paper, we have proposed an approach to automatically annotated PDF document with the SciAnnotDoc model. The evaluation of the annotation show not only that the model is realistic because it is amenable to automatic production (many previously proposed annotation models have never been used in practice because the require a manual annotation, see <ref type="bibr" target="#b10">[11]</ref>, for a more complete review of the different systems and model), but also that the precision is good.</p><p>To improve the recall index, a solution could be to create more JAPE rules. However, introducing a larger number of rules might also increase the risk of adding noise to the annotation. Another solution could be to test if some hybrid approach mixing a rules-based approach and a machine-learning approach may improve precision and recall. An other solution is to ask experts not to classify the sentence in categories, but to confirm the type of categories a sentence is already classify into. By using this kind of methodology, we can improve and enlarge a training corpus that we could use to improve the precision and recall of the actual annotation process.</p><p>The evaluation with users shows that despite these inaccuracies and a small sample, we were able to build a query system that already outperforms keyword search in many cases, especially in the case where the recall is very important. Google allow to query a term for the definition with "define" + the term. In Google the set of answer is extracted from glossaries, dictionaries and Wikipedia for the first ranked answer, and for the next answers the system seems to work by looking at the pattern "define"+ term. For scientists this is not enough, first because of the source of the information is not accurate enough and second because of the lack of answers. For Google Scholar, scientists make the assumption that the sources are more accurate because the IRs is indexing scientific documents. The system query the index with the pattern "define" AND "feminism", ignoring all the other definition that use some other sentence construction than ".... define feminism...". And as we have shown above, the number of definition of the term fund with this pattern is from a very long shot not enough, especially for scientists. By consequence, when the task is to find a definition and the user need a very high recall, Google or Google Scholar are not performing well. One of the difficulty we have to deal in the evaluation was the lack of a good evaluation corpus, helping us to calculate the precision and the recall of the system. This problem is very often mentioned in the literature, conference and workshops. We hope than in the future, with the different evaluation campaign that was created these last years, this recurrent problem should diminish.</p><p>The user evaluation also shown that the user seems to be less frustrated by the FSAD system than the keywords-search and seems to find that the level of irrelevance in the set of results is less important in FSAD than in keywordssearch. Some of the results could be non significant because of the size of the sample that is a little bit too small.</p><p>In the case of very precise query such as the task 2 or 3 we still have to analyse the precision and recall of our system. We also want to compare the result with some today IR's, but we can hypothesis that in contrary of the first task, it will be the precision that will be missing because it is not searching at the sentence level. The reason is that Google and similar system are indexing the text by the terms they find in the metadata (title, abstract, keywords) and sometimes by the terms contained in the entire document, but they don't take into account the context of the term, or even the distance between the terms. For example, in the task 2, users will certainly types as keywords "academic" or "university" and "gender inequality", but the problem is that those terms could appear everywhere in the text, even in the references, and document that have for example a reference that was publish in the Oxford University Press and contains in an other part of the text "gender inequality" could appear in the top ranked answers.</p><p>In the future, we will conduct some additional usability testing and collect data to scientifically assess the quality of the system and to determine the influence of the precision/recall of the automated annotation process on the system performance. We will also conduct some experiment to analyse which kind of task is more demanding of a good precision, versus the one that need a good recall.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. SciAnnotDoc model</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>(</head><label></label><figDesc>({Lookup.classURI==".../genderStudies.owl#concept"}) (PUNCT)? (VERBE_BE))</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2</head><label>2</label><figDesc>Fig. 2. Gate Pipeline</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Annotation algorithm model</figDesc><graphic coords="7,193.43,209.89,138.32,58.57" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>ANNIE tag's sequence</figDesc><table><row><cell>the</cell><cell>definition</cell><cell>of</cell><cell>the</cell><cell>part-of-speech</cell><cell>tags</cell><cell>can</cell><cell>be</cell><cell>found</cell><cell>at</cell></row><row><cell cols="6">http://gate.ac.uk/sale/tao/splitap7.html#x39-789000G</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Definition sentences and JAPE rules (simplified)</figDesc><table><row><cell>gender</cell><cell>is</cell><cell>typically thought to</cell><cell>refer to</cell></row><row><cell>gender</cell><cell>has</cell><cell>become used to</cell><cell>refer to</cell></row><row><cell>gender</cell><cell>was</cell><cell>a term used to</cell><cell>refer to</cell></row><row><cell cols="2">NN TO BE HAVE(macro)</cell><cell>Token[2,5]</cell><cell>REFER(macro)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Annotated corpus statistics by Discourse elements</figDesc><table><row><cell>Journal Name</cell><cell>Def.</cell><cell cols="4">Find. Hypo. Meth. Related</cell><cell>Referenced</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>Work</cell><cell>documents</cell></row><row><cell>Gender and Society</cell><cell>745</cell><cell cols="4">2945 1021 1742 986</cell><cell>4855</cell></row><row><cell>Feminist Studies</cell><cell cols="5">2201 4091 2545 3660 177</cell><cell>5377</cell></row><row><cell>Gender Issues</cell><cell>280</cell><cell cols="2">1126 414</cell><cell>611</cell><cell>267</cell><cell>1566</cell></row><row><cell>Signs</cell><cell>789</cell><cell cols="2">1566 712</cell><cell cols="2">1221 516</cell><cell>3129</cell></row><row><cell>American Historical Re-</cell><cell>97</cell><cell>219</cell><cell>87</cell><cell>170</cell><cell>15</cell><cell>440</cell></row><row><cell>view</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>American Journal Of</cell><cell cols="5">1776 10160 4316 6742 2907</cell><cell>13323</cell></row><row><cell>Sociology</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Feminist economist</cell><cell cols="5">1381 6940 2025 4169 2288</cell><cell>9600</cell></row><row><cell>Total</cell><cell cols="5">7269 27047 11120 18315 7156</cell><cell>38290</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Precision/recall values</figDesc><table><row><cell>Discourse element type</cell><cell cols="2">No of sentences Precision</cell><cell cols="2">Recall F1.0s</cell></row><row><cell>Findings</cell><cell>168</cell><cell>0.82</cell><cell>0.39</cell><cell>0.53</cell></row><row><cell>Hypothesis</cell><cell>104</cell><cell>0.62</cell><cell>0.29</cell><cell>0.39</cell></row><row><cell>Definitions</cell><cell>111</cell><cell>0.80</cell><cell>0.32</cell><cell>0.46</cell></row><row><cell>Methodology</cell><cell>172</cell><cell>0.83</cell><cell>0.46</cell><cell>0.59</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Proceedings of the 5th International Workshop on Semantic Digital Archives(SDA 2015)   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">CiTO is used to describe the different type of citation or reference between the documents or discourse element</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://open.xerox.com/Services/XIPParser</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">http://www.excom.fr</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">http://gate.ac.uk</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">http://nlp.stanford.edu/software/classifier.shtml</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6">http://www.cs.waikato.ac.nz/ml/weka/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_7">https://mahout.apache.org/users/basics/algorithms.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_8">see<ref type="bibr" target="#b4">[5]</ref> for a complete review of the different algorithms Proceedings of the 5th International Workshop on Semantic Digital Archives(SDA 2015)   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_9">3rd person singular present</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_10">Ontology OWLIM2, OntoRoot GazetteerProceedings of the 5th International Workshop on Semantic Digital Archives(SDA 2015)   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_11">https://pdfbox.apache.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_12">http://franz.com/agraph/allegrograph/ Proceedings of the 5th International Workshop on Semantic Digital Archives (SDA 2015)</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Acknowledgments</head><p>This work is supported by the Swiss National Fund (200020 138252)   </p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Working long hours and having no choice: time poverty in guinea</title>
		<author>
			<persName><forename type="first">E</forename><surname>Bardasi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Wodon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Feminist Economics</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="45" to="78" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">As we may think</title>
		<author>
			<persName><forename type="first">V</forename><surname>Bush</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The atlantic monthly</title>
		<imprint>
			<biblScope unit="volume">176</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="101" to="108" />
			<date type="published" when="1945">1945</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Constraints into preferences: Gender, status, and emerging career aspirations</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Correll</surname></persName>
		</author>
		<ptr target="http://asr.sagepub.com/content/69/1/93.abstract" />
	</analytic>
	<monogr>
		<title level="j">American Sociological Review</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="93" to="113" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Towards automatic extraction of epistemic items from scientific publications</title>
		<author>
			<persName><forename type="first">T</forename><surname>Groza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Handschuh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bordea</surname></persName>
		</author>
		<idno type="DOI">10.1145/1774088.1774377</idno>
		<ptr target="http://doi.acm.org/10.1145/1774088.1774377" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2010 ACM Symposium on Applied Computing</title>
				<meeting>the 2010 ACM Symposium on Applied Computing<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="1341" to="1348" />
		</imprint>
	</monogr>
	<note>SAC &apos;10</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Supervised machine learning: A review of classification techniques</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">B</forename><surname>Kotsiantis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Informatica</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="249" to="268" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Automatic recognition of conceptualization zones in scientific articles and two life science applications</title>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dobnik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Batchelor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rebholz-Schuhmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="991" to="1000" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Interpreting gender</title>
		<author>
			<persName><forename type="first">L</forename><surname>Nicholson</surname></persName>
		</author>
		<ptr target="http://www.jstor.org/stable/3174928" />
	</analytic>
	<monogr>
		<title level="j">Signs</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="79" to="105" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">CORAAL-Dive into publications, bathe in the knowledge</title>
		<author>
			<persName><forename type="first">V</forename><surname>Nováček</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Groza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Handschuh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Decker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Web Semantics: Science, Services and Agents on the World Wide Web</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="176" to="181" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Aggregating search results for social science by extracting and organizing research concepts and relations</title>
		<author>
			<persName><forename type="first">S</forename><surname>Ou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S G</forename><surname>Kho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR 2008 Workshop on Aggregated Search</title>
				<meeting><address><addrLine>Singapour</addrLine></address></meeting>
		<imprint>
			<publisher>Singapour</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Identifying comparative claim sentences in full-text scientific articles</title>
		<author>
			<persName><forename type="first">D</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Blake</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">50th Annual Meeting of the Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Ribaupierre</surname></persName>
		</author>
		<title level="m">Precise information retrieval in semantic scientific digital libraries</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
		<respStmt>
			<orgName>University of Geneva</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">New trends for reading scientific documents</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Ribaupierre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Falquet</surname></persName>
		</author>
		<idno type="DOI">10.1145/2064058.2064064</idno>
		<ptr target="http://doi.acm.org/10.1145/2064058.2064064" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th ACM workshop on Online books, complementary social media and crowdsourcing</title>
				<meeting>the 4th ACM workshop on Online books, complementary social media and crowdsourcing<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="19" to="24" />
		</imprint>
	</monogr>
	<note>BooksOnline &apos;11</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A user-centric model to semantically annotate and retrieve scientific documents</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Ribaupierre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Falquet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the sixth international workshop on Exploiting semantic annotations in information retrieval</title>
				<meeting>the sixth international workshop on Exploiting semantic annotations in information retrieval</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="21" to="24" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Un modèle d&apos;annotation sémantique centré sur les utilisateurs de documents scientifiques: cas d&apos;utilisation dans les études genre</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Ribaupierre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Falquet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IC-25èmes Journées francophones d&apos;Ingénierie des Connaissances</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="99" to="104" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">User-centric design and evaluation of a semantic annotation model for scientific documents</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Ribaupierre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Falquet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">13th International Conference on Knowledge Management and Knowledge Technologies, I-KNOW &apos;14</title>
				<meeting><address><addrLine>Graz, Austria</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">September 16-19, 2014. 2014</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Using argumentation to extract key sentences from biomedical abstracts</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ruch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Boyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chichester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Tbahriti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Geissbühler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fabry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gobeill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pillet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rebholz-Schuhmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lovis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Veuthey</surname></persName>
		</author>
		<ptr target="http://www.sciencedirect.com/science/article/pii/S1386505606001183,connect-ingMedicalInformaticsandBio-Informatics-{MIE}2005" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Medical Informatics</title>
		<imprint>
			<biblScope unit="volume">76</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="195" to="200" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Detecting key sentences for automatic assistance in peer reviewing research articles in educational sciences</title>
		<author>
			<persName><forename type="first">Á</forename><surname>Sándor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vorndran</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2009 Workshop on Text and Citation Analysis for Scholarly Digital Libraries</title>
				<meeting>the 2009 Workshop on Text and Citation Analysis for Scholarly Digital Libraries</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="36" to="44" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">CiTO, the Citation Typing Ontology, and its use for annotation of reference lists and visualization of citation networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Shotton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Bio-Ontologies 2009 Special Interest Group meeting at ISMB</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Argumentative zoning: Information extraction from scientific text</title>
		<author>
			<persName><forename type="first">S</forename><surname>Teufel</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
		<respStmt>
			<orgName>University of Edinburgh</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Unpublished PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Autour du projet scientext: étude des marques linguistiques du positionnement de l&apos;auteur dans les écrits scientifiques</title>
		<author>
			<persName><forename type="first">A</forename><surname>Tutin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Grossmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Falaise</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kraif</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journées Linguistique de Corpus</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">12</biblScope>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
