<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">An attempt to combine features in classifying argument components in persuasive essays</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Yunda</forename><surname>Desilia</surname></persName>
							<email>yunda.desilia@binus.ac.id</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">Bina Nusantara University Jakarta</orgName>
								<address>
									<country key="ID">Indonesia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Thasya</forename><surname>Velizya</surname></persName>
							<email>velizya.utami@binus.ac.id</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">Bina Nusantara University Jakarta</orgName>
								<address>
									<country key="ID">Indonesia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cecilia</forename><surname>Utami</surname></persName>
							<email>cecilia.arta@binus.ac.id</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">Bina Nusantara University Jakarta</orgName>
								<address>
									<country key="ID">Indonesia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Derwin</forename><surname>Arta</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">Bina Nusantara University Jakarta</orgName>
								<address>
									<country key="ID">Indonesia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><surname>Suhartono</surname></persName>
							<email>dsuhartono@binus.edu</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">Bina Nusantara University Jakarta</orgName>
								<address>
									<country key="ID">Indonesia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">An attempt to combine features in classifying argument components in persuasive essays</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B56B7AA932C029DD6F0037041ED5D955</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T04:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>argument component</term>
					<term>feature</term>
					<term>word embedding</term>
					<term>argumentation mining</term>
					<term>persuasive essay</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>So far, several approaches have been done in detecting and classifying argumentation in persuasive essays. In this paper, we proposed some new features on top of the state-of-the-art researches in argumentation mining. We grouped 68 features into 8 categories; they are structural, lexical, indicators, contextual, syntactic, prompt similarity, word embedding, and discourse features. Instead of handcrafted features, we utilized word embedding as the feature. At the end of this paper, we presented the comparison between each group of features to classify the argument components. 402 persuasive essays were utilized. We found that structural features were the most significant feature while discourse features were not. After combining all features, we obtained 79.96% as the accuracy; it was slightly outperforming the state-ofthe-art accuracy which was 77.3%.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>Argumentation is a process of building arguments, exchanging arguments, and evaluating arguments in terms of interaction with the other arguments. An argument is a set of premises or evidence/fact, which are given to support the claim <ref type="bibr" target="#b10">(Palau and Moens, 2009)</ref>. The objective of argumentation is to make the audiences believe the idea, thought, or opinion stated are true and proved. Argumentation mining aims to detect the arguments in a text document, relation between them, and internal structure of each argument. By integrating argumentation mining in writing environments, human will be able to inspect their text for plausibility and to improve the quality of their argumentation.</p><p>A minimum definition of an argument is a set of statement that consists of 3 parts: conclusion, premises, and inference <ref type="bibr" target="#b21">(Walton, 2009)</ref>. On the other hand, it is stated that argument is a statement with 3 components: claim/point of view that is argued, actual argument/evidence, a statement that links first claim to the argument and makes sure the function of argument can be understood. <ref type="bibr" target="#b7">(Moens, 2014)</ref> Palau <ref type="bibr">(2008)</ref> stated that argumentation detection can help to facilitate understanding of argumentation paragraph, demonstrate a good identification for important information, increase the possibility of indexing implementation or document searching, represent reasoning system. The classification of argument component and visualization has several advantages, such as to show clear, strong, and structured/organized arguments. Besides, it also facilitates evaluation of opinion, facilitates understanding of other's opinions, helps in giving the teaching of general thoughts, and helps in teaching critical thinking. Thus, having a better accuracy in classifying argument components becomes a compulsory problem.</p><p>In this work, we proposed some new features on top of the state-ofthe-art research in argumentation mining. We implemented 68 subfeatures that are grouped into 8 main categories of features. They are structural, lexical, contextual, indicators, prompt similarity, syntactic, word embedding, and discourse. We also provided accuracy comparison to previous systems that were related to our work. We propose approach that consists of two main steps in our research. First, we did component identification, which include a process of identification and detection of argument component. We separated argumentative text units from non-argumentative text units and also identified the presence of argument component. Secondly, we did component classification, which include classification process of argument component type into major claim, claim, premise, or non-argumentative.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">RELATED WORKS</head><p>There are several works that are related with this research, specifically in the field of argument detection and classification. Moens, Boiy, Palau, and Reed (2007) did a research of automatic detection of arguments in legal texts. They used lexical, syntactic, semantic, and discourse features. In this research, they used Araucaria corpus as the dataset and Multinomial naïve Bayes and maximum entropy model as the classifiers. As the result, they obtained 74% accuracy of all features extraction in the variant of texts and 68% in legal texts. The detection and classification of argument component and the identification of argument structure was proposed by <ref type="bibr" target="#b10">Palau and Moens (2009)</ref>. They used Araucaria corpus and European Court of Human Rights (ECHR) as the data and feature extraction as the method. This research obtained 73% of accuracy in Araucaria and 80% of accuracy in ECHR. On the other hand, the accuracy was 74.07% for premise and conclusion classification and it yielded 60% for detecting the argument structure. <ref type="bibr" target="#b6">Lippi and Torroni (2015)</ref> proposed several methods to detect claims. They used IBM corpus dataset and 90 persuasive essays. As the result, they achieved 71.4% of accuracy in the 90 persuasive essays and 20.6% in IBM corpus. Al- <ref type="bibr" target="#b1">Khatib et al. (2016)</ref> proposed a distant supervision approach in classifying argumentative parts in text automatically from online debate portal. They used corpus of Webis-debate-16 and did a cross-domain comparison with 90 persuasive essays and web discourse corpus. This research achieved 66.8% of accuracy in 90 persuasive essays corpus, 87.7% of accuracy in web discourse corpus, and 91.8% of accuracy in Webis-debate-16. For the experiment of cross-domain comparison, the highest accuracy was obtained by web discourse corpus tested in Webis-debate-16, which reached 84.4% of accuracy.</p><p>The other focus to classify the arguments by identifying argumentation schemes was done by <ref type="bibr">Feng and Hirst (2011)</ref>. They used Araucaria database, features extraction, and two methods of classification. The features used in this research were general and scheme-specific features. The highest accuracy was 90.8% in scheme target of reasoning while the lowest accuracy was 63.2% in scheme target of classification for one-against-othersclassification. For pairwise classification, the highest accuracy was 98.3% in scheme target of classification-reasoning and the lowest accuracy was 64.2% in scheme target of classificationconsequences.</p><p>To identify the argumentative discourse, some researchers did annotation study to create the corpus. <ref type="bibr" target="#b14">Stab and Gurevych (2014a)</ref> did the annotation study and created corpus of 90 persuasive essays. They continued the research by identifying the argument component and the argumentative relations in persuasive essays. Support Vector Machine (SVM) was used and it obtained 77.3% of accuracy with structural feature as the best performing feature. On further research, they created an approach to parse the argumentation structures in persuasive essays <ref type="bibr">(Stab and Gurevych, 2016)</ref>. They created corpus of 402 persuasive essays and extracted the features to identify the argument component, classified the argument component, identified the argumentative relation, tree generation, and stance recognition. They obtained 77.3% of accuracy and structural was the best performing features. They also proposed approach to recognize the absence of opposing arguments in persuasive essays. They used both corpus of 90 persuasive essays and 402 persuasive essays. As the result, they got 75.6% of accuracy. The combination of unigrams, production rules, and adversative transitions obtained the highest accuracy among all of combinations. Habernal and Gurevych (2016) annotated and analyzed the arguments automatically in user-generated web discourse by extracting 5 (five) feature sets to detect the argument component. As the result, they obtained 75.4% of accuracy. Some researchers focused on the approach to identify the argumentation structures. <ref type="bibr" target="#b11">Peldszus (2014)</ref> proposed an approach to identify argumentation structures in micro text automatically with the various level of granularity. They used 115 micro text as the dataset and extracted the features and did a comparison with some types of classifiers. The most outperformed classifiers were Support Vector Machine (SVM) and Maximum Entropy Classifiers (MaxEnt). SVM obtained 64% of accuracy and MaxEnt obtained 63% of accuracy. The best features to obtain the high accuracy were lemma unigrams and lemma bigrams. Lawrence and Reed (2015) proposed 3 (three) methods to extract argumentation structures. They used AIFdb corpus and implemented discourse indicators, topic similarity, and schematic structure as the methods. The combination of those methods reached 83% of accuracy with the best performing feature was schematic feature.</p><p>Further implementation of argumentation detection and classification, such as accessing the quality of arguments have been done by some researchers. Wachsmuth, Al-Khatib, and Stein (2016) investigated mining structure to access the argumentation quality of persuasive essays. They used corpus that contains essays from International Corpus of Learner English, extracted the features, and classified the argument component into ADU types: thesis, conclusion, premise, and none. They obtained 74.5% of accuracy with the sentence position as the best performing feature.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">METHODS 3.1 Data</head><p>We utilized a corpus of persuasive essays compiled by <ref type="bibr">Stab and Gurevych (2016)</ref>. It consists of 402 annotated persuasive essays with different kind of topics. This corpus contains argument component annotation in the clause-level as well as argumentative relations and argument structure in a different level of discourse. It also contains annotation about major claim, claim, premise in each of essay and consists of 7.116 sentences with 147.271 tokens.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Current Features</head><p>We implemented 68 sub-features that were categorized into 8 groups: structural, lexical, indicators, contextual, syntactic, prompt similarity, word embedding, and discourse features. The features described in this section were combined from some researches in argument components classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">Structural Features</head><p>Structural features are features that identified argument component based on structure of the text. Covering sentence is a sentence that contains the argument component in it. Structural includes 3 subfeatures, which are token statistics, location, and punctuation. For token statistics, we defined the number of tokens from argument component, the number of tokens from covering sentence, the number of tokens preceding and following an argument component in the covering sentence, the token ratio between covering sentence and argument component, the number of tokens from covering paragraph, the number of covering sentences preceding and following paragraph, the token ratio between covering sentence and covering paragraph, the token ratio between covering sentence and essay, the average number of token at sentence, the ratio and a Boolean feature that indicates if the argument component covers all tokens of its covering sentence as token statistics features. For location, we defined a set of location-based features for exploiting the structural properties of essay. 4 Boolean features that indicate if the argument component is present in the introduction or conclusion of an essay and if it is present in the first or the last sentence of a paragraph. Secondly, we add the position of the covering sentence in the essay and the position of the covering sentence in the paragraph as a numeric feature. We also count the ratio of covering sentence and paragraph, the ratio of covering sentence and essay, and the ratio of paragraph and essay. For punctuation, we define a set of punctuation-based feature to identify characteristics of argument component. This features will return the number of punctuation marks of the covering sentence and the number of punctuation marks of the argument component, the number of punctuation marks preceding and following an argument component in its covering sentence and a Boolean feature that indicates whether the sentence is closed with a question mark or not.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2">Lexical Features</head><p>These features are defined by N-grams, POS N-grams, verbs, adverbs, modals auxiliary, comparative and superlative adjective, the ratio of pronouns, and word couples.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3">Indicator Features</head><p>Boolean features indicating the presence of question indicators, time indicators, evidence indicator, conclusion indicator, compareand-contrast, and cue phrases. We used 55 discourse markers as well and modelled each as a Boolean feature set to true if one of them is present in the covering sentence. The discourse markers were taken from the Penn Discourse Treebank 2.0 Annotation Manual <ref type="bibr" target="#b13">(Prasad et. al., 2007)</ref>. Furthermore, we also define 4 (four) Boolean features that indicate the presence of type indicators including forward indicators, backward indicators, thesis indicators and rebuttal indicators. In addition, we defined 5 (five) Boolean features to identify possessive pronoun (I, me, mine, myself, my) in covering sentence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.4">Contextual Features</head><p>These features return the number of punctuations, number of tokens and sub-clauses from the sentence preceding and following the covering sentence, the number of covering sentence preceding and following the covering sentence. We also defined Boolean features indicating the presence of modal verbs, question indicator, comparative and superlative adjective, and type of indicators. In addition, we defined 4 (four) Boolean features and numeric that indicate if the shared noun and shared verb is present in the introduction or conclusion of an essay.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.5">Syntactic Features</head><p>We count the number of sub-clauses in each sentence and return numeric value. We also count the depth of parse tree, extract the production rules, and identify whether the sentence is in past tense, present tense, or not in both.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.6">Prompt Similarity Features</head><p>These features were created to count the similarity of cosine value between current sentence and the prompt, with the first sentence in each paragraph, with the last sentence in each paragraph, with its preceding sentence, and with its following sentence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.7">Word Embedding Features</head><p>They were created to count the vector representation of each word. Glove was used to obtain the vector representation for each word. We count the average of vector values per argument component.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.8">Discourse Features</head><p>We implemented discourse doubles, which return: (1) count of explicit and implicit relation in a sentence and then return the count of which type present the most, (2) the ratio of explicit and implicit relation. Explicit discourse connectives are drawn primarily from well-defined syntactic classes, while implicit discourse connectives are inserted between paragraph-internal adjacent sentence pairs not related explicitly by any of the syntactically defined set of explicit connectives.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Additional Features</head><p>To explore further in classifying argument components, we defined some features which are quite promising to boost the accuracy of classification. Our additional features included 7 main features, which were structural, lexical, indicators, contextual, syntactic, prompt similarity, and discourse features.</p><p>Structural features were number of token in covering paragraph, number of preceding and following covering sentence in covering paragraph, and position of covering sentence in paragraph. Lexical features were POS N-grams and word couples. Indicator features were forward, backward, rebuttal, thesis indicators, and cue phrases. Contextual features were type of indicators in context, number of shared noun and shared verb that are present in introduction and conclusion in essay, and 4 binary features that indicates shared noun and verbs that are present in introduction or conclusion in essay. Syntactic feature was POS distribution.</p><p>Prompt similarity feature was the similarity of cosine value between current sentence with the prompt. Word embedding feature was defined to extract the vector representation of each word.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">RESULTS AND DISCUSSION 4.1 Performance</head><p>There are 8 categories of features that were implemented for the features extraction: structural, indicator, contextual, lexical, syntactic, prompt similarity, word embedding, and discourse with total of 68 sub-features. We used Support Vector Machine (SVM) as classifier by using 10-folds cross validation and utilized a corpus of 402 annotated persuasive essays by <ref type="bibr">Stab and Gurevych (2016)</ref>.</p><p>The accuracy result of this system was 79.96%. It indicated that a higher accuracy was achieved in comparison to the argument component detection and classification systems conducted in the previous works as shown in Table <ref type="table" target="#tab_0">1</ref>. Even though this comparison did not show a proper objective evaluation due to task differences among them, our accuracy was quite promising to surpass previous works, especially to Stab and Gurevych (2014b). We guessed the accuracy to identify claims was very low due to class imbalance where claim had the lowest amount of data. Beside using 10-folds cross validation for training, we also conducted experiments using 5-folds cross validation with 79.74% accuracy. We conducted experiments by using each feature group to capture which feature sets were significant in classifying the argument components. Based on Table <ref type="table" target="#tab_2">3</ref>, the best feature set to classify argument components is structural feature with 77.83% accuracy result. Contextual and lexical features consecutively were the next significant features among all.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Combining the Features</head><p>We attempted to combine all features as the next experiment. It was to identify which features combination has the best and the least impact in improving the system's accuracy. Based on Table <ref type="table" target="#tab_3">4</ref>, we can conclude that the most influential feature is structural, because all combination of features without structural has the lowest accuracy result with 69.74%, while the least influential feature is discourse as without discourse feature, the accuracy result is 79.98%.</p><p>From 8 trials of features combination, 7 of them showed significant accuracy, where 7 of them achieved an accuracy of 77.7% to 79.9%. This result indicates that the accuracy achieved by the combination of features produces higher accuracy compared to the accuracy of previous works (Table <ref type="table" target="#tab_0">1</ref>). In addition, we can see from the experiments that the accuracy of the system significantly decreased as a result of the feature extraction without structural features. Therefore, we also did an experiment with combination of 3 (three) features that achieved the highest accuracy, i.e. structural, lexical, and contextual features which produced 77.87% as the accuracy result.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Comparing Each Group of Features</head><p>We conducted other experiments by comparing system's accuracy among implementation by using the features presented by <ref type="bibr" target="#b15">Stab and Gurevych (2014b)</ref>, handcrafted features proposed by authors, and additional features from previous works. The system was trained using the same corpus consisting 402 annotated persuasive essays compiled by <ref type="bibr">Stab and Gurevych (2016)</ref>.</p><p>Stab and Gurevych (2014b) implemented structural, indicator, contextual, lexical, and syntactic features with total of 28 subfeatures. Our system's accuracy result using features extraction based on Stab and Gurevych (2014b) is 76.32% (Table <ref type="table" target="#tab_4">5</ref>), while the original accuracy result of their research was 77.3% by using 90 persuasive essays where the highest accuracy is achieved by structural features. The result's difference can be caused by the different number of the training data. We proposed some handcrafted features to develop algorithm to identify and classify argument components and to increase the accuracy of system. This experiment used 24 sub-features which produced 68.46% as the accuracy result. In addition, we ran the system using each feature's category to identify each feature's performance (Table <ref type="table" target="#tab_5">6</ref>). From the result presented in Table <ref type="table" target="#tab_5">6</ref>, the system achieved 68.46% accuracy with the highest accuracy achieved by structural features which followed by contextual and prompt similarity features as the second and the third most performing features.</p><p>The experiments also implemented additional features which were obtained from previous works conducted from state-of-the-art researches. There are 16 additional sub-features implemented in this scenario. Based on Table <ref type="table" target="#tab_6">7</ref>, the system achieved 71.08% of accuracy with the most significant accuracy was achieved by structural and contextual features. Word embedding feature was less performing feature in this experiment. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">CONCLUSIONS</head><p>After all the experiments, we have done to detect and classify the argument component, we found that 79.96% of accuracy was achieved by implementing all features set. We defined 68 subfeatures which were summarized into 8 categories of features: they were structural, lexical, indicator, contextual, syntactic, word embed-ding, prompt similarity, and discourse features. We found that structural features were the best feature that had the most significant impact to the system's accuracy, which obtained 77.83% accuracy. The other significant features are contextual and lexical, with the accuracy of 63.10% and 61.06%.</p><p>The most significant features combination was the combination of all features without discourse features. This combination obtained 79.98% accuracy, which was higher than the total accuracy of all features. The combination of all features without structural performed the lowest accuracy, so that we conclude that structural features was the most significant feature while discourse features was not. Besides, the combination of 3 (three) structural, contextual, and lexical features also performed a significant accuracy, which was 77.87%. Features proposed by Stab and Gurevych (2014b) performed the highest accuracy, which was 76.32%. Each of experiment in comparing features classification could obtain more than 67% of accuracy. It means that each of the experiment could identify argument components for more than 67%.</p><p>Since the experiments showed that the most significant features were structural, contextual, and lexical, we concern to develop these groups for our next experiment. We also find that the data training in bigger number with various topics and characteristics will probably increase the accuracy of system. Besides, we also must define the other features or the other method that can help in differentiate the premise and claim further.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 . Previous works performance</head><label>1</label><figDesc></figDesc><table><row><cell>Related Work</cell><cell>Accuracy</cell></row><row><cell>Palau and Moens (2007)</cell><cell>74%</cell></row><row><cell>Palau and Moens (2009)</cell><cell>74.04%</cell></row><row><cell>Stab and Gurevych (2014b)</cell><cell>77.3%</cell></row><row><cell>Lippi and Torroni (2015)</cell><cell>71.4%</cell></row><row><cell>Stab and Gurevych (2016)</cell><cell>77.3%</cell></row><row><cell>Wachsmuth, Al-Khatib, and Stein (2016)</cell><cell>74.5%</cell></row><row><cell>Habernal and Gurevych (2016)</cell><cell>75.4%</cell></row><row><cell>Al-Khatib et al. (2016)</cell><cell>66.8%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 . Confusion matrix of the system accuracy results (SVM) for argument component classification</head><label>2</label><figDesc></figDesc><table><row><cell></cell><cell>MC</cell><cell>Cl</cell><cell>Pr</cell><cell>No</cell></row><row><cell>MC</cell><cell>578</cell><cell>130</cell><cell>43</cell><cell>0</cell></row><row><cell>Cl</cell><cell>226</cell><cell>309</cell><cell>970</cell><cell>1</cell></row><row><cell>Pr</cell><cell>28</cell><cell>147</cell><cell>3656</cell><cell>1</cell></row><row><cell>No</cell><cell>0</cell><cell>0</cell><cell>0</cell><cell>1638</cell></row><row><cell cols="5">Table 2 explains that the system correctly identifies 578 major</cell></row><row><cell cols="5">claims (MC), 309 claims (Cl), 3656 premises (Pr), and 1638 non-</cell></row><row><cell cols="5">argumentative (No). The errors occurred in identifying claims.</cell></row><row><cell cols="5">Most of them were identified as premise. The accuracy in</cell></row><row><cell cols="5">identifying each component was 76.96% for major claim, 20.52%</cell></row><row><cell cols="5">for claim, 95.41% for premise, and 100% for non-argumentative.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 . Previous works performance</head><label>3</label><figDesc></figDesc><table><row><cell>Feature</cell><cell>Accuracy</cell><cell>Feature Name</cell><cell>Accuracy</cell></row><row><cell>Name</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Structural</cell><cell>77.83%</cell><cell>Syntactic</cell><cell>51.35%</cell></row><row><cell>Indicator</cell><cell>54.73%</cell><cell>Prompt Similarity</cell><cell>54.79%</cell></row><row><cell>Contextual</cell><cell>63.10%</cell><cell>Word Embedding</cell><cell>49.46%</cell></row><row><cell>Lexical</cell><cell>61.06%</cell><cell>Discourse</cell><cell>49.41%</cell></row><row><cell>All Features</cell><cell></cell><cell>79.96%</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 . Accuracy result of combination of feature without one feature category in system</head><label>4</label><figDesc></figDesc><table><row><cell>Feature Name</cell><cell>Accura</cell><cell cols="2">Feature Name</cell><cell>Accuracy</cell></row><row><cell></cell><cell>cy</cell><cell></cell><cell></cell></row><row><cell cols="4">Without Structural 69.74% Without Syntactic</cell><cell>78.21%</cell></row><row><cell>Without Lexical</cell><cell cols="3">77.72% Without Prompt</cell><cell>79.93%</cell></row><row><cell></cell><cell></cell><cell>Similarity</cell><cell></cell></row><row><cell>Without</cell><cell cols="2">77.98% Without</cell><cell>Word</cell><cell>78.48%</cell></row><row><cell>Indicators</cell><cell></cell><cell>Embedding</cell><cell></cell></row><row><cell>Without</cell><cell cols="2">78.05% Without</cell><cell></cell><cell>79.98%</cell></row><row><cell>Contextual</cell><cell></cell><cell>Discourse</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 . Accuracy result of implementation features by Stab and Gurevych (2014b)</head><label>5</label><figDesc></figDesc><table><row><cell>Feature Name</cell><cell>Accuracy</cell></row><row><cell>Structural</cell><cell>74.33%</cell></row><row><cell>Indicator</cell><cell>61.11%</cell></row><row><cell>Contextual</cell><cell>52.38%</cell></row><row><cell>Lexical</cell><cell>58.69%</cell></row><row><cell>Syntactic</cell><cell>50.94%</cell></row><row><cell>All Features</cell><cell>76.32%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 . Accuracy result of proposed handcrafted features</head><label>6</label><figDesc></figDesc><table><row><cell>Feature Name</cell><cell>Accuracy</cell></row><row><cell>Structural</cell><cell>63.81%</cell></row><row><cell>Indicator</cell><cell>49.45%</cell></row><row><cell>Contextual</cell><cell>59.94%</cell></row><row><cell>Lexical</cell><cell>49.58%</cell></row><row><cell>Prompt Similarity</cell><cell>54.70%</cell></row><row><cell>Discourse</cell><cell>49.43%</cell></row><row><cell>All Features</cell><cell>68.46%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7 . Accuracy result of additional features from state-of- the-art researches</head><label>7</label><figDesc></figDesc><table><row><cell>Feature Name</cell><cell>Accuracy</cell></row><row><cell>Structural</cell><cell>61.15%</cell></row><row><cell>Indicator</cell><cell>53.95%</cell></row><row><cell>Contextual</cell><cell>50.69%</cell></row><row><cell>Lexical</cell><cell>59.27%</cell></row><row><cell>Syntactic</cell><cell>50.72%</cell></row><row><cell>Prompt Similarity</cell><cell>54.79%</cell></row><row><cell>Word Embedding</cell><cell>49.46%</cell></row><row><cell>All Features</cell><cell>71.08%</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="18" xml:id="foot_0">18th Workshop on Computational Models of Natural Argument Floris Bex, Floriana Grasso, Nancy Green (eds) 16th July 2017, London, UK</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">ACKNOWLEDGMENTS</head><p>This research work was supported by Bina Nusantara University and partly supported by research grant from Directorate General of Research and Development Reinforcement, Ministry of Research, Technology and Higher Education of the Republic of Indonesia.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Cross-domain mining of argumentative text through distant supervision</title>
		<author>
			<persName><forename type="first">A</forename><surname>Khatib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><surname>Matthias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kohler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">15th Conf. Of the North American Chapter of the Association for Computational Linguistics (NAACL&apos;16) (</title>
				<meeting><address><addrLine>San Diego, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Classifying arguments by scheme</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">W</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Graeme</forename><forename type="middle">H</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 49th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>49th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Portland, Oregon</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2011">2011. 2011</date>
			<biblScope unit="page" from="987" to="996" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Exploiting debate portals for semi-supervised argumentation mining in user-generated web discourse</title>
		<author>
			<persName><forename type="first">I</forename><surname>Habernal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP 2015)</title>
				<meeting>the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP 2015)<address><addrLine>Lisbon, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="page" from="2127" to="2137" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Argumentation mining in user-generated web discourse</title>
		<author>
			<persName><forename type="middle">I</forename><surname>Habernal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Linguistics</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note>in press</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Combining argument mining techniques</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lawrence</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Reed</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd Workshop on Argumentation Mining</title>
				<meeting>the 2nd Workshop on Argumentation Mining<address><addrLine>Denver, Colorado</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="page" from="127" to="136" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Context-independent claim detection for argumentation mining</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lippi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Torroni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI</title>
				<meeting>the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI</meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Tutorial Argumentation Mining</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<pubPlace>Belgium</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Automatic detection of arguments in legal texts</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Boiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Palau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Reed</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th International Conference on Artificial Intelligence and Law, ICAIL &apos;07</title>
				<meeting>the 11th International Conference on Artificial Intelligence and Law, ICAIL &apos;07<address><addrLine>Stanford, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="225" to="230" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Automatic argumentation detection. Project ACILA -Automatic Detection and Classification of Arguments in a Legal Case</title>
		<author>
			<persName><surname>Palau</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<pubPlace>Leuven, Belgium</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Argumentation mining: the detection, classification and structure of arguments in text</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Palau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Moens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th International Conference on Artificial Intelligence and Law, ICAIL&apos;09</title>
				<meeting>the 12th International Conference on Artificial Intelligence and Law, ICAIL&apos;09<address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="98" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Towards segment-based recognition of argumentation structure in short texts</title>
		<author>
			<persName><forename type="first">A</forename><surname>Peldszus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the First Workshop on Argumentation Mining</title>
				<meeting>the First Workshop on Argumentation Mining<address><addrLine>Baltimore, Maryland USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014-06-26">2014. June 26, 2014</date>
			<biblScope unit="page" from="88" to="97" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Form argument diagrams to argumentation mining in texts: a survey</title>
		<author>
			<persName><forename type="first">A</forename><surname>Peldszus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stede</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Cognitive Informatics and Natural Intelligence</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="31" />
			<date type="published" when="2013-01">2013. January 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">The Penn Discourse Treebank 2.0 annotation manual</title>
		<author>
			<persName><forename type="first">R</forename><surname>Prasad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Miltsakaki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Dinesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Robaldo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">L</forename><surname>Webber</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>Institute for Research in Cognitive Science, University of Pennsylvania</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Annotating argument components and relations in persuasive essays</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th International Conference on Computational Linguistics (COLING 2014)</title>
				<meeting>the 25th International Conference on Computational Linguistics (COLING 2014)<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014a. 2014</date>
			<biblScope unit="page" from="1501" to="1510" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Identifying argumentative discourse structures in persuasive essays</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Empirical Methods in Natural Language Processing (EMNLP 2014)</title>
				<meeting><address><addrLine>Doha, Qatar</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014b. 2014</date>
			<biblScope unit="page" from="46" to="56" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Parsing argumentation structures in persuasive essays</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016-04">2016. April 2016</date>
			<pubPlace>Germany</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Technische Universität Darmstadt</orgName>
		</respStmt>
	</monogr>
	<note>arXiv preprint, under review</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Recognizing the absence of opposing arguments in persuasive essays</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd Workshop on Argument Mining held in conjunction with the 2016 Annual Meeting of the Association for Computational Linguistics (ACL 2016)</title>
				<meeting>the 3rd Workshop on Argument Mining held in conjunction with the 2016 Annual Meeting of the Association for Computational Linguistics (ACL 2016)</meeting>
		<imprint>
			<date type="published" when="2016-08">2016. August 2016</date>
			<biblScope unit="page" from="113" to="118" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Detecting argument components and structures</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Habernal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Report of Dagstuhl Seminar on Debating Technologies</title>
				<imprint>
			<date type="published" when="2015">2015. 2016</date>
			<biblScope unit="volume">15512</biblScope>
			<biblScope unit="page" from="32" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">The Uses of Argument</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Toulmin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1958">1958</date>
			<publisher>Cambridge University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Using Argument Mining to Assess the Argumentation Quality of Essays</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Khalid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<pubPlace>Germany</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Bauhaus-Universität Weimar</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Argumentation Theory: A Very Short Introduction</title>
		<author>
			<persName><forename type="first">D</forename><surname>Walton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Argumentation in Artificial Intelligence</title>
				<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="1" to="24" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
