<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automatic Recognition of Narrative Drama units: a structured learning approach</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Danilo</forename><surname>Croce</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Enterprise Engineering</orgName>
								<orgName type="institution" key="instit1">University of Roma</orgName>
								<orgName type="institution" key="instit2">Tor Vergata (Italy)</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Roberto</forename><surname>Basili</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Dept. of Enterprise Engineering</orgName>
								<orgName type="institution" key="instit1">University of Roma</orgName>
								<orgName type="institution" key="instit2">Tor Vergata (Italy)</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vincenzo</forename><surname>Lombardo</surname></persName>
							<email>vincenzo.lombardo@unito.it</email>
							<affiliation key="aff2">
								<orgName type="department">Dept. of Informatics</orgName>
								<orgName type="institution">University of Torino</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Eleonora</forename><surname>Ceccaldi</surname></persName>
							<email>eleonoraceccaldi@gmail.com</email>
							<affiliation key="aff3">
								<orgName type="institution">DIBRIS University of Genova</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automatic Recognition of Narrative Drama units: a structured learning approach</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">3F70756EC965E0A5B7A5E7B8A96FE495</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T00:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Drama is a story told through the live actions of characters; dramatic writing is characterized by aspects that are central to identify, interpret, and relate the different elements of a story. The Drammar ontology has been proposed to represent core dramatic qualities of a dramatic text, namely Actions, Agents, Scenes and Conflicts, evoked by individual text units. The automatic identification of such elements in a drama is the first step in the recognition of their evolution, both at coarse and fine grain text level. In this paper, we address the issue of segmentation, that is, the partition of the drama into meaningful unit sequences We study the role of editorial as well as content-based text properties, without relying on deep ontological relations. We propose a generative inductive machine learning framework, combining Hidden Markov models and SVM and discuss the role of event information (thus involving agents and actions) at the lexical and grammatical level.</p><p>LOPAKHIN. The train's arrived, thank God. What's the time?</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Drama is a story told through the live actions of characters. The Drammar ontology <ref type="bibr" target="#b8">[LDP18,</ref><ref type="bibr" target="#b4">DLPar]</ref> identifies the core dramatic qualities of a dramatic text, namely Actions, Agents, Units/Scenes, and Conflicts, implicitly evoked by the dramatic text, as claimed by the scientific literature on drama analysis.</p><p>Drama relies on an internal coherence and a rich set of eventualities, related to the interactions among characters and the insurgence and resolution of conflicts. Dramas are very well structured. As a running example, we address the incipit of Anton Chekhov's "The Cherry Orchard ", in its English translation <ref type="bibr" target="#b2">[Che17]</ref>:</p><p>A room which is still called the nursery. One of the doors leads into ANYA'S room. It is close on sunrise. ... DUNYASHA comes in with a candle, and LOPAKHIN with a book in his hand.</p><p>DUNYASHA. It will soon be two. [Blows out candle] It is light already.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Schematically:</head><p>• individual utterances are denoted by the correspondent acting characters;</p><p>• some editorial notes (in italics) interleave with the spoken parts where the authors suggest the environment changes or specific happenings;</p><p>• a strict separation between spoken and editorial fragments is imposed.</p><p>Our current research objective aims at supporting the automatic annotation of a drama, able to outline the evolution of the dramatic elements above through discrete events.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Events in narrative texts</head><p>Following the observer point of view, <ref type="bibr" target="#b16">[SZR07]</ref> propose the following definitions for the events:</p><p>• An Event is "a segment of time at a given location, that is conceived by an observer to have a beginning and an end; granularity of events can go from a second or less to tens of minutes".</p><p>• An Event model is "an actively maintained representation of the current event, which is updated at perceptual event boundaries".</p><p>• The Event segmentation is "the perceptual and cognitive process by which a continuous activity is segmented into meaningful events". From the psychological literature, we have that readers structure a narrative text into a series of events in order to understand and remember the text (cf. the experiments of [ZT01, ZS07, ZSSM10]). Events are coded at clause level. Relevant information for the narrative coding includes, e.g. <ref type="bibr" target="#b16">[SZR07]</ref>:</p><p>• Time and Space information (as the presence of spatial changes, e.g., moving from one room to another inside a house can be meaningful);</p><p>• Objects, given the interaction of characters with elements of a scene;</p><p>• change of Character, revealed by the changes of the subject of a clause;</p><p>• Causes (causal relationship over activities) and Goals (new goal-directed activities), to be coded as core dimensions of Events.</p><p>Usually, clauses are also coded for terminal punctuation (e.g., periods and question marks) and non terminal punctuation (e.g., commas and semicolons). As the annotation of such dramatic aspects is time consuming, we aim at automatizing it, relying upon the lexical, grammatical, and editorial information, expressed by individual clauses. In this way, events can be recognized and properly segmented along the dramatic text. We will refer hereafter this process as the event segmentation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">Event Segmentation: Related Work</head><p>Event segmentation is a task traditionally tackled in NLP according to sentence boundary detection methods (e.g. <ref type="bibr" target="#b6">[Hea94]</ref>, <ref type="bibr" target="#b14">[SDDK11]</ref>) or cohesion-based clustering models (e.g. <ref type="bibr" target="#b3">[Cho00]</ref>). Text segmentation methods usually search for the set of segments in a text that optimize some form of coherence of the content. Word usage is modeled in TextTiling <ref type="bibr" target="#b6">[Hea94]</ref> for each sentence in a sequence and the two sides of a potential boundary are selected when large lexical difference is found. Prosodic features and lexical features are taken into account to model discourse as in the Hidden Markov Model segmentation proposed in [YXX + 16]. The lexical connectivity strength between two adjacent fragments of a text is used as hint in DivSeg ( <ref type="bibr" target="#b14">[SDDK11]</ref>). Unsupervised approaches are based on probabilistic models (e.g. <ref type="bibr" target="#b6">[Hea94]</ref>, C99 <ref type="bibr" target="#b3">[Cho00]</ref> or the DotPlotting algorithm <ref type="bibr" target="#b12">[Rey94]</ref>) or agglomerative clustering <ref type="bibr" target="#b18">[Yaa99]</ref>). In the former group, terms frequencies are used to identify topical segments (dense dot clouds on the graphic). e.g. DotPlotting <ref type="bibr" target="#b12">[Rey94]</ref>. In the latter group, dendrograms are induced over paragraphs and transformed into a hierarchical segmentation <ref type="bibr" target="#b18">[Yaa99]</ref>. Lexical chains methods are applied in an unsupervised manner as they exploit semantic lexicons to model word associations and semantic relations. In these methods, a chain links multiple occurrences of a term in the document: it is considered broken when there are too many sentences between two occurrences of a term. The Segmenter system ([KKM98]) detects such broken points across a document according to possibly multiple chains. Some of the methods use lexical resources or forms of ontological similarity to model similarity metrics between text blocks (sentences or paragraphs), based on semantic information (e.g. recognized named entity in the text). Wordnet or Wikipedia-based methods have been proposed to define semantic similarity metrics between text units. Recently, deep learning methods have been applied to Text Segmentation, specifically to the Topic-based segmentation task. In particular <ref type="bibr" target="#b10">[LSJ18]</ref> presents an end-to-end segmentation model: first, a bidirectional recurrent neural network is used to encode input text sequences, and then, another recurrent neural network is used together with a pointer network to select text boundaries in the input sequence. Although very appealing, since it does not require hand-crafted features definition, this method requires a significant amount of training material, made of several hundreds of annotated documents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">A structured learning approach to drama segmentation</head><p>In line with most of the above event segmentation approaches, we will rely on a machine learning perspective by assuming a set of textual resources as the triggering observations:</p><p>• L is a set of fully annotated drama fragments, whose segments are completely known (e.g., the example units 4 to 6 in Checkov's The Cherry Orchard reported in the paper Appendix or the nunnery scene from Shakespeare's Hamlet <ref type="bibr" target="#b9">[LPD16]</ref>, respectively);</p><p>• O L is a very small scale corpus, made of fragments from the possibly partially annotated opera (e.g., the complete annotated dramas "The Cherry Orchard" or "Hamlet", respectively, though they are usually neither segmented, nor annotated);</p><p>• O A (L) is a large scale corpus of unannotated texts of the same author (e.g., all of Checkov's plays or Shakespeare's plays, respectively);</p><formula xml:id="formula_0">• O E (A(L)</formula><p>) is a comprehensive corpus of the drama works of the same epoch (e.g., Contemporary play or Elizabethan theatre plays, respectively).</p><p>So, we rely on the chain</p><formula xml:id="formula_1">L ⊂ O L ⊂ O A (L) ⊂ O E (A(L)).</formula><p>We propose an integration of unsupervised and supervised learning processes acting: our first attempt is to use the comprehensive O E (A(L)) to generate a lexical resource focused on the work and author style: according to unsupervised methods, such as <ref type="bibr" target="#b11">[MCCD13]</ref>, we can rely on word embedding for a large scale dictionary of lexical items: these generalize lexical semantics within the underlying targeted text genre. The proposal is to inject this information into the supervised steps that address the labeled material L, in order to fully label the entire work O L in an accurate manner. Annotated examples in L are the basic source of information for the segmentation stage.</p><p>Hereafter, we concentrate on the variety of lexical, grammatical and aspectual features (e.g. the mode and transitivity of a number of verbs involved in the dramatic action), suitably exploited for training a sequence labeling component over O L . We propose a structured learning paradigm based on independent kernels for training SVMs over L ( <ref type="bibr" target="#b15">[STC04]</ref>) and apply them within a Markovian modeling, isomorphic to HMM. The major steps are thus:</p><formula xml:id="formula_2">• (PreTraining) Use O E (A(L))</formula><p>to acquire lexical information in the form of a neural language model (in line with <ref type="bibr" target="#b11">[MCCD13]</ref>), expressing general semantic properties of individual words. A specific treatment of some classes of words is here applied. For example, character names (e.g. Dunyasha and Lopakhin in The Cherry Orchard or Hamlet and Ophelia in Hamlet). This is a standard a-priori information for a drama that is mapped into the category label Character, in order to minimize sparsity.</p><p>• (Feature Modeling and Extraction) Feature extraction is applied to derive textual, editorial and narrative features, as discussed in Section 2.2</p><p>• (Model Optimization) Then, a structured Machine Learning model is applied to achieve segmentation as an IOB-like sentence labeling process, in order to organize sentences in units and hierarchies of scenes. The adopted algorithm is known as SVM-HMM ([TJHA05], adopted in [CB11, BCV + 16]).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">A Markovian Support Vector Machine</head><p>The aim of a Markovian formulation of SVM is to make the classification of a input example x i ∈ R n (belonging to a sequence of examples) dependent on the labels assigned to the previous elements in a history of length m, i.e., x i−m , . . . , x i−1 . In our classification task, a drama is a sequence of utterances x = (x 1 , . . . , x s ), each of them representing the example x i , i.e., the specific i-th paragraph. Given the corresponding sequence of expected labels y = (y 1 , . . . , y s ), a sequence of m step-specific labels (from a dictionary of d symbols) can be retrieved, in the form y i−m , . . . , y i−1 . In our machine learning setting, labels are related to the Segmentation task: we will thus adopt the IOB notation so that each element in the drama will be associated to the label B if it is at the Beginning of a Unit,</p><formula xml:id="formula_3">I if it is Inside it, O if it is Out of the Unit itself.</formula><p>In order to make the classification of x i also dependent on the previous decisions, we augment the feature vector of x i by introducing a projection function ψ m (x i ) ∈ R md that associates each example with a md−dimensional feature vector where each dimension set to 1 corresponds to the presence of one of the d possible labels observed in a history of length m, i.e. m steps before the target element x i .</p><p>In order to apply a SVM, a projection function φ m (•) can be defined to consider both the observations x i and the transitions ψ m (x i ) by concatenating the two representations as follows:</p><formula xml:id="formula_4">φ m (x i ) = x i || ψ m (x i ) with φ m (x i ) ∈ R n+md .</formula><p>Notice that the symbol || here denotes the vector concatenation, so that ψ m (x i ) does not interfere with the original feature space, where x i lies. Kernel-based methods can be applied in order to model meaningful representation spaces, encoding both the feature representing individual examples together with the information about the transitions. According to kernel-based learning <ref type="bibr" target="#b15">[STC04]</ref>, we can define a kernel function K m (x i , z j ) between a generic item of a sequence x i and another generic item z j from the same or a different sequence, parametric in the history length m. It surrogates the dot product between φ m (•) such that:</p><formula xml:id="formula_5">K m (x i , z j ) = φ m (x i )φ m (z j ) = K obs (x i , z j ) + K tr ψ m (x i ), ψ m (z j )</formula><p>We define a kernel that is the linear combination of two further kernels: K obs operating over the individual examples x i and a K tr operating over the feature vectors encoding the involved transitions. It is worth noticing that K obs neither depends on the position nor on the context of individual examples, in line with Markov assumption that characterizes a large class of these generative models, e.g. HMM. For simplicity, we define K tr as a linear kernel between input instances, i.e. a dot-product in the space generated by ψ m (•):</p><formula xml:id="formula_6">K m (x i , z j ) = K obs (x i , x j ) + ψ m (x i )ψ m (z j )</formula><p>At training time, we use the kernel-based SVM in a One-Vs-All schema over the feature space derived by K m (•, •). The learning process provides a family of classification functions f (x i ; m) ⊂ R n+md × R d , which associate each x i to a distribution of scores with respect to the different d labels, depending on the context size m. At classification time, all possible sequences y ∈ Y + should be considered in order to determine the best labeling ŷ, where m is the size of the history used to enrich x i , that is: ŷ = arg max y∈Y + { i=1...m f (x i ; m)}</p><p>In order to reduce the computational cost, a Viterbi-like decoding algorithm is adopted<ref type="foot" target="#foot_0">1</ref> as described in Fig. <ref type="figure" target="#fig_0">1</ref>. The next section defines the kernel function K obs applied to specific turns in the drama.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Modeling dramatic properties as ML features</head><p>Three types of kernels are applied for different types of features. Lexical features include sentence embeddings as linear combinations of individual word embeddings, grammatical patterns, such as verb-objects or subject-verb pairs, POS n-grams (n=3) and, finally, sentence properties such as length and complexity (e.g. number of different active mode verbs). Narrative features are strictly dependent on the narrative structure and express possible Characters and Actions in a turn. Named Entity Recognition is first run on the individual utterances, to capture character mentions. A narrative vector including the acting character (e.g. LOPAKHIN in line 0036 or 0038 in the Appendix) as well as all the other recently mentioned characters (e.g. LOPAKHIN and DUNYASHA in the editorial note at line 0042). Individual features modeling the number of mentioned or recently mentioned characters for each turn will be adopted. An aging mechanism defines lower scores for no longer mentioned characters. Finally, narrative features denoting the Actions mentioned in a turn will be adopted in order to account for the interaction (and possible conflicts) in an explicit way. Examples are motion verbs such as to come, to go, social verbs, such as to meet (see LOPAKHIN in unit 0040) or even emotional verbs (e.g. or to faint as in unit 0041. Specific dictionaries of English verbs and their nominalizations will be used here to denote narratively interesting Actions. Editorial features will depend on the material that includes the author's suggestions in the environment (see, for example, the sentence "A room which is still called the nursery" in the incipit). In this case, a representation that is similar to the one for the lexical features for individual acting turns is adopted, but the editorial material will be expressed through a separated vector, in order to play an independent role. Table <ref type="table">1</ref>: Performance scores and ablation analysis for the segmentation based on different lexical features. Tokenbased Accuracy figures are Strict when applied only to B-labeled paragraphs in the oracle, and Greedy when also all the consistently aligned I-labeled paragraphs are considered as correct. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Basic</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Experimental Evaluation and Discussion</head><p>In the current experimental stage, we applied ablation analysis to the set of Lexical Features as described in the previous section. The lexical model is tested via the HMM SVM framework (implemented within KeLP, [FCM + 18]) on the annotated version of The Cherry Orchard, in its English translation: the Appendix reports a short excerpt. The work includes 4 acts made of about 904 paragraphs, segmented into 67 units. The different tokens in our labeled corpus L = O L are thus about 22,800. Every paragraph has been considered part of a sequence of length k=5 that corresponds to the local input to the tagger. Every paragraph in the sequence is represented via feature vectors and several lexical representations have been adopted:</p><p>• (Simple Lexical ) A bag-of-word feature vector including lemmas, bi-grams and POS tags occurring in the target paragraph are represented.</p><p>• (Baseline) As a baseline, a set of simple heuristics from narrative features is used to simulate a blind typographic approach. The synthetic vector encodes only a label with the guessed editorial role of the paragraph. In this way, individual utterance and editorial notes are just kept separated.</p><p>• (Contextualized lexical ) Similar to the Lex feature vector, but extended with the vector of the preceding paragraph, in order to contextualize the model.</p><p>• (Word Embeddings) A real-valued vector that corresponds to the sentence embedding of the target paragraph is adopted. Each training paragraph is labeled according to the IOB notation, i.e. as B or I. A macro n-fold cross validation is applied with one fold per act, i.e n = 4. In one evaluation step, one act is removed from the dataset: training on the three remaining acts is carried out by leaving a 10% of paragraphs as development data (i.e. for tuning of the SVM parameters): the automatic tagging over the left-out act allows to measure and average the labeling accuracy. Each time, the training set is randomly split 90/10 to derive a development set used to tune the SVM parameters.</p><p>Measures of performance are class-based precision and recall, while accuracy is the percentage of paragraphs that are correctly re-labeled with respect to the original IOB label. Micro-average across the four different 4 folds is applied. Notice that for the unbalanced presence of the I tag (i.e. 92.6% of the paragraphs), the simple baseline model achieved a 93.5% of accuracy across all paragraphs. For this reason, in Table <ref type="table">1</ref>, we just report precision, recall and F1 for the two separated classes. Moreover, we report the strict accuracy, just computing the accuracy restricted to the B gold labeled paragraph. Notice that this class is defined by only 67 positive examples in the training dataset. Finally, the accuracy measured on only the aligned B and I paragraph is reported: it considers correct a paragraph labeled as "inner" by the system only when this does not violate any boundary B in the oracle annotation. As Table <ref type="table">1</ref> shows more complex lexical features brings more information as they increase performance for each measure. Moreover, (last column in last line) the token based accuracy suggests that the current model correctly annotate about 70% of the paragraphs of the work, thus representing a large advantage against manual annotation.</p><p>Examples of mistaken segmentations are hereafter reported where the gold and automatic labels are shown after the row number for the different paragraphs, respectively: ... According to the gold labels, the B-labeled paragraphs in lines 803 and 903 are all wrong, while 801,804 and 901,902,905 are correctly aligned I-labeled paragraphs: these latter are retained in the greedy version of the Token-based Accuracy scores. Notice how the mistakes are mainly due to mismatches in the way editorial material is used by the human annotators. In the first example, lines 802, 803, the beginning is annotated to the sitting act of YASHA (line 802). In the second, the first speech of LOPAKHIN is used to start a new segment (line 904). On the contrary, in both cases the system has focused on the entrance of the new character to suggest the start (i.e. B labels in line 803 and 903).</p><p>These mild errors suggest how the generalization of the system at this current stage of development is already acceptable in several cases. Accuracy rates are thus expected to grow when more complex features (for example the narrative features that will better express the ontological information) will be adopted. This will be part of future work.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The overall sequence labeling architecture for event segmentation.</figDesc><graphic coords="5,131.82,54.07,351.96,226.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>801 I I (Goes off ) 802 B I YASHA remains, sitting beside the shrine. 803 I B Enter RANYEVSKAYA, GAYEV, and LOPAKHIN. 804 I I LOPAKHIN. It has to be settled once and for all -time won't wait. Look, it's a simple enough question. Do you agree to lease out the land for summer cottages or not? Answer me one word: yes or no? Just one word! ... 901 I I YEPIKHODOV (Off, behind the door ). I'll tell about you! 902 I I VARYA. Oh, coming back, are you? (Seizes the stick that FIRS left besides the door.) Come on,then...Come on... Come on... I'll show you... Are you coming? My word, you're going to be for it...! (Raises the stick threateningly.) 903 I B Enter LOPAKHIN. 904 B I LOPAKHIN. Thank you kindly. 905 I I VARYA (angrily and sarcastically). Sorry! My mistake. ...</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">When applying f (x i ; m) the classification scores are normalized through a softmax function and probability scores are derived.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A discriminative approach to grounded spoken language understanding in interactive robotics</title>
		<author>
			<persName><surname>Bcv + ; Emanuele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Danilo</forename><surname>Bastianelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrea</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roberto</forename><surname>Vanzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniele</forename><surname>Basili</surname></persName>
		</author>
		<author>
			<persName><surname>Nardi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016</title>
				<meeting>the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-07-15">9-15 July 2016. 2016</date>
			<biblScope unit="page" from="2747" to="2753" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Structured learning for semantic role labeling</title>
		<author>
			<persName><forename type="first">Danilo</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roberto</forename><surname>Basili</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Artificial Intelligence Around Man and Beyond -XIIth International Conference of the Italian Association for Artificial Intelligence</title>
				<meeting><address><addrLine>Palermo, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2011-09-15">2011. September 15-17, 2011. 2011</date>
			<biblScope unit="page" from="238" to="249" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">Anton</forename><surname>Chekhov</surname></persName>
		</author>
		<title level="m">The Cherry Orchard. Plays, by Anton Tchekoff</title>
				<meeting><address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<publisher>Scribner&apos;s</publisher>
			<date type="published" when="1917">1917</date>
		</imprint>
	</monogr>
	<note>2d series, tr. with an introduction by Julius West</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Advances in domain independent linear text segmentation</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Y</forename><surname>Choi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st NAACL Conference</title>
				<meeting>the 1st NAACL Conference</meeting>
		<imprint>
			<publisher>ACL</publisher>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="26" to="33" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The ontology of drama</title>
		<author>
			<persName><forename type="first">Rossana</forename><surname>Damiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vincenzo</forename><surname>Lombardo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antonio</forename><surname>Pizzo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Ontology</title>
		<imprint/>
	</monogr>
	<note>to appear</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Kelp: a kernel-based learning platform</title>
		<author>
			<persName><forename type="first">Giuseppe</forename><surname>Fcm + ; Simone Filice</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Giovanni</forename><surname>Castellucci</surname></persName>
		</author>
		<author>
			<persName><surname>Da San</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aless</forename><surname>Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Danilo</forename><surname>Moschitti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roberto</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><surname>Basili</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">191</biblScope>
			<biblScope unit="page" from="1" to="5" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Multi-paragraph segmentation of expository text</title>
		<author>
			<persName><forename type="first">Marti</forename><forename type="middle">A</forename><surname>Hearst</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL</title>
				<imprint>
			<publisher>Morgan Kaufmann Publishers / ACL</publisher>
			<date type="published" when="1994">1994</date>
			<biblScope unit="page" from="9" to="16" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Linear segmentation and segment significance</title>
		<author>
			<persName><forename type="first">Min-Yen</forename><surname>Kan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Judith</forename><forename type="middle">L</forename><surname>Klavans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kathleen</forename><forename type="middle">R</forename><surname>Mckeown</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">VLC@COLING/ACL</title>
				<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Drammar: A comprehensive ontological resource on drama</title>
		<author>
			<persName><forename type="first">Vincenzo</forename><surname>Lombardo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rossana</forename><surname>Damiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antonio</forename><surname>Pizzo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ISWC 2018 -17th Int. Semantic Web Conf</title>
				<meeting><address><addrLine>Monterey, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">October 8-12, 2018. 2018</date>
			<biblScope unit="page" from="103" to="118" />
		</imprint>
	</monogr>
	<note>Proceedings, Part II</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Safeguarding and accessing drama as intangible cultural heritage</title>
		<author>
			<persName><forename type="first">Vincenzo</forename><surname>Lombardo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antonio</forename><surname>Pizzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rossana</forename><surname>Damiano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JOCCH</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">26</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Segbot: A generic neural text segmentation model with pointer network</title>
		<author>
			<persName><forename type="first">Jing</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aixin</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shafiq</forename><surname>Joty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18</title>
				<meeting>the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page">7</biblScope>
		</imprint>
	</monogr>
	<note>International Joint Conferences on Artificial Intelligence Organization</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Efficient Estimation of Word Representations in Vector Space</title>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kai</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Greg</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeffrey</forename><surname>Dean</surname></persName>
		</author>
		<idno>CoRR, abs/1301.3781</idno>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">An automatic method of finding topic boundaries</title>
		<author>
			<persName><forename type="first">C</forename><surname>Jeffrey</surname></persName>
		</author>
		<author>
			<persName><surname>Reynar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL</title>
				<imprint>
			<biblScope unit="page" from="331" to="333" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">/</forename><surname>Acl</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1994">1994</date>
			<publisher>Morgan Kaufmann Publishers</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">An iterative approach to text segmentation</title>
		<author>
			<persName><forename type="first">Fei</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">William</forename><forename type="middle">M</forename><surname>Darling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adnan</forename><surname>Duric</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fred</forename><forename type="middle">W</forename><surname>Kroon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECIR</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="volume">6611</biblScope>
			<biblScope unit="page" from="629" to="640" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Kernel Methods for Pattern Analysis</title>
		<author>
			<persName><forename type="first">John</forename><surname>Shawe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">-</forename><surname>Taylor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nello</forename><surname>Cristianini</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
			<publisher>Cambridge University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Human brain activity time-locked to narrative event boundaries</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">K</forename><surname>Speer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Reynolds</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychological Science</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="449" to="455" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Large margin methods for structured and interdependent output variables</title>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Tsochantaridis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thorsten</forename><surname>Joachims</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Hofmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yasemin</forename><surname>Altun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Machine Learning Reserach</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Segmentation of expository text by hierarchical agglomerative clustering</title>
		<author>
			<persName><forename type="first">Yaari</forename><surname>Yaakov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Recent Advances in NLP (RANLP&apos;97)</title>
				<imprint>
			<publisher>ACL</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A dnn-hmm approach to story segmentation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">S</forename><surname>Chng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">INTERSPEECH</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1527" to="1531" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Event segmentation</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Swallow</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Current Directions in Psychological Science</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="80" to="84" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">The brain&apos;s cutting-room floor: Segmentation of narrative cinema</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">K</forename><surname>Speer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Swallow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Maley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers in human neuroscience</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Event structure in perception and conception</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zacks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tversky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychological bulletin</title>
		<imprint>
			<biblScope unit="volume">127</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">3</biblScope>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
	<note>Appendix: a segmentation example: unit 4-6</note>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">Anton</forename><surname>Checkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Yepikhodov</surname></persName>
		</author>
		<title level="m">From The Cherry Orchard by</title>
				<imprint>
			<date>0034</date>
		</imprint>
	</monogr>
	<note>Stumbles against the table, which falls over.) There you are... (As if exulting in it.) You see what I&apos;m up against! I mean, it&apos;s simply amazing! (Goes out</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<author>
			<persName><forename type="first">Unit</forename><surname>Id</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">UNIT NAME: Dunyasha struts around</title>
				<imprint>
			<date>0004</date>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page">37</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title/>
		<author>
			<persName><surname>Dunyasha</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>To tell you the truth, he&apos;s proposed to me</note>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Lopakhin</forename><surname>Ah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">!</forename></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">what to say</title>
		<author>
			<persName><surname>Dunyasha. I Don't Know</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">UNIT NAME: Lopakhin and Dunyasha welcome the masters</title>
				<imprint>
			<date>0005</date>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page">42</biblScope>
		</imprint>
	</monogr>
	<note>He&apos;s all right, he doesn&apos;t give any trouble, it&apos;s just sometimes when he starts to talk you can&apos;t understand a word of it. It&apos;s very nice, and he puts a lot of feeling into it, only you can&apos;t understand it. I quite like him in a way, even. He&apos;s madly in love with me. He&apos;s the kind of person who never has any luck. Every day something happens. They tease him in our part of the house -they call him Disasters by the Dozen</note>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">I think they&apos;re coming</title>
		<author>
			<persName><surname>Lopakhin</surname></persName>
		</author>
		<imprint>
			<date>listens</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">They&apos;re coming! What&apos;s the matter with me? I&apos;ve gone all cold</title>
		<author>
			<persName><surname>Dunyasha</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Will she recognize me? Five years we haven&apos;t seen each other</title>
		<author>
			<persName><surname>Lopakhin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">They are indeed coming</title>
				<imprint/>
	</monogr>
	<note>Let&apos;s go and meet them</note>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<title/>
		<author>
			<persName><surname>Dunyasha</surname></persName>
		</author>
		<imprint>
			<biblScope unit="volume">43</biblScope>
		</imprint>
	</monogr>
	<note>in agitation. I&apos;ll faint this very minute. I will, I&apos;ll faint clean away! 0042 Two carriages can be heard coming up to the house. LOPAKHIN and DUNYASHA harry out. The stage is empty. UNIT ID: 0006, UNIT NAME: The owners settle down</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
