<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Deep Contextual Punctuator for NLG Text</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Vandan</forename><forename type="middle">Mujadia</forename><surname>Pruthwik</surname></persName>
							<email>vandan.mu@research.iiit.ac.in</email>
							<affiliation key="aff0">
								<orgName type="department">Language Technologies Research Center Kohli Center On Intelligent Systems</orgName>
								<orgName type="institution">IIIT Hyderabad</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mishra</forename><surname>Dipti</surname></persName>
							<email>dipti@iiit.ac.in</email>
							<affiliation key="aff0">
								<orgName type="department">Language Technologies Research Center Kohli Center On Intelligent Systems</orgName>
								<orgName type="institution">IIIT Hyderabad</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Misra</forename><surname>Sharma</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Language Technologies Research Center Kohli Center On Intelligent Systems</orgName>
								<orgName type="institution">IIIT Hyderabad</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Deep Contextual Punctuator for NLG Text</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">327884D273E113FA3733216B5258A22F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T19:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes our team oneNLP's (LTRC, IIIT-Hyderabad) participation for the SEPP-NLG 2021 tasks 1 , Sentence End and Punctuation Prediction in NLG Text-2021. We applied sequence to tag prediction over contextual embedding as fine-tuning for both of these tasks. We also explored the use of multilingual Bert and multitask learning for these tasks on English, German, French and Italian.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Dataset</head><p>As a part of SEPP-NLG 2021, the organizers released an Europarl corpus of spoken texts by lower-</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Generally, the output of automatic speech recognition (ASR) systems ignore the prediction of punctuation marks. Similarly, output of OCR systems <ref type="bibr" target="#b9">(Nguyen et al., 2019)</ref> need automatic validation for punctuation. Apart from the omission of punctuation markers, some automatic tools generated texts e.g. PDF to text extraction may erroneously displace sentences for several reasons. Here, detecting the end of a sentence and placing an appropriate punctuation mark significantly improves the quality of such outputs by preserving the original meaning. Thus, missing punctuation or inappropriate punctuation degrade the readability of the presented text and leads to poor user experiences in real-world scenarios <ref type="bibr" target="#b2">(Che et al., 2016;</ref><ref type="bibr" target="#b12">Ueffing et al., 2013)</ref> as well as erroneous input to the subsequent automatic systems such as Machine Translation, Summarization, Question Answering, NLU etc. Therefore it is necessary to restore or correct punctuation marks for these automatic outputs.</p><p>Traditionally, automatic punctuation marking approaches <ref type="bibr" target="#b8">(Lu and Ng, 2010</ref>) can be divided into three broad categories <ref type="bibr" target="#b13">(Vandeghinste et al., 2018)</ref> based on the used features. They can be prosody based features <ref type="bibr" target="#b6">(Kim and Woodland, 2001;</ref><ref type="bibr" target="#b3">Christensen et al., 2001)</ref>, lexical features <ref type="bibr" target="#b0">(Augustyniak et al., 2020;</ref><ref type="bibr" target="#b10">Peitz et al., 2014)</ref> or combined or hybrid features of the previous two features based methods. Recent lexical based punctuation prediction methods build upon deep neural networks where it is modeled as a sequence to tag prediction task <ref type="bibr" target="#b7">(Li and Lin, 2020)</ref> or a sequence to sequence prediction task <ref type="bibr" target="#b13">(Vandeghinste et al., 2018)</ref>.</p><p>The simplest and basic form of punctuation prediction is the discovery of sentence boundaries, here the problem is the binary classification (where classes are period or empty as label). An incremental and a bit harder problem is the prediction of each individual punctuation, here the class labels for subtask2 are ": -, ? . 0" (0 indicating no punctuation). SEPP-NLG 2021 presents both these tasks as a challenge for the English, German, French and Italian languages.</p><p>In a recent advance of deep learning, pretrained language models such as ELMo <ref type="bibr" target="#b11">(Peters et al., 2018)</ref>, ULMFiT <ref type="bibr" target="#b5">(Howard and</ref><ref type="bibr">Ruder, 2018), OpenAI Transformer (Lee and</ref><ref type="bibr">Hsiang, 2020)</ref> and BERT <ref type="bibr" target="#b4">(Devlin et al., 2018)</ref> have resulted in a massive jump in the state-of-the-art performance for many NLP tasks, i.e text classification <ref type="bibr" target="#b1">(Büyüköz et al., 2020)</ref>, natural language inference and question-answering, dialogue system (Budzianowski and Vulić, 2019) etc. All these approaches pre-train an unsupervised language model on a large corpus of data such as all wikipedia articles, news articles and then fine-tune these pretrained models on different downstream tasks.</p><p>Here, for our experiments on the two punctuation prediction tasks, we try to use multi-lingual Bert and ALBert (for English) as a fine-tuning task along with the baseline experiments with CRF.  casing and removing all punctuations in the transcripts available in multiple languages. Table <ref type="table" target="#tab_0">1</ref> describes the corpora details for Training and Development corpus for all languages in terms of sentences and Tokens. Table <ref type="table" target="#tab_1">2</ref>, Table <ref type="table" target="#tab_2">3</ref>, Table <ref type="table" target="#tab_3">4</ref> and Table 5 describe the training corpora details in terms of punctuation classes and their respective distribution. Here data numbers are given after ''! ;'' are mapped to ''.''.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Lang</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Approach</head><p>We primarily used two broad categories of approaches. We model the problem as a sequence labeling task. In Machine Learning approaches, we trained a CRF model to identify the different kinds of labels correctly. Transformer based BERT fine tuning is also used as the other technique.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">CRF</head><p>We split the training data in English into sequences of 25 tokens each. This decision of setting the maximum sequence length to 25 was based on the average sentence length of the training data in English. We only used words as features and utilized a continuous window of 5 words over the full corpus as the required features for the CRF.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Class</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Fine-tuning Contextual Embedding</head><p>Multi-task learning (MTL) is a technique which aims to improve generalization, strengthen representations and enable adaptation in machine learning <ref type="bibr" target="#b14">(Worsham and Kalita, 2020)</ref> for related tasks. For our case, we enabled multi-task learning for our AlBert and mBert based contextual experiments as presented in Figure <ref type="figure" target="#fig_0">1</ref> where contextual embeddings are shared across the sub-tasks. We applied a sequence to tag classifier on the output contextualized token embeddings of Albert/mBert for the tag prediction. Here, we have used Albert<ref type="foot" target="#foot_1">2</ref> for English </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results and Discussion</head><p>As the results of CRF with word level features for English were poor (shown in table 7), we did not conduct CRF experiments on other languages.</p><p>We could observe that the results of Bert with Multi task learning is superior to the results of CRF. This is due to the better sentence or sequence representations learnt from the transformers. Simple surface level word features fail to capture the end sentence or punctuation markers in CRF.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion and Future work</head><p>We have successfully applied contextual embedding for the task of punctuation prediction and achieved comparable results on both of the subtasks. We believe that fine-tuning Bert on more data would benefit the overall punctuation task. Also,</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Dataset</head><p>Lang Pr Re F1 the language specific contextual embedding would improve performance in other languages. We will be incorporating both of these points in our future work.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Multi Task Learning</figDesc><graphic coords="3,120.56,62.79,356.42,154.21" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Training and Development Data Detail    </figDesc><table><row><cell></cell><cell cols="5">#Train Sents #Train Toks Avg Train Sent Len #Dev Sents #Dev Toks</cell></row><row><cell>English</cell><cell>1406577</cell><cell>33779095</cell><cell>24.015</cell><cell>321333</cell><cell>7743489</cell></row><row><cell>German</cell><cell>1308508</cell><cell>28645112</cell><cell>21.891</cell><cell>291443</cell><cell>6358683</cell></row><row><cell>French</cell><cell>1236504</cell><cell>32690367</cell><cell>26.438</cell><cell>332330</cell><cell>8781593</cell></row><row><cell>Italian</cell><cell>1132554</cell><cell>28167993</cell><cell>24.871</cell><cell>290089</cell><cell>7194189</cell></row><row><cell>Class</cell><cell>#Count</cell><cell>#Percentage</cell><cell></cell><cell></cell><cell></cell></row><row><cell>:</cell><cell>43133</cell><cell>0.128</cell><cell></cell><cell></cell><cell></cell></row><row><cell>-</cell><cell>80916</cell><cell>0.240</cell><cell></cell><cell></cell><cell></cell></row><row><cell>,</cell><cell>1759686</cell><cell>5.209</cell><cell></cell><cell></cell><cell></cell></row><row><cell>?</cell><cell>44290</cell><cell>0.131</cell><cell></cell><cell></cell><cell></cell></row><row><cell>.</cell><cell>1396166</cell><cell>4.133</cell><cell></cell><cell></cell><cell></cell></row><row><cell>0</cell><cell>30454904</cell><cell>90.159</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Total 33779095</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>English : Class Details for Training Data</figDesc><table><row><cell>Class</cell><cell>#Count</cell><cell>#Percentage</cell></row><row><cell>:</cell><cell>51192</cell><cell>0.179</cell></row><row><cell>-</cell><cell>81710</cell><cell>0.285</cell></row><row><cell>,</cell><cell>2208970</cell><cell>7.712</cell></row><row><cell>?</cell><cell>40511</cell><cell>0.141</cell></row><row><cell>.</cell><cell>1290282</cell><cell>4.504</cell></row><row><cell>0</cell><cell>24972447</cell><cell>87.179</cell></row><row><cell cols="2">Total 28645112</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>German : Class Details for Training Data</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 :</head><label>4</label><figDesc>French : Class Details for Training Data</figDesc><table><row><cell></cell><cell>#Count</cell><cell>#Percentage</cell></row><row><cell>:</cell><cell>46128</cell><cell>0.141</cell></row><row><cell>-</cell><cell>68523</cell><cell>0.210</cell></row><row><cell>,</cell><cell>1657880</cell><cell>5.071</cell></row><row><cell>?</cell><cell>41005</cell><cell>0.125</cell></row><row><cell>.</cell><cell>1223802</cell><cell>3.744</cell></row><row><cell>0</cell><cell>29653029</cell><cell>90.709</cell></row><row><cell cols="2">Total 32690367</cell><cell></cell></row><row><cell>Class</cell><cell>#Count</cell><cell>#Percentage</cell></row><row><cell>:</cell><cell>55080</cell><cell>0.196</cell></row><row><cell>-</cell><cell>52983</cell><cell>0.188</cell></row><row><cell>,</cell><cell>1503502</cell><cell>5.338</cell></row><row><cell>?</cell><cell>38807</cell><cell>0.138</cell></row><row><cell>.</cell><cell>1138669</cell><cell>4.042</cell></row><row><cell>0</cell><cell>25378952</cell><cell>90.099</cell></row><row><cell cols="2">Total 28167993</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 :</head><label>5</label><figDesc>Italian : Class Details for Training Data</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 :</head><label>6</label><figDesc>Subtask1 Results using BERT MTL</figDesc><table><row><cell></cell><cell>EN</cell><cell cols="2">0.92 0.92 0.92</cell></row><row><cell></cell><cell>DE</cell><cell cols="2">0.93 0.95 0.94</cell></row><row><cell>Test</cell><cell>FR</cell><cell cols="2">0.9 0.89 0.9</cell></row><row><cell></cell><cell>IT</cell><cell cols="2">0.88 0.89 0.89</cell></row><row><cell></cell><cell cols="3">AVG 0.91 0.91 0.91</cell></row><row><cell></cell><cell>EN</cell><cell cols="2">0.81 0.67 0.73</cell></row><row><cell></cell><cell>DE</cell><cell cols="2">0.85 0.72 0.78</cell></row><row><cell>Surprise Test</cell><cell>FR</cell><cell cols="2">0.77 0.62 0.69</cell></row><row><cell></cell><cell>IT</cell><cell cols="2">0.78 0.58 0.67</cell></row><row><cell></cell><cell cols="3">AVG 0.8 0.65 0.72</cell></row><row><cell></cell><cell>EN</cell><cell cols="2">0.92 0.92 0.92</cell></row><row><cell></cell><cell>DE</cell><cell cols="2">0.94 0.95 0.94</cell></row><row><cell>Dev</cell><cell>FR</cell><cell cols="2">0.9 0.89 0.9</cell></row><row><cell></cell><cell>IT</cell><cell cols="2">0.88 0.89 0.89</cell></row><row><cell></cell><cell cols="3">AVG 0.91 0.91 0.91</cell></row><row><cell cols="2">Subtask# Pr</cell><cell cols="2">Re F1-score</cell></row><row><cell>1</cell><cell cols="2">0.73 0.52</cell><cell>0.61</cell></row><row><cell>2</cell><cell cols="2">0.71 0.32</cell><cell>0.35</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7 :</head><label>7</label><figDesc>CRF Results of Subtask 1 and 2 for English Dev data</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://sites.google.com/view/sentence-segmentation/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://tfhub.dev/google/albert base/3</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://tfhub.dev/tensorflow/bert multi cased L-12 H-768 A-12/4</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">Łukasz</forename><surname>Augustyniak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Szymanski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mikołaj</forename><surname>Morzy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Zelasko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adrian</forename><surname>Szymczak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jan</forename><surname>Mizgajski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yishay</forename><surname>Carmiel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Najim</forename><surname>Dehak</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.05774</idno>
		<title level="m">Hello, it&apos;s gpt-2-how can i help you? towards the use of pretrained language models for task-oriented dialogue systems</title>
				<imprint>
			<date type="published" when="2019">2020. 2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Punctuation prediction in spontaneous conversations: Can Paweł Budzianowski and Ivan Vulić</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Analyzing elmo and distilbert on sociopolitical news classification</title>
		<author>
			<persName><forename type="first">Berfu</forename><surname>Büyüköz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ali</forename><surname>Hürriyetoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arzucan</forename><surname>Özgür</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Workshop on Automated Extraction of Sociopolitical Events from News 2020</title>
				<meeting>the Workshop on Automated Extraction of Sociopolitical Events from News 2020</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="9" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Punctuation prediction for unsegmented transcript based on word vector</title>
		<author>
			<persName><forename type="first">Xiaoyin</forename><surname>Che</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Cheng</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Haojin</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Meinel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC&apos;16)</title>
				<meeting>the Tenth International Conference on Language Resources and Evaluation (LREC&apos;16)</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="654" to="658" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Punctuation annotation using statistical prosody models</title>
		<author>
			<persName><forename type="first">Heidi</forename><surname>Christensen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yoshihiko</forename><surname>Gotoh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steve</forename><surname>Renals</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">Jacob</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ming-Wei</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenton</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristina</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Jeremy</forename><surname>Howard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sebastian</forename><surname>Ruder</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1801.06146</idno>
		<title level="m">Universal language model fine-tuning for text classification</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The use of prosody in a combined system for punctuation generation and speech recognition</title>
		<author>
			<persName><forename type="first">Ji-Hwan</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Philip</forename><forename type="middle">C</forename><surname>Woodland</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Seventh European conference on speech communication and technology</title>
				<imprint>
			<publisher>Jieh-Sheng Lee and Jieh Hsiang</publisher>
			<date type="published" when="2001">2001. 2020</date>
			<biblScope unit="volume">62</biblScope>
			<biblScope unit="page">101983</biblScope>
		</imprint>
	</monogr>
	<note>Patent claim generation by fine-tuning openai gpt-2</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A 43 language multilingual punctuation prediction neural network model</title>
		<author>
			<persName><forename type="first">Xinxing</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Edward</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Interspeech 2020</title>
				<meeting>Interspeech 2020</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1067" to="1071" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Better punctuation prediction with dynamic conditional random fields</title>
		<author>
			<persName><forename type="first">Wei</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hwee Tou</forename><surname>Ng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2010 conference on empirical methods in natural language processing</title>
				<meeting>the 2010 conference on empirical methods in natural language processing</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="177" to="186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Deep statistical analysis of ocr errors for effective post-ocr processing</title>
		<author>
			<persName><forename type="first">Thi-Tuyet-Hai</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adam</forename><surname>Jatowt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mickael</forename><surname>Coustaty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antoine</forename><surname>Nhu-Van Nguyen</surname></persName>
		</author>
		<author>
			<persName><surname>Doucet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM/IEEE Joint Conference on Digital Libraries (JCDL)</title>
				<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="29" to="38" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Better punctuation prediction with hierarchical phrase-based translation</title>
		<author>
			<persName><forename type="first">Stephan</forename><surname>Peitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Markus</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hermann</forename><surname>Ney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Int. Workshop on Spoken Language Translation (IWSLT)</title>
				<meeting>of the Int. Workshop on Spoken Language Translation (IWSLT)<address><addrLine>South Lake Tahoe, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Deep contextualized word representations</title>
		<author>
			<persName><forename type="first">Matthew</forename><forename type="middle">E</forename><surname>Peters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mohit</forename><surname>Iyyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matt</forename><surname>Gardner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenton</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luke</forename><surname>Zettlemoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of NAACL</title>
				<meeting>of NAACL</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Improved models for automatic punctuation prediction for spoken and written text</title>
		<author>
			<persName><forename type="first">Nicola</forename><surname>Ueffing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maximilian</forename><surname>Bisani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paul</forename><surname>Vozila</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Interspeech</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="3097" to="3101" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A comparison of different punctuation prediction approaches in a translation context</title>
		<author>
			<persName><forename type="first">Vincent</forename><surname>Vandeghinste</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lyan</forename><surname>Verwimp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joris</forename><surname>Pelemans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Wambacq</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st Annual Conference of the European Association for Machine Translation: 28-30</title>
				<meeting>the 21st Annual Conference of the European Association for Machine Translation: 28-30<address><addrLine>Universitat d&apos;Alacant, Alacant, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>European Association for Machine Translation</publisher>
			<date type="published" when="2018-05">2018. May 2018</date>
			<biblScope unit="page" from="269" to="278" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Multitask learning for natural language processing in the 2020s: Where are we going?</title>
		<author>
			<persName><forename type="first">Joseph</forename><surname>Worsham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jugal</forename><surname>Kalita</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">136</biblScope>
			<biblScope unit="page" from="120" to="126" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
