<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Commonsense Knowledge and Controllable Techniques for an Effective and Efficient Approach to Text Generation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Iván</forename><surname>Martínez-Murillo</surname></persName>
							<email>ivan.martinezmurillo@ua.es</email>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Software and Computing Systems</orgName>
								<orgName type="institution">University of Alicante</orgName>
								<address>
									<addrLine>Apdo. de Correos 99</addrLine>
									<postCode>E-03080</postCode>
									<settlement>Alicante</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Commonsense Knowledge and Controllable Techniques for an Effective and Efficient Approach to Text Generation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">046BC294EC97738B7D7DCFE1D6FE5718</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:05+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Generation</term>
					<term>Controllable techniques</term>
					<term>Hallucination</term>
					<term>Efficient architectures</term>
					<term>Taskagnostic</term>
					<term>Commonsense Knowledge</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The Natural Language Generation (NLG) field has advanced at a breakneck speed, favoured by the development of Large Language Models (LLMs). Notwithstanding, these models also have some drawbacks. On the one hand, these models can introduce some risks such as hallucination or bias which can be used in an unethical way to potentially generate dis-and mis-information. On the other hand, the expense of time and cost of training these models is too high. In account of this, the purpose of this paper is to propose a new research line for my PhD thesis. During the research, I will propose an efficient architecture, that could generate quality text in a controllable way, while integrating external commonsense knowledge. The objective is that this proposed architecture could achieve similar performance to state-of-the-art models while being more efficient.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Justification of the research</head><p>The rapid development of generative Artificial Intelligence (AI) has caused an augment of interest in society in AI tools. These tools can produce a positive impact in lots of areas, saving the time and effort of solving some tasks <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>.</p><p>In particular, state-of-the-art Natural Language Generation (NLG) tools can produce text that, in some cases, can be indistinguishable from human-generated texts. This could have lots of benefits in some sectors such as academia, tourism or marketing <ref type="bibr" target="#b3">[4]</ref>. Nonetheless, these tools also have some drawbacks. First of all, text generated by these tools may contain hallucinations, which is the phenomena that occur when a text is nonsensical or unfaithful to the provided source <ref type="bibr" target="#b4">[5]</ref>. Secondly, AI-generated text could be biased in some cases, which is the misrepresentation or attribution errors that result in favouring certain groups or ideas <ref type="bibr" target="#b5">[6]</ref>. Finally, these tools also lack of logical reasoning, a fact that it is essential to human intelligence <ref type="bibr" target="#b6">[7]</ref>. In the wake of these limitations, these tools can be used in a bad and unethical way to potentially generate dis-and mis-information.</p><p>Moreover, the core of these tools are Large Language Models (LLMs). The expense of time and cost needed to train these models is extremely high, being only within the reach of large companies.</p><p>Therefore, the motivation for the present research arises from the need in the academia to find efficient architectures that could produce text in a controlled manner, achieving a similar performance to state-of-the-art models, but solving the hallucination issue.</p><p>The remainder of this article is organised as follows: Section 2 presents an overview of the relevant literature concerning NLG; Section 3 shows the main hypotheses and objectives planned for this research; Finally, Section 4, and Section 5 detail the methodology this PhD will follow and some relevant research topics for discussion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background and Related Work</head><p>Before introducing my proposal, this section aims to contextualise this study within the state of the art of the NLG.</p><p>NLG is the subfield in the Natural Language Processing (NLP) area that aims to produce meaningful sentences to meet a communicative goal <ref type="bibr" target="#b7">[8]</ref>. Depending on several aspects of the generation, NLG can be classified according to two criteria:</p><p>• Type of input: Depending on the type of input, NLG can be catalogued as (1) text-to-text generation (T2T) and ( <ref type="formula">2</ref>) data-to-text generation (D2T) <ref type="bibr" target="#b8">[9]</ref>. In D2T, input data can assume different types such as binary data, images, voice, database, ontologies, etc. Recently, another concept of NLG has emerged, (3) none-to-text generation (N2T) <ref type="bibr" target="#b9">[10]</ref>, which corresponds to the generation to which no input is received. • Task typology: Based on the communicative goal, NLG can be grouped into (1) text abbreviation;</p><p>(2) text expansion; and (3) text rewriting and reasoning. Text abbreviation tasks consist in detecting the most important information in a text and fusing that information into a short text, e.g. text summarisation. Text expansion tasks aim to generate complete sentences from some meaningful words, e.g. topic-to-essay. Finally, Text rewriting and reasoning tasks try to rewrite a text into another style or apply reasoning methods, e.g. text simplification.</p><p>To achieve the communicative goal of these tasks, the NLG area has been studied for a long time. First researches date by the end of 1970 <ref type="bibr" target="#b10">[11]</ref>. Notwithstanding, it has not been until recent years that the NLG field has achieved an exponential improvement, producing text in a very similar way to humans. But, how did we get to this?</p><p>In a first stage, the NLG task was seen as a sequential scheme of four different stages (preprocessing, macroplanning, microplanning and realisation). Modular architectures followed this scheme, making a clear distinction between the distinct sub-tasks of each stage. The most famous modular architecture was proposed by Reiter <ref type="bibr" target="#b11">[12]</ref>. Figure <ref type="figure" target="#fig_0">1</ref> shows the sub-task division in this architecture.</p><p>Other works within this architecture can be found in <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16]</ref>.</p><p>Later, that clear distinction between the distinct sub-tasks became more flexible originating what is known as planning perspectives. This scheme was similar to the employed in modular architectures, but it allows to combine and implement two or more different sub-tasks in one sub-task, e.g. to combine text structuring and sentence aggregation sub-tasks. Some examples of this approach are present in <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24]</ref>.</p><p>Finally, the sub-task division started to disappear, originating global approaches. This type of architecture does not make a distinction among sub-tasks, performing every task as a whole, and relying on statistical learning and neural networks. Some proposed architectures within global approaches are: Graph Neural Networks <ref type="bibr" target="#b24">[25]</ref>, Generative Adversarial Nets <ref type="bibr" target="#b25">[26]</ref>, Recurrent Neural Networks <ref type="bibr" target="#b26">[27]</ref>, Pre-trained Models <ref type="bibr" target="#b27">[28]</ref>, Memory Networks <ref type="bibr" target="#b28">[29]</ref>, Transformers <ref type="bibr" target="#b29">[30]</ref> and Copy and Pointing Mechanism <ref type="bibr" target="#b30">[31]</ref>. This group of approaches have made the major development in the NLG area. The most important proposal in this group was the Transformers architecture and its concept of attention. Models based on this architecture achieve a high performance at NLG tasks. The best-performing models based on Transformers are LLMs such as GPT4 <ref type="bibr" target="#b31">[32]</ref> or LLaMa <ref type="bibr" target="#b32">[33]</ref>, which have neural networks with billions of parameters. Nowadays, most of the research in the industry is focused on developing bigger LLMs, as it is thought that a bigger LLM would achieve better performance. The cost and time of training these models are unassumable for the academia. On account of that issue, there is a need in the academia to find more efficient architectures that could perform similarly to LLMs.</p><p>Consequently, my line of work will focus on exploring efficient architectures that could generate text with similar results to state-of-the-art models. Moreover, controllable generation methods, techniques to integrate external commonsense knowledge and task-agnostic architectures will be studied in order to reduce the phenomena known as hallucination.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Main Hypothesis and Objectives</head><p>This PhD thesis is based on the hypothesis that integrating external commonsense knowledge along with controllable text generation techniques in an efficient architecture will help to reduce the hallucination issue, and besides performing similarly to state-of-the-art models. Thus, the main objective of this research is the proposal of an efficient architecture that could achieve a good performance in different NLG tasks, e.g. text summarisation, and text simplification, and could reduce hallucination as much as possible. In order to complete this main objective, several sub-objectives have been proposed:</p><p>• A1. To explore optimal controllable text generation techniques. The planned schedule of these sub-objectives can be seen in Figure <ref type="figure" target="#fig_2">2</ref>, starting from February 2023. Group A corresponds with the study and test of state-of-the-art techniques. After an initial study, during Group B, an efficient architecture will be proposed, tested and compared with other open-source architectures using a common benchmark. Finally, in Group C the proposed architecture will be adapted to perform in different NLG tasks. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology and proposed experiment</head><p>The proposed methodology to carry out this research is based on a complete and comprehensive training in all areas of NLG, including general training on NLP. After having the basic notions of NLG, the research focuses on an exhaustive analysis of the state of the art of NLG, especially on deep learning techniques that allow controlled language generation and integrate commonsense knowledge. Subsequently, the experimentation also starts, testing different open-source architectures, along with the most relevant studied techniques. After having tested several architectures, an efficient base model will be proposed, integrating commonsense knowledge and controllable generation techniques into it. Then, it will be evaluated against other architectures using a common benchmark. Finally, the proposed architecture will be adapted to perform different tasks.</p><p>At present, I am experimenting with the CommonGen dataset <ref type="bibr" target="#b33">[34]</ref>. The CommonGen dataset consists of a set of common concepts and some reference sentences using those concepts and the main idea is to test machines for the ability of generative commonsense reasoning. I am testing with different types of approaches such as SimpleNLG, Factorised Language Models, or Neural Models over this dataset. With the proposed experiment, the main idea is to combine the best-obtained architecture with controllable generation techniques in order to obtain a base model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Research issues to discuss</head><p>In order to advance towards an effective and efficient approach for controllable text generation, several research issues are suggested and briefly discussed.</p><p>What does controllable text generation mean, and what are the most efficient methods to incorporate it? Controllable text generation is the task of producing text in a way that its attributes can be controlled <ref type="bibr" target="#b34">[35]</ref>. These attributes can adopt a wide variety of ranges, such as stylistics, to include specific information in the content, based on the demographic attributes of the interlocutor, etc. As seen in <ref type="bibr" target="#b35">[36]</ref>, there are three ways to approach controllable text generation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Via hyperparameters:</head><p>Training data in LLMs can be unbalanced due to the fact that it is difficult to balance that huge amount of data. Modifying hyperparameters may generalise the knowledge better and consequently raise obtained results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Via additional input:</head><p>To fine-tune a pre-trained model with more information than just the text could enhance its performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Via conditional training:</head><p>Using internal control variables could enrich the generation with specific capabilities.</p><p>What is hallucination and what are the ways to reduce its occurrence? Hallucination in NLG occurs when a text generated by an AI lacks of coherence or deviates from the intended sense of the source input <ref type="bibr" target="#b4">[5]</ref>. It can be classified into two categories: intrinsic hallucinations, which appear when the generated text contradicts the source input, and extrinsic hallucinations which arise when the source input cannot substantiate the generated text.</p><p>There exist different types of approaches to minimise the occurrence of hallucinations. Firstly, constructing a reliable dataset, which does not contain any type of contradiction in the data. Secondly, modifying the encoder/decoder architecture can enhance the ability to better understand and represent the knowledge. Thirdly, proposing an optimal training strategy such as controllable text generation could benefit the model. Finally, one important approach is to integrate external commonsense knowledge into the models.</p><p>How to integrate external commonsense knowledge? Commonsense knowledge is an important factor in human communication, as it facilitates inference without the explicit mention of context <ref type="bibr" target="#b36">[37]</ref>. Although current state-of-the-art models exhibit some common sense abilities, it is not complete yet. Traditionally, commonsense has been injected into NLG systems in the form of rules and ontologies. Nowadays, the approaches have focused on injecting commonsense into neural NLG models through pre-trained models and using commonsense graphs. But there is still much work to do in this field in order to reach a complete commonsense knowledge.</p><p>Can a smaller architecture obtain similar performance than LLMs? There are some structures such as Plug and Play models or Variational Autoencoders that are more efficient than LLMs. Integrating commonsense knowledge and controllable generation techniques into these models could help to perform like LLMs while being smaller and more efficient models.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure1: Sub-task division in the modular architecture for the stages proposed by Reiter<ref type="bibr" target="#b7">[8]</ref> </figDesc><graphic coords="3,154.66,84.19,283.47,113.39" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>•</head><label></label><figDesc>A2. To examine hallucination mitigation techniques. • A3. To study how to integrate external commonsense knowledge. • A4. To analyse and test different task-agnostic architectures incorporating the previously studied techniques. • B1. To compare the performance of open-source state-of-the-art architectures using a common benchmark. • B2. To propose a cost-effective architecture that can generate text in a controllable way and evaluate it. • C1. To adapt the proposed architecture to perform in some NLG tasks, e.g., summarisation or text simplification.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: PhD project schedule</figDesc><graphic coords="4,154.66,315.09,283.47,85.04" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This research work is part of the R&amp;D projects "CORTEX: Conscious Text Generation" (PID2021-123956OB-I00), funded by MCIN/ AEI/10.13039/501100011033/ and by "ERDF A way of making Europe"</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Assessing the impact of generative ai on medicinal chemistry</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">P</forename><surname>Walters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Murcko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature biotechnology</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="143" to="145" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Mayahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vidrih</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2211.12660</idno>
		<title level="m">The impact of generative ai on the future of visual content marketing</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Examining science education in chatgpt: An exploratory study of generative artificial intelligence</title>
		<author>
			<persName><forename type="first">G</forename><surname>Cooper</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Science Education and Technology</title>
		<imprint>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">so what if chatgpt wrote it?&quot; multidisciplinary perspectives on opportunities, challenges and implications of generative conversational ai for research, practice and policy</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">K</forename><surname>Dwivedi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kshetri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hughes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">L</forename><surname>Slade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jeyaraj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Kar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Baabdullah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Koohang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Raghavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ahuja</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Information Management</title>
		<imprint>
			<biblScope unit="volume">71</biblScope>
			<biblScope unit="page">102642</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Survey of hallucination in natural language generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Frieske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ishii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">J</forename><surname>Bang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Madotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fung</surname></persName>
		</author>
		<idno type="DOI">10.1145/3571730</idno>
		<idno>doi:</idno>
		<ptr target="10.1145/3571730" />
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Ferrara</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2304.03738</idno>
		<title level="m">Should chatgpt be biased? challenges and risks of bias in large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Evaluating the logical reasoning ability of chatgpt and gpt-4</title>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Teng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2304.03439</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Building applied natural language generation systems</title>
		<author>
			<persName><forename type="first">E</forename><surname>Reiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Dale</surname></persName>
		</author>
		<idno type="DOI">10.1017/S1351324997001502</idno>
	</analytic>
	<monogr>
		<title level="j">Natural Language Engineering</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="57" to="87" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">La generación de lenguaje natural: análisis del estado actual</title>
		<author>
			<persName><forename type="first">M</forename><surname>Vicente</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Barros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">S</forename><surname>Peregrino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Agulló</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Lloret</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computación y Sistemas</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="721" to="756" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Positioning yourself in the maze of neural text generation: A task-agnostic survey</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Chandu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W</forename><surname>Black</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2010.07279</idno>
		<ptr target="https://arxiv.org/abs/2010.07279.doi:10.48550/ARXIV.2010.07279" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Natural language generation</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">D</forename><surname>Mcdonald</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Handbook of natural language processing</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="121" to="144" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Has a consensus nl generation architecture appeared, and is it psycholinguistically plausible?</title>
		<author>
			<persName><forename type="first">E</forename><surname>Reiter</surname></persName>
		</author>
		<idno type="arXiv">arXiv:cmp-lg/9411032</idno>
		<imprint>
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Computer generation of multiparagraph english text</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">C</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Moore</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">American Journal of Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="17" to="29" />
			<date type="published" when="1981">1981</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Generating natural language under pragmatic constraints</title>
		<author>
			<persName><forename type="first">E</forename><surname>Hovy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Pragmatics</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="689" to="719" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">W</forename><surname>Levelt</surname></persName>
		</author>
		<title level="m">Speaking: From intention to articulation mit press</title>
				<meeting><address><addrLine>Cambridge, MA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1989">1989</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Controlling a language generation planner</title>
		<author>
			<persName><forename type="first">S</forename><surname>Nirenburg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">R</forename><surname>Lesser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nyberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IJCAI</title>
				<imprint>
			<date type="published" when="1989">1989</date>
			<biblScope unit="page" from="1524" to="1530" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Strips: A new approach to the application of theorem proving to problem solving</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">E</forename><surname>Fikes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">J</forename><surname>Nilsson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial intelligence</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="189" to="208" />
			<date type="published" when="1971">1971</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Planning english sentences</title>
		<author>
			<persName><forename type="first">D</forename><surname>Appelt</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1985">1985</date>
			<publisher>cambridge university press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Approaches to the planning of coherent text</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Hovy</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1991">1991</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Enabling technology for multilingual natural language generation: the kpml development environment</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Bateman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Natural Language Engineering</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="15" to="55" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Sentence generation as a planning problem</title>
		<author>
			<persName><forename type="first">A</forename><surname>Koller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stone</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/P07-1043" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Association for Computational Linguistics</title>
				<meeting>the 45th Annual Meeting of the Association of Computational Linguistics, Association for Computational Linguistics<address><addrLine>Prague, Czech Republic</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="336" to="343" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Natural language generation as planning under uncertainty for spoken dialogue systems, Empirical Methods in Natural Language Generation: Data-oriented Methods and Empirical Evaluation</title>
		<author>
			<persName><forename type="first">V</forename><surname>Rieser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Lemon</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="105" to="120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Generating with discourse combinatory categorial grammar</title>
		<author>
			<persName><forename type="first">C</forename><surname>Nakatsu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>White</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Linguistic Issues in Language Technology</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Learning what to say and how to say it: Joint optimisation of spoken dialogue management and natural language generation</title>
		<author>
			<persName><forename type="first">O</forename><surname>Lemon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Speech &amp; Language</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="210" to="221" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">The graph neural network model</title>
		<author>
			<persName><forename type="first">F</forename><surname>Scarselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Tsoi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagenbuchner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Monfardini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE transactions on neural networks</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="61" to="80" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Generative adversarial nets</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mirza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Warde-Farley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ozair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pouget-Abadie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="2672" to="2680" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Sequence to sequence learning with neural networks</title>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Vinyals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Distributed representations of words and phrases and their compositionality</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">S</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">End-to-end memory networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sukhbaatar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Weston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.03762</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>See</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1704.04368</idno>
		<title level="m">Get to the point: Summarization with pointer-generator networks</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<title/>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.08774</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">Gpt-4 technical report</note>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Rozière</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hambro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Azhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rodriguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.13971</idno>
		<title level="m">Llama: Open and efficient foundation language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">CommonGen: A constrained text generation challenge for generative commonsense reasoning</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bhagavatula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ren</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.findings-emnlp.165" />
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2020, Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1823" to="1840" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Exploring controllable text generation techniques</title>
		<author>
			<persName><forename type="first">S</forename><surname>Prabhumoye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W</forename><surname>Black</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Salakhutdinov</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.coling-main.1</idno>
		<ptr target="https://aclanthology.org/2020.coling-main.1.doi:10.18653/v1/2020.coling-main.1" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th International Conference on Computational Linguistics, International Committee on Computational Linguistics</title>
				<meeting>the 28th International Conference on Computational Linguistics, International Committee on Computational Linguistics<address><addrLine>Barcelona, Spain (Online</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="14" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Neural natural language generation: A survey on multilinguality, multimodality, controllability and learning</title>
		<author>
			<persName><forename type="first">E</forename><surname>Erdem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kuyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Yagcioglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Parcalabescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Plank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Babii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Turuta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Erdem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Calixto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">73</biblScope>
			<biblScope unit="page" from="1131" to="1207" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Mahamood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Clinciu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gkatzia</surname></persName>
		</author>
		<title level="m">It&apos;s common sense, isn&apos;t it? demystifying human evaluations in commonsense-enhanced nlg systems</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
