<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Let the Database Talk Back: Natural Language Explanations for SQL</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Stavroula</forename><surname>Eleftherakis</surname></persName>
							<email>seleftheraki@athenarc.gr</email>
							<affiliation key="aff0">
								<orgName type="institution">Athena Research Center</orgName>
								<address>
									<settlement>Athens</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Orest</forename><surname>Gkini</surname></persName>
							<email>orestg@athenarc.gr</email>
							<affiliation key="aff1">
								<orgName type="institution">Athena Research Center</orgName>
								<address>
									<settlement>Athens</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Georgia</forename><surname>Koutrika</surname></persName>
							<email>georgia@athenarc.gr</email>
							<affiliation key="aff2">
								<orgName type="institution">Athena Research Center</orgName>
								<address>
									<settlement>Athens</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Let the Database Talk Back: Natural Language Explanations for SQL</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">082240BC77982AB9D9985A6B717DD2A7</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T08:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Database interaction is often characterized as a non-trivial and timeconsuming process due to user's inexperience with the data or the query language. Therefore, there is a need for the databases to be able to "talk back" in order to assist the users during data exploration and eventually lead them to the desired results. In this paper, we tackle the problem of SQL-to-NL by extending the graph-based model of Logos <ref type="bibr" target="#b2">[3]</ref>. Our novel extensions include improvements in terms of the system's translation capabilities and the fluency of the generated explanations. Finally, we report several challenges, highlighted by experiments on different user cases, i.e, astronomy and policy making.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Nowadays, the availability of data coming from many and different application fields urges the deployment of sophisticated tools which help us effectively explore and extract knowledge out of this data. However, the variety of the data sources (astronomical, biomedical, etc.) as well as the complexity of query languages such as SQL and SPARQL often pose obstacles during data exploration due to the users' unfamiliarity with the database content or the query language. Explaining queries in text can help tackle both problems as it enables users to understand the SQL queries that are used to retrieve the answers through the data exploration process <ref type="bibr" target="#b7">[8]</ref>.</p><p>The problem of translating SQL queries to natural language, or SQL-to-NL in short, appears to be deceivingly simple, as SQL queries are using a restricted vocabulary, comprising SQL elements (clauses, operators, etc.) and database elements (i.e., relations, attributes), and there is no ambiguity in interpreting such elements. Hence, initially a straightforward translation appears adequate. Reality teaches us quite the opposite, as the resulting text should be accurate in capturing the respective SQL query, and effective allowing fast and unique interpretation of it. Achieving both of these qualities is very difficult and raises several technical challenges that need to be addressed <ref type="bibr" target="#b7">[8]</ref>. Furthermore, several different explanations can be generated for the same SQL query, making evaluation challenging.</p><p>Logos <ref type="bibr" target="#b2">[3]</ref> is a well-established system which translates SQL queries into narratives. In contrast to neural approaches, Logos' model can be interpreted. This gives us the advantage of offering precise translations, knowing exactly how translations are produced. Nevertheless, the system's translation capabilities are highly depended on the query's syntax, meaning that for each type of query the system should have the necessary tools to model it. Moreover, depended on the query's complexity (number of attributes, bridge tables, etc.) translations can become unnatural.</p><p>In our work, we provide NL explanations as part of a data exploration platform used by our collaborators representing different fields, including astrophysics and policy making <ref type="bibr" target="#b0">[1]</ref>. In this platform, NL explanations are provided for explaining the internal SQL queries generated by different system components used for recommendations, NL-to-SQL translation, and data exploration. Thus, we have to translate a variety of queries containing different clauses and operators. Furthermore, there is a need for these translations to be as natural is possible.</p><p>In an effort to tackle the above-mentioned challenges, we extend Logos to two important directions: (a) translation capabilities, where we focus on the system's ability of translating different types of queries, (b) fluent explanations, where we focus on improving the system's translations. Furthermore, we present experimental results as well as results of a user study over two different databases: astronomical data and policy-making data. Our evaluation shows the effectiveness of our approach and provides several insights regarding challenges that arise due to the ambiguous nature of the NL explanations including: (a) scoring textual explanations, and (b) generating explanations suitable for different groups of people.</p><p>The rest of the paper is organised as follows. In Section 2, related work is discussed. In Section 3, we provide background information on Logos. In Section 4, we present our novel extensions. In Section 5, the effectiveness of the system is explored. Finally, concluding remarks are provided in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>Existing approaches for SQL-to-NL can be broadly divided into two categories. Template-/rule-based approaches (e.g., <ref type="bibr" target="#b3">[4]</ref>, <ref type="bibr" target="#b4">[5]</ref>) require the design of special templates and rules that are used to compose sentences. Due to the non-ambiguous nature of the SQL queries, templates can help limit the syntactic variability of their output and smooth explanations out <ref type="bibr" target="#b6">[7]</ref>. Logos <ref type="bibr" target="#b2">[3]</ref> falls into this category.</p><p>The second line of research tackles the problem as neural machine translation and uses sequence-to-sequence (Seq2Seq) models (e.g., <ref type="bibr" target="#b1">[2]</ref>, <ref type="bibr" target="#b11">[12]</ref>) which treat both the natural language description and  the query as sequences. In contrast to the previous category, those models automatically learn how to translate queries without the need of predefined query patterns. However, they require plenty of training data and fine tuning and their effectiveness is still very low.</p><p>Lately, hybrid models (e.g., <ref type="bibr" target="#b9">[10]</ref>, <ref type="bibr" target="#b10">[11]</ref>) have been proposed to solve a similar problem to ours, i.e., data to natural language. Those models automatically learn templates and use them in order generate textual explanations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">BACKGROUND</head><p>Logos' approach to NL generation comprises graph representations of the database and the queries, template phrases associated with parts of the graph, and graph traversal in particular directions to compose the templates found on the way into the final text formation. We provide an overview of the main ingredients: the database graph, the query graph, labels, templates, and the algorithms.</p><p>The database graph captures the relationships between the database relations and attributes (nodes of the graph). Its edges are divided into three types: (a) membership edges (attributes to relations), representing attribute projections in queries, (b) selection edges (relations to attributes), representing attributes in predicates, and (c) join edges (attribute to attribute), representing joins.</p><p>The query graph of a query is the part of the database graph that the query refers to, extended with additional nodes (e.g., for functions) and edges (e.g., for group-by's) in order to capture the entire query meaning. Each node or edge of the query graph can be annotated with a label that signifies its meaning in natural language. There are default labels that can be used for any database. Additionally, labels can be provided by a domain expert.</p><p>Example. Let us consider the CORDIS<ref type="foot" target="#foot_0">1</ref> database, which stores information about research projects funded by the European Union. In Figure <ref type="figure" target="#fig_0">1</ref>, we see a subgraph of the corresponding database graph. Note that for simplicity, we do not show the attributes used to join the various relations. By default, the name of each node is also its label. We also see how edges are annotated with default labels. We can easily override these. For instance, for the relation PROJECT_MEMBERS, instead of the system's default label "project members", we can use the short label "participants".  NL explanations are created by traversing the query graph accompanied with a template mechanism <ref type="bibr" target="#b8">[9]</ref>. The template mechanism uses the provided labels (designer or default) to form meaningful phrases. A template label 𝑙 (𝑢) or 𝑙 ((𝑢, 𝑣)) is assigned to a node 𝑢 or a path (𝑢, 𝑣), respectively. For instance, a join template label may have the form: 𝑙 ((𝑢, 𝑣)) = 𝑙 (𝑢) + 𝑒𝑥𝑝𝑟 + 𝑙 (𝑢, 𝑣) + 𝑙 (𝑣), where 𝑢 and 𝑣 are relation nodes of the query graph. Using the example of the previous paragraph (see Figure <ref type="figure" target="#fig_1">2</ref>), a template for the path connecting the tables PEOPLE and PROJECTS is the following: 𝑙 ( (PEOPLE, PROJECTS)) = 𝑙 (PEOPLE) + " as " + "principal investigators of" +𝑙 (PROJECTS) = "people as principal investigators of projects" . Synthesis of the NL explanations is performed as graph traversal of the query graph. There are three traversal strategies: (a) Binary Search Tree (BST) algorithm, where the translation consists of a composition of clauses each one of them focusing on specific query semantics, (b) Multiple Reference Points (MRP) algorithm, where information from all parts of the query graph is blended in the translation, (c) Template composition (TMT) algorithm, where predefined, richer templates corresponding to different query parts are used in an effort to produce more concise translations. In what follows, our examples were created using the MRP traversal strategy and correspond to queries targeting the CORDIS database.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">TOWARDS RICHER TRANSLATIONS</head><p>Logos was extended in two different directions. The first direction aimed to enable the translation of more query types, while the other one aimed to create fluent explanations. For this purpose, we have upgraded the query parsing, graph generation, traversal strategies, and template mechanism.</p><p>Translation Capabilities. Our first goal was to extend the translation capabilities of Logos towards translating more clauses and operators. To that end, Logos should be able to analyze those new types of queries, enrich (if necessary) the query graph with new types of nodes and edges, and translate those new elements using any of the traversal strategies (BST, MRP, TMT). Thus, we have implemented changes in the system's parser, introduced new graph elements, and created new translation rules. As a result, the system is now capable of translating queries with the SELECT TOP and LIMIT clauses, as well as the (NOT) IN and (NOT) LIKE operators.</p><p>Once an input query is given to the system, the parser analyzes it and generates a parsing tree that stores important information for the creation of the query graph. For the parser, we have introduced to the system several new parsing nodes (e.g., for limit) capturing information about all the aforementioned clauses and operators.</p><p>Queries having SELECT TOP or LIMIT clauses are marked as "limited" and their associated limitation number is temporally stored in the system. The translation algorithm (BST, MRP, TMT) checks if the input query is limited and if that is the case an explanation is separately created and added to the end of the translation. As an example, we provide that of query 1 in Table <ref type="table" target="#tab_0">1</ref>.</p><p>In order for Logos to translate queries having IN and NOT IN operators, new types of query graph elements were created: the in edge, the not-in edge, and the value-list node. An illustrative example of this case is query 2 in Table <ref type="table" target="#tab_0">1</ref> whose query graph is given in Figure <ref type="figure" target="#fig_2">3</ref>. The translation algorithm detects those new elements of the query graph and produces a textual explanation.</p><p>Regarding the translation of queries having the (NOT) LIKE operator, we created two new types of edges: the like edge and the not-like edge. An example of this case is that of query 3 in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Our last improvement, in terms of translation capabilities, is the creation of a new query graph node, the star node. In combination with the function node that represents the COUNT operator, we now get explanations like that of query 4 in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Fluent Explanations. Our second goal was to improve the translations generated by the system in terms of fluency. This kind of changes require only the modification of the traversal strategies and the template mechanism.</p><p>While exploring the system's capabilities, we noticed that many database schemas include bridge tables, i.e., special tables used to transform many-to-many relationships into one-to-many relationships. Bridge tables (usually) appear in queries in order to join two tables. Although those tables play an important role in terms of data modeling, they cause "noisy" explanations. For instance, let us consider query 1 of Table <ref type="table" target="#tab_1">2</ref>. Apparently, the previously produced explanation of the system is unnatural. Bridge tables are manually stored in a designer table and excluded from the translation. The template mechanism has been modified so that if 𝑢, 𝑣, and 𝑤 are table nodes with 𝑣 corresponding to a bridge table, the template label of the path connecting 𝑢 to 𝑤 is of the form</p><formula xml:id="formula_0">𝑙 ((𝑢, 𝑤)) = 𝑙 (𝑢) + 𝑙 (𝑢, 𝑤) + 𝑙 (𝑤).</formula><p>Moreover, a mini dictionary has been developed providing the translation process with the plural form of all the attribute labels. B: "Create groups according to the title of topics. Find the number of projects and the title of topics associated with project topics associated with the projects. " A: "Find the titles of topics and the number of projects on these topics grouped by the titles of topics. " 3. SELECT m.member_name FROM project_members m, projects p WHERE m.project = p.unics_id; B: "Find the member name of project members associated with projects. " A: "Find participants participated in projects. "</p><p>Essentially, for every attribute node 𝑢 of the query graph we take (if that exists) the plural form 𝑝 of its label 𝑙 (𝑢), i.e., 𝑝 (𝑙 (𝑢)).</p><p>In addition, we have modified the way the MRP algorithm translates the group by clause. Previously the translation for this part would result in a separate sentence. Leveraging the nature of the algorithm which aims on generating translations by blending different parts of the query, we now blend this translation more naturally.</p><p>An example that captures all the aforementioned changes (bridge tables, plural form, group by on MRP) is that of query 2 in Table <ref type="table" target="#tab_1">2</ref>.</p><p>Lastly, we have improved the translation of queries which include only heading attributes <ref type="bibr" target="#b3">[4]</ref> in the SELECT clause. The heading attribute is the most characteristic attribute of its relation. As an example we offer that of query 3 in Table <ref type="table" target="#tab_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">EXPERIMENTAL RESULTS</head><p>The evaluation of Logos is divided into two parts: (a) the automated evaluation part, where we evaluate our results using the Bilingual Evaluation Understudy (BLEU) automated metric <ref type="bibr" target="#b5">[6]</ref>, and (b) the human evaluation part, where we evaluate our results using the help of SQL experts. The purpose of the first part is to use a wellestablished metric to show how good the NL explanations are, while the second part aims at evaluating qualitative aspects of the NL explanations, such as clarity and fluency.</p><p>For both types of evaluation, we created 28 queries (14 for the CORDIS database, and 14 for the SDSS<ref type="foot" target="#foot_1">2</ref> database) (see Tables <ref type="table" target="#tab_6">6  and 7</ref>, respectively). Using the MRP algorithm, we translated those queries twice. One time using the default version of the system, which works by considering only the database schema (internal knowledge), and one time using both the database schema and the designer tables (internal and external knowledge), which, as mentioned in the previous sections, store information about heading attributes, bridge tables, and node or edge labels. First of all, we want to investigate the effect of the designer tables to the translations. Moreover, we want to know how close to the ground truth (textual explanations given by SQL experts of the databases) the system's explanations are. In what follows, we denote the default version of the system as Logos v.1 and the version that takes the advantage of the designer tables as Logos v.2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Automated Evaluation</head><p>Automated evaluation was carried out to compare the generated explanations to the ground truth i.e., textual explanations of SQL queries given by SQL experts of the databases, members of the INODE project <ref type="bibr" target="#b0">[1]</ref>. The quality of the results is measured using the BLEU-4 score. BLEU is a score for comparing a candidate translation of text to one or more reference translations. The scores range between 0% and 100%. A score of 100% means that the estimated, by the system, explanation matches completely the ground truth.</p><p>The results are summarized in Table <ref type="table" target="#tab_2">3</ref>. We also report the minimum and the maximum BLEU score. Furthermore, we noticed a large variation between the scores; thus, we decided to report the median BLEU score per translation system instead of the average. For both databases, median scores are under 10%. Looking at the medians, we conclude that Logos v.2 produces translations closer to the ground truth than those obtained from Logos v.1. Especially for the CORDIS database, the median BLEU score of Logos v.2 is more than two times higher than that of Logos v.1.</p><p>The low scores do not indicate that the translations produced by Logos are not correct. BLEU scores work by counting matching n-grams in the candidate translation to n-grams in the reference text, where 1-gram or unigram would be each token and a bigram comparison would be each word pair. That means that the score is higher the more common parts a NL explanation has with the ground truth. Manual examination of the NL explanations that the two versions of Logos generated versus the ground truth showed that the automatically generated translations were in fact correct. However, they looked very different from the ground truth. Indicatively, we show the translations of query 9 (from Table <ref type="table" target="#tab_5">6</ref>).</p><p>CORDIS query 9 explanations:</p><p>-Logos v.1: "Find the institutions names of institutions associated with countries whose country name is France. "</p><p>-Logos v.2: "Find institutions located in countries whose name is</p><p>France. "</p><p>-Ground truth: "Show names of institutions from France. ".</p><p>We see that Logos would not necessarily produce translations the way that a human mind would produce. And even different people would provide different explanations for the same SQL query (albeit all correct). This shows the opportunity of enhancing the translation capabilities of the system with learning that not only leverages the database schema but is also performed on previously defined human translations. It also shows the challenge of creating automated evaluation metrics which do not judge the explanations quality only by estimating their similarity with the ground truth. These observations lead to the need of conducting human evaluation as well, which will be presented in the next subsection.</p><p>Focusing now on the results of CORDIS, we see that there is significant improvement on the translations of queries with id 3-6, 11, and 12. This is mainly due to the exclusion of bridge tables from the translation procedure and the heading attribute addition. For queries with id 9, and 10 we noticed a score reduction. Indicatively, looking at the translations of query 9 above, we observe that although the translation of Logos v.2 is more natural than that of Logos v.1, the presence of the sentence "names of institutions" in the translation of the latter leads to a higher BLEU score.</p><p>Let us now focus on the results of SDSS. The scores are lower than those of the CORDIS database. This is due to the nature of the SDSS database that uses abbreviated names and letter symbols in order to describe the content of its tables and attributes. For instance, "photoobj" instead of "photometric objects", or the letter "u" to denote the magnitude of a photometric object in "u" (ultraviolet) filter. During the experiments, we realized that by transforming those names and symbols into meaningful textual sentences, we increase the size of the explanations compared to the size of explanations provided by the astrophysicist SQL expert. This shows another challenge for the automatic generation of NL explanations: different styles of explanations may be given by domain experts in different fields. Thus, different explanations are suitable for different groups of people (e.g., astronomers and data scientists). For example, for query 14, the BLEU score of Logos v.2 is substantially lower than that of Logos v.1.</p><p>SDSS query 14 explanations:</p><p>-Logos v.1: "Find the u, g, r, i and z of photoobj associated with specobj whose class is QSO. "</p><p>-Logos v.2: "Find the magnitude u, magnitude g, magnitude r, magnitude i and magnitude z of photometric objects corresponding to spectroscopic objects whose class is QSO. "</p><p>-Ground truth: "Show me the u, g, r, i, z magnitudes of spectroscopic quasars. ".</p><p>We concluded that this kind of notation (abbreviated names, and letter symbols), for the attributes of the SDSS tables, is sometimes preferred over full descriptions. Lastly, it has been observed that the SDSS database includes many discrete variables (attributes) that define different types of objects, e.g., stars. A fine example of that case is that of query 7.</p><p>SDSS query 7 (Table <ref type="table" target="#tab_6">7</ref>) explanations:</p><p>-Logos v.1: "Find the specobjids of specobj whose subclass is OB and class is STAR. "</p><p>-Logos v.2: "Find spectroscopic objects whose spectroscopic subclass is OB and class is STAR. "</p><p>-Ground truth: "Find all spectroscopic stars which are massive and hot. ".</p><p>We see that both versions of Logos do not understand that "subclass = OB" and "class = STAR" means massive and hot stars. This justifies the low BLEU scores in both versions of the system. For this purpose, additional knowledge is required. We are in fact in the process of integrating an ontology that provides such mappings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Human Evaluation</head><p>For this experimental setting, an online survey was conducted. A total of 21 people, all SQL experts, participated in the survey. The experts were members of the INODE project <ref type="bibr" target="#b0">[1]</ref>. People who contributed to the creation of labels (ground truth explanation) were excluded from the evaluation process. From the pool of 28 SQL queries (see Tables <ref type="table" target="#tab_6">6 and 7</ref>), participants were asked to rate the textual explanations of 4 randomly chosen queries (2 per database). The queries were equally distributed to all participants. For each query, the participant rated 3 explanations (1 per translation): (a) the explanation produced by Logos v.1, (b) the explanation produced by Logos v.2, and (c) the ground truth explanation, resulting in a total of 84 explanations rated by humans.</p><p>The participants judged the quality of the translations on the 7-point Likert-scales. They provided ratings for:</p><p>• clarity: how clear and understandable the explanation is.</p><p>• fluency: how natural the explanation is.</p><p>• precision: how well the information of the provided SQL query is captured on its textual explanation.</p><p>A score of 7 to all the aforementioned categories means that the provided explanation is extremely clear, natural, and precise, respectively. We score neutral as a 4. Data associated with respondents that completed the survey in less than 5 minutes (half the approximate time for filling out the survey) were deleted. Furthemore, we deleted the data of participants which have selected the same response to every question, regardless of the question. After cleaning the data, we ended up having 2 different scores (per explanation, and feature), corresponding to 2 different participants. The final score of an explanation for a given feature is obtained by taking the average of the 2 different scores. Therefore, we ended up having 1 single score per explanation, and feature. Indicatively, in Table <ref type="table" target="#tab_4">5</ref> we show the data collected for the explanations of the CORDIS query with id 9, and the obtained final scores. From those final scores, 18 rating sets (2 databases x 3 translation systems x 3 features) consisting of 14 elements each (1 for every query), were created.</p><p>In Table <ref type="table" target="#tab_3">4</ref>, we show the averages of those sets, accompanied with their standard deviation between brackets. Logos v.2 leads to better translations in terms of clarity and fluency for both databases (average score increases). However, we see that these scores do not surpass those of the ground truth. An interesting observation is that as the explanations become clearer and more fluent, precision decreases. In other words, as the explanations become more natural, they tend to lose their ability to explicitly explain each part of their associated SQL query. Lastly, we see that the difference between the average fluency scores of Logos v.2 and Logos v.1, increases for the SDSS database. As mentioned in the previous Section, this is due to the nature of the SDSS database which has a less explainable database schema in terms of NL explanation. By adopting labels for the components of the database schema (tables, attributes, and joins), we increase the average fluency score (Logos v.2).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">CONCLUDING REMARKS</head><p>In this paper, the translation of SQL queries has been discussed as a potential solution to problems that rise in data exploration. In that vein, we have extended the graph-based model of Logos <ref type="bibr" target="#b2">[3]</ref>. The system's capabilities have been extended to two different directions: (a) that of translating different types of queries and (b) that of creating fluent explanations. To this end, we have implemented changes in the system's parser, introduced new types of nodes and edges for the query graph, created new translation rules and templates, and finally modified the available traversal strategies (BST, MRP, TMT). In addition, a data set composed of 28 queries coming from different user cases (astronomy, policy making) has been created to perform both automated and human evaluation.</p><p>The experiments highlighted the following two challenges: (a) the need for creating automated metrics which do not consider only the ground truth and (b) the need for generating different explanations for different groups of people. Regarding the latter, we observed that: (a) scientific related databases are far more difficult to be explained by natural language and (b) sometimes the scientists themselves ask for less detailed interpretations. Lastly, it has been observed that as the explanations become more natural, they tend to lose their ability to explicitly explain each part of the query. Depending on the application, one may want to have more precise or more natural explanations.</p><p>Having said all this, SQL to text is a hard problem to solve and there is no single solution. Different explanations may be suitable for different applications or even different groups of people, albeit all correct. As future work, we plan to improve our system by adding an ontology which provides mappings for the interpretation of several parts of the query graph.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">APPENDIX</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: A subgraph of the CORDIS database graph.</figDesc><graphic coords="2,53.80,83.69,240.25,90.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: A join on the CORDIS database graph.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: A query with IN and NOT IN operators.</figDesc><graphic coords="2,317.96,83.69,240.25,82.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2</head><label>2</label><figDesc>Figure 2 zooms in on the join path connecting relation PEOPLE with relation PROJECTS and provides a more detailed view. A possible label which expresses this connection is "principal investigators of", while the default one is "associated with". Such designer labels are stored in special tables called designer tables. □ NL explanations are created by traversing the query graph accompanied with a template mechanism [9]. The template mechanism uses the provided labels (designer or default) to form meaningful phrases. A template label 𝑙 (𝑢) or 𝑙 ((𝑢, 𝑣)) is assigned to a node 𝑢 or a path (𝑢, 𝑣), respectively. For instance, a join template label may have the form: 𝑙 ((𝑢, 𝑣)) = 𝑙 (𝑢) + 𝑒𝑥𝑝𝑟 + 𝑙 (𝑢, 𝑣) + 𝑙 (𝑣), where</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Example SQL queries and generated interpretations. SQL Query Interpretations 1. SELECT TOP 10 title FROM projects WHERE start_year=2021; "Find projects whose start year is 2021. Limit the results to top 10. " 2. SELECT total_cost, ec_max_contribution, framework_program, ec_fund_scheme FROM projects WHERE framework_program IN ('FP7') AND ec_fund_scheme NOT IN ('ERC-SG', 'ERC-CG'); "Find the total costs, max contributions, framework programs and fund schemes of projects whose framework program is FP7 and fund scheme is not ERC-SG and ERC-CG. " 3. SELECT title FROM topics WHERE title LIKE '%climate%'; "Find the titles of topics whose title is like %climate%. " 4. SELECT COUNT( * ) FROM projects WHERE start_year=2012; "Find the number of projects whose start year is 2012. "</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Example SQL queries and generated interpretations, where B and A denotes before and after, respectively. SQL Query Interpretations 1. SELECT pr. * FROM projects pr, project_subject_areas psa, subject_areas sa WHERE pr.unics_id = psa.project AND psa.subject_area = sa.code AND sa.title = 'Mathematics and Statistics'; B: "Find everything about projects associated with project subject areas, and for project subject areas associated with subject areas whose title is Mathematics and Statistics. " A: "Find everything about projects on subject areas whose title is Mathematics and Statistics. " 2. SELECT COUNT(p.title), t.title FROM topics t, project_topics pt, projects p WHERE t.code = pt.topic AND pt.project = p.unics_id GROUP BY t.title;</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>BLEU scores for the textual explanations of the 14 CORDIS queries, and the 14 SDSS queries.</figDesc><table><row><cell>Query</cell><cell cols="2">CORDIS BLEU Scores</cell><cell cols="2">SDSS BLEU Scores</cell></row><row><cell>ID</cell><cell cols="4">Logos v.1 Logos v.2 Logos v.1 Logos v.2</cell></row><row><cell>1</cell><cell>3.21</cell><cell>3.98</cell><cell>4.99</cell><cell>12.55</cell></row><row><cell>2</cell><cell>4.37</cell><cell>6.27</cell><cell>2.45</cell><cell>2.66</cell></row><row><cell>3</cell><cell>2.26</cell><cell>9.26</cell><cell>11.71</cell><cell>15.73</cell></row><row><cell>4</cell><cell>3.51</cell><cell>12.87</cell><cell>4.86</cell><cell>5.01</cell></row><row><cell>5</cell><cell>2.01</cell><cell>15.46</cell><cell>2.84</cell><cell>3.67</cell></row><row><cell>6</cell><cell>3.40</cell><cell>11.20</cell><cell>3.38</cell><cell>4.07</cell></row><row><cell>7</cell><cell>4.03</cell><cell>4.32</cell><cell>3.74</cell><cell>4.37</cell></row><row><cell>8</cell><cell>8.23</cell><cell>9.55</cell><cell>14.72</cell><cell>14.01</cell></row><row><cell>9</cell><cell>12.30</cell><cell>9.29</cell><cell>4.46</cell><cell>18.80</cell></row><row><cell>10</cell><cell>18.30</cell><cell>7.41</cell><cell>4.03</cell><cell>4.07</cell></row><row><cell>11</cell><cell>14.46</cell><cell>24.81</cell><cell>18.46</cell><cell>27.36</cell></row><row><cell>12</cell><cell>10.70</cell><cell>17.40</cell><cell>6.87</cell><cell>8.56</cell></row><row><cell>13</cell><cell>4.07</cell><cell>4.20</cell><cell>4.30</cell><cell>22.24</cell></row><row><cell>14</cell><cell>3.67</cell><cell>4.46</cell><cell>35.61</cell><cell>5.50</cell></row><row><cell>MAX</cell><cell>18.30</cell><cell>24.81</cell><cell>35.61</cell><cell>27.36</cell></row><row><cell>MIN</cell><cell>2.01</cell><cell>4.20</cell><cell>2.45</cell><cell>2.66</cell></row><row><cell>MEDIAN</cell><cell>4.05</cell><cell>9.28</cell><cell>4.66</cell><cell>7.03</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 :</head><label>4</label><figDesc>Average clarity, fluency, and precision per database, and translation system (their standard deviation in parentheses).</figDesc><table><row><cell>Features</cell><cell>Logos v.1</cell><cell cols="3">CORDIS Logos v.2 Ground truth Logos v.1</cell><cell>SDSS Logos v.2 Ground truth</cell></row><row><cell>Clarity</cell><cell cols="2">4.25 (1.29) 5.79 (1.13)</cell><cell>6.79 (0.31)</cell><cell cols="2">4.32 (1.11) 6.04 (0.77)</cell><cell>6.50 (0.57)</cell></row><row><cell>Fluency</cell><cell cols="2">3.64 (1.43) 4.75 (1.06)</cell><cell>6.86 (0.35)</cell><cell cols="2">3.39 (0.97) 5.50 (0.87)</cell><cell>6.57 (0.56)</cell></row><row><cell cols="3">Precision 6.04 (1.46) 5.79 (1.10)</cell><cell>5.18 (1.75)</cell><cell cols="2">6.50 (0.50) 6.21 (0.70)</cell><cell>5.18 (1.01)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 :</head><label>5</label><figDesc>Final scores per feature for the explanations of the CORDIS query with id 9.</figDesc><table><row><cell>Features</cell><cell>Rating</cell><cell cols="3">CORDIS QUERY ID 9 Logos v.1 Logos v.2 Ground truth</cell></row><row><cell></cell><cell>Expert A</cell><cell>2</cell><cell>6</cell><cell>7</cell></row><row><cell>Clarity</cell><cell>Expert B</cell><cell>4</cell><cell>6</cell><cell>7</cell></row><row><cell></cell><cell>Final Score</cell><cell>3</cell><cell>6</cell><cell>7</cell></row><row><cell></cell><cell>Expert A</cell><cell>1</cell><cell>3</cell><cell>7</cell></row><row><cell>Fluency</cell><cell>Expert B</cell><cell>4</cell><cell>4</cell><cell>7</cell></row><row><cell></cell><cell>Final Score</cell><cell>2.5</cell><cell>3.5</cell><cell>7</cell></row><row><cell></cell><cell>Expert A</cell><cell>7</cell><cell>7</cell><cell>7</cell></row><row><cell>Precision</cell><cell>Expert B</cell><cell>6</cell><cell>7</cell><cell>7</cell></row><row><cell></cell><cell>Final Score</cell><cell>6.5</cell><cell>7</cell><cell>7</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 :</head><label>6</label><figDesc>The 14 SQL queries of the CORDIS database. SELECT distinct t.title FROM projects pr, project_topics pt, topics t WHERE pr.unics_id = pt.project AND pt.topic = t.code AND pr.end_year = 2014; 5 SELECT p.full_name FROM people p, projects pr, project_topics pt, topics t WHERE p.unics_id = pr.principal_investigator AND pr.unics_id = pt.project AND pt.topic = t.code AND t.title = 'Systems'; 6 SELECT m.title FROM people p, projects pr, project_programmes pm, programmes m WHERE p.unics_id = pr.principal_investigator AND pr.unics_id = pm.project AND pm.programme = m.code AND p.full_name = 'Thomas Bell'; 7 SELECT p.acronym FROM projects p, project_members pm, institutions i, countries c WHERE p.unics_id = pm.project AND pm.institution_id = i.unics_id AND i.country_id = c.unics_id AND c.country_name = 'Greece';</figDesc><table><row><cell cols="2">ID CORDIS SQL Queries</cell></row><row><cell></cell><cell>SELECT sum(pm.ec_contribution) AS funding_received</cell></row><row><cell>1</cell><cell>FROM projects p, project_members pm</cell></row><row><cell></cell><cell>WHERE pm.project=p.unics_id AND p.framework_program='H2020';</cell></row><row><cell>2</cell><cell>SELECT c.name FROM institutions i, countries c WHERE c.unics_id=i.country_id AND i.name='Athena';</cell></row><row><cell></cell><cell>SELECT pr.title FROM projects pr, project_subject_areas</cell></row><row><cell>3</cell><cell>psa, subject_areas sa WHERE pr.unics_id = psa.project AND</cell></row><row><cell></cell><cell>psa.subject_area = sa.code AND sa.title = 'Mathematics and Statistics';</cell></row><row><cell>4</cell><cell></cell></row><row><cell></cell><cell>SELECT pe.full_name FROM projects pr, people pe</cell></row><row><cell>8</cell><cell>WHERE pr.principal_investigator = pe.unics_id AND</cell></row><row><cell></cell><cell>pr.start_year = 2014;</cell></row><row><cell></cell><cell>SELECT i.institutions_name FROM institutions i, countries</cell></row><row><cell>9</cell><cell>c WHERE i.country_id = c.unics_id AND c.country_name =</cell></row><row><cell></cell><cell>'France';</cell></row><row><cell></cell><cell>SELECT mb.member_name FROM project_members mb,</cell></row><row><cell>10</cell><cell>activity_types a WHERE a.code = mb.activity_type AND</cell></row><row><cell></cell><cell>a.description = 'Research Organisations';</cell></row><row><cell>11</cell><cell>SELECT count(p.title) FROM projects p GROUP BY p.start_year;</cell></row><row><cell>12</cell><cell>SELECT count(i.name) FROM institutions i, countries c WHERE i.country_id=c.unics_id GROUP BY c.name;</cell></row><row><cell cols="2">13 SELECT title FROM topics WHERE title like '%climate%';</cell></row><row><cell>14</cell><cell>SELECT count(p.title) FROM projects p WHERE p.start_year=2018;</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7 :</head><label>7</label><figDesc>The 14 SQL queries of the SDSS database.</figDesc><table><row><cell cols="2">ID SDSS SQL Queries</cell></row><row><cell>1</cell><cell>SELECT objid FROM photoobj WHERE clean=1;</cell></row><row><cell>2</cell><cell>SELECT specobjid, z FROM specobj WHERE class = 'QSO' AND zwarning = 0;</cell></row><row><cell>3</cell><cell>SELECT objid FROM photoobj WHERE ra &gt;185 AND ra &lt;185.1 AND dec &lt;5;</cell></row><row><cell>4</cell><cell>SELECT specobjid FROM specobj WHERE survey = 'segue2';</cell></row><row><cell>5</cell><cell>SELECT specobjid FROM specobj WHERE class = 'STAR' AND zwarning = 0;</cell></row><row><cell>6</cell><cell>SELECT s.specobjid FROM specobj as s WHERE s.subclass = 'STARFORMING' AND s.class= 'GALAXY';</cell></row><row><cell>7</cell><cell>SELECT s.specobjid FROM specobj as s WHERE s.subclass = 'OB' AND s.class= 'STAR';</cell></row><row><cell>8</cell><cell>SELECT * FROM photoobj WHERE ra &gt;100 and dec &lt;100 AND type = 3;</cell></row><row><cell>9</cell><cell>SELECT DISTINCT type FROM photoobj;</cell></row><row><cell cols="2">10 SELECT class, count(*) FROM specobj GROUP BY class;</cell></row><row><cell>11</cell><cell>SELECT g. * FROM specobj s, galspecline g WHERE s.specobjid = g.specobjid and ra &lt;185 AND dec &lt;25;</cell></row><row><cell>12</cell><cell>SELECT n. * FROM neighbors n, photoobj p WHERE p.objid = n.objid AND p.b = 1.072 and p.l = 174.535;</cell></row><row><cell>13</cell><cell>SELECT s.</cell></row></table><note>* FROM specobj s, galspecline g WHERE s.specobjid = g.specobjid; 14 SELECT p.u, p.g, p.r, p.i, p.z FROM specobj s, photoobj p WHERE s.bestobjid = p.objid AND s.class = 'QSO';</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://data.europa.eu/euodp/en/data/dataset/cordisH2020projects</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://www.sdss.org/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGMENTS</head><p>This work was supported by the European Union's Horizon 2020 research and innovation programme under grant agreement (No 863410). We also thank the Intelligent Open Data Exploration (IN-ODE) team for contributing on the system's evaluation.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">Sihem</forename><surname>Amer-Yahia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Georgia</forename><surname>Koutrika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frederic</forename><surname>Bastian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Theofilos</forename><surname>Belmpas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Braschler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ursin</forename><surname>Brunner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Diego</forename><surname>Calvanese</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maximilian</forename><surname>Fabricius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Orest</forename><surname>Gkini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Catherine</forename><surname>Kosten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Davide</forename><surname>Lanti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antonis</forename><surname>Litke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hendrik</forename><surname>Lücke-Tieke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Francesco</forename><forename type="middle">Alessandro</forename><surname>Massucci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tarcisio</forename><surname>Mendes De Farias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Mosca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Francesco</forename><surname>Multari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nikolaos</forename><surname>Papadakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dimitris</forename><surname>Papadopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yogendra</forename><surname>Patil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aurélien</forename><surname>Personnaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guillem</forename><surname>Rull</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana</forename><forename type="middle">Claudia</forename><surname>Sima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ellery</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dimitrios</forename><surname>Skoutas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Srividya</forename><surname>Subramanian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guohui</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kurt</forename><surname>Stockinger</surname></persName>
		</author>
		<title level="m">INODE: Building an End-to-End Data Exploration System in Practice</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note>Extended Vision</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Summarizing Source Code using a Neural Attention Model</title>
		<author>
			<persName><forename type="first">Srinivasan</forename><surname>Iyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Konstas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alvin</forename><surname>Cheung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luke</forename><surname>Zettlemoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL. The Association for Computer Linguistics</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Logos: a system for translating queries into narratives</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Kokkalis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Panagiotis</forename><surname>Vagenas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alexandros</forename><surname>Zervakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alkis</forename><surname>Simitsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Georgia</forename><surname>Koutrika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yannis</forename><surname>Ioannidis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data</title>
				<meeting>the 2012 ACM SIGMOD International Conference on Management of Data</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="673" to="676" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Explaining structured queries in natural language</title>
		<author>
			<persName><forename type="first">Georgia</forename><surname>Koutrika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alkis</forename><surname>Simitsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yannis</forename><forename type="middle">E</forename><surname>Ioannidis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICDE. IEEE</title>
				<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="333" to="344" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Sorry, i don&apos;t speak SPARQL: translating SPARQL queries into natural language</title>
		<author>
			<persName><forename type="first">Axel-Cyrille Ngonga</forename><surname>Ngomo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lorenz</forename><surname>Bühmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christina</forename><surname>Unger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jens</forename><surname>Lehmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Gerber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd international conference on World Wide Web</title>
				<meeting>the 22nd international conference on World Wide Web</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="977" to="988" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Bleu: a method for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">Kishore</forename><surname>Papineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Salim</forename><surname>Roukos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Todd</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wei-Jing</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th annual meeting of the Association for Computational Linguistics</title>
				<meeting>the 40th annual meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="311" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Building applied natural language generation systems</title>
		<author>
			<persName><forename type="first">Ehud</forename><surname>Reiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Dale</surname></persName>
		</author>
		<idno type="DOI">10.1017/S1351324997001502</idno>
		<ptr target="https://doi.org/10.1017/S1351324997001502" />
	</analytic>
	<monogr>
		<title level="j">Nat. Lang. Eng</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="57" to="87" />
			<date type="published" when="1997">1997. 1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">DBMSs Should Talk Back Too</title>
		<author>
			<persName><forename type="first">Alkis</forename><surname>Simitsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yannis</forename><forename type="middle">E</forename><surname>Ioannidis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIDR</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Synthesizing structured text from logical database subsets</title>
		<author>
			<persName><forename type="first">Alkis</forename><surname>Simitsis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Georgia</forename><surname>Koutrika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yannis</forename><surname>Alexandrakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yannis</forename><surname>Ioannidis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">EDBT</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="428" to="439" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Automated learning of templates for data-to-text generation: comparing rule-based, statistical and neural methods</title>
		<author>
			<persName><forename type="first">Chris</forename><surname>Van Der Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Emiel</forename><surname>Krahmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sander</forename><surname>Wubben</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th International Conference on Natural Language Generation</title>
				<meeting>the 11th International Conference on Natural Language Generation</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="35" to="45" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Learning neural templates for text generation</title>
		<author>
			<persName><forename type="first">Sam</forename><surname>Wiseman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stuart</forename><forename type="middle">M</forename><surname>Shieber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alexander</forename><forename type="middle">M</forename><surname>Rush</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1808.10122</idno>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">SQL-to-Text Generation with Graph-to-Sequence Model</title>
		<author>
			<persName><forename type="first">Kun</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lingfei</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhiguo</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yansong</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vadim</forename><surname>Sheinin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</title>
				<editor>
			<persName><forename type="first">Ellen</forename><surname>Riloff</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">David</forename><surname>Chiang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Julia</forename><surname>Hockenmaier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Jun'ichi</forename><surname>Tsujii</surname></persName>
		</editor>
		<meeting>the 2018 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="931" to="936" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
