<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Solving Morphological Analogies Through Generation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Kevin</forename><surname>Chan</surname></persName>
							<email>chan3@etu.univ-lorraine.fr</email>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Université de Lorraine</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">LORIA</orgName>
								<address>
									<postCode>F-54000</postCode>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Shane</forename><forename type="middle">P</forename><surname>Kaszefski-Yaschuk</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Université de Lorraine</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">LORIA</orgName>
								<address>
									<postCode>F-54000</postCode>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Camille</forename><surname>Saran</surname></persName>
							<email>camille.saran5@etu.univ-lorraine.fr</email>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Université de Lorraine</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">LORIA</orgName>
								<address>
									<postCode>F-54000</postCode>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Esteban</forename><surname>Marquer</surname></persName>
							<email>esteban.marquer@loria.fr</email>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Université de Lorraine</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">LORIA</orgName>
								<address>
									<postCode>F-54000</postCode>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Miguel</forename><surname>Couceiro</surname></persName>
							<email>miguel.couceiro@loria.fr</email>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Université de Lorraine</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">LORIA</orgName>
								<address>
									<postCode>F-54000</postCode>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Solving Morphological Analogies Through Generation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">BE5E43A86DB9957D0E9EEC1561949F16</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T07:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Morphological analogy</term>
					<term>Analogy solving</term>
					<term>Representation learning</term>
					<term>Word generation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This contribution is a first attempt at solving morphological analogies through generation, instead of relying on retrieval approaches. Our preliminary experiments show promising results for some languages and reveal the feasibility of the approach in generating solutions of analogical equations in the morphology setting.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Analogical proportions are understood as statements of the form "𝐴 is to 𝐵 as 𝐶 is to 𝐷" denoted 𝐴 : 𝐵 :: 𝐶 : 𝐷, and they are the basis of analogical inference. Analogical inference is a remarkable capability of human reasoning, and that has been used to solve hard reasoning tasks. To some extent, it can be thought of as transferring knowledge from a source domain to a different, but somewhat similar, target domain by relying simultaneously on similarities and dissimilarities. Analogy based reasoning (AR) is closely related to case-based reasoning and has gained increasing interest from the artificial intelligence (AI) community, and has shown its potential in multiple machine learning (ML) tasks such as classification, decision making and recommendation with competitive results <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref>. Furthermore, analogical inference can support data augmentation through analogical extension and extrapolation for model learning, especially in environments with few labeled examples <ref type="bibr" target="#b4">[5]</ref>. Also, it has been successfully applied to several classical NLP tasks such as machine translation <ref type="bibr" target="#b5">[6]</ref>, several semantic and morphological tasks <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref>, as well as (visual) question answering and solving puzzles and scholastic aptitude tests <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>There are two basic tasks associated with AR. The first is analogy detection that corresponds to the task of deciding whether a quadruple 𝐴, 𝐵, 𝐶, 𝐷 constitutes a valid analogical proportion. This task asks for a common theoretical framework. However, the notion of analogy is not consensual, and there have been several efforts that follow different axiomatic and logical approaches <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>. For instance, <ref type="bibr" target="#b13">[14]</ref> introduces the following 4 postulates in the linguistic context as a guideline for formal models of analogical proportions: symmetry (if 𝐴 : 𝐵 :: 𝐶 : 𝐷, then 𝐶 : 𝐷 :: 𝐴 : 𝐵), central permutation (if 𝐴 : 𝐵 :: 𝐶 : 𝐷, then 𝐴 : 𝐶 :: 𝐵 : 𝐷), strong inner reflexivity (if 𝐴 : 𝐴 :: 𝐶 : 𝐷, then 𝐷 = 𝐶), and strong reflexivity (if 𝐴 : 𝐵 :: 𝐴 : 𝐷, then 𝐷 = 𝐵). Such postulates appear reasonable in the word domain, but they can be criticized in other application domains <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref>.</p><p>The second basic task is analogy solving that refers to the task of extrapolating or generating, for a given triple 𝐴, 𝐵, 𝐶 the value 𝑋 such that 𝐴 : 𝐵 :: 𝐶 : 𝑋 is a valid analogy. One approach to tackling this task is by retrieval and adaptation, i.e., defining an 𝑋 from a pool of retrieved candidate solutions to be suitably adapted. In fact, analogy solving is somewhat related to case-based reasoning (CBR) <ref type="bibr" target="#b16">[17]</ref> where, given a set 𝑃 of problems, a set 𝑆 of solutions and a set 𝒞 of cases (𝑥, 𝑦) ∈ 𝑃 × 𝑆, the CBR task is to find a solution 𝑦 𝑡 to a given target problem 𝑥 𝑡 . CBR basically consists in (1) selecting 𝑘 source cases in the case base according to some criteria related to the target problem (retrieval step), and (2) reusing the 𝑘 retrieved cases for proposing a target solution (adaptation step). Despite being a reasonable approach in controlled settings, it suffers from several drawbacks: it requires a suitable choice of examples and is intrinsically limited by case based approaches, that prevent creative inference and innovation.</p><p>More recent approaches to analogy solving take advantage of recent deep neural network frameworks that rely on vector representations and on the structure of the underlying multidimensional space. Essentially, analogical proportions are formalized in terms of the parallelogram rule by which four vectors 𝑒 𝐴 , 𝑒 𝐵 , 𝑒 𝐶 , and 𝑒 𝐷 (representing four elements 𝐴, 𝐵, 𝐶, and 𝐷) are in analogical proportion if 𝑒 𝐷 − 𝑒 𝐶 = 𝑒 𝐵 − 𝑒 𝐴 . Such an arithmetic view of analogical proportions has been used since the first works on analogy <ref type="bibr" target="#b17">[18]</ref>, and it was the key element in the methodology employed by earlier neural-based approaches <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>. In the absence of a decoder, the authors implicitly generate a representation 𝑒 𝑋 and then retrieve the closest candidate 𝐷 from the vocabulary to solve the analogical equation 𝐴 : 𝐵 :: 𝐶 : 𝑋 (see brief discussion of Subsection 2.2). However, Chen et al. <ref type="bibr" target="#b20">[21]</ref> argue that the latter two methods significantly differ from human performance.</p><p>In the case of sentence analogies (i.e., where 𝐴, 𝐵, 𝐶 are sentences), <ref type="bibr" target="#b21">[22]</ref> overcomes this issue by training a decoder that is then used to decode 𝑒 𝑋 . In this paper, we employ a similar approach in the setting of word analogies. More precisely, following the tracks of <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25]</ref>, we address morphological issues on words and tackle the problem of solving morphological analogies. Inspired by the work of <ref type="bibr" target="#b21">[22]</ref> to solving sentence analogies, the novelty in our contribution is to make use of autoencoders to solving morphological analogies on words. More precisely, the main contributions of this paper are as follows: (i) we propose a model to generate words at character level from word embeddings with high reconstruction performance, and (ii) we achieve encouraging results to solving morphological analogies by generation, thus indicating the feasibility of the approach. Nonetheless, this constitutes ongoing research that requires further investigations.</p><p>The paper is organized as follows. We first briefly survey previous work on both main tasks dealing with morphological analogies in Section 2. We then describe the key components of the deep learning architecture as well as the analogy solving procedure we use in Section 3. The empirical setting setting is then presented in Section 4 where we also discuss the experimental results. We conclude with a general overview of this contribution in Section 5 and propose further directions of future research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Approaches</head><p>In this paper, we focus on morphological analogies, i.e., analogies on words 𝐴, 𝐵, 𝐶, and 𝐷 that capture morphological transformations of words (e.g., conjugation or declension). In this section we introduce key approaches of analogy detection and solving in morphology. The main trend follows the seminal work of <ref type="bibr" target="#b25">[26]</ref> by exploiting the postulates of analogical proportions mentioned in introduction, but some approaches including ours take a slightly different approach. As deep learning approaches to morphological analogies are strongly related to approaches on semantic word analogies, the latter will also be discussed here.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Analogy Detection</head><p>As mentioned above, the analogy detection task corresponds to classifying quadruples 𝐴, 𝐵, 𝐶, 𝐷 into valid or invalid analogies. The tools in <ref type="bibr" target="#b26">[27]</ref> detect morphological analogies using the number of characters occurrences and the length of the longest common subword. Their approach is designed to generate analogical grids, i.e., matrices of transformations of various words, similar to paradigm tables in linguistics <ref type="bibr" target="#b6">[7]</ref>. A data-driven alternative was implemented by <ref type="bibr" target="#b7">[8]</ref> for semantic word analogies. Using a dataset of semantic analogies, they learn a neural network to classify quadruples 𝐴, 𝐵, 𝐶, 𝐷 into valid or invalid analogies, using their embedding 𝑒 𝐴 , 𝑒 𝐵 , 𝑒 𝐶 , and 𝑒 𝐷 . This approach was applied to morphological analogies in <ref type="bibr" target="#b23">[24]</ref> by replacing the GloVe <ref type="bibr" target="#b27">[28]</ref> semantic embeddings used by Lim et al. with a morphology-oriented word embedding model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Analogy Solving</head><p>Approaches to analogy solving usually generate the fourth element to solve the analogy, but it is also possible to leverage a list of candidates and retrieve the most fitting fourth term to solve the analogy. In Subsubsection 2.2.2 we describe key approaches using the former method to solve morphological analogies, and similarly in Subsubsection 2.2.1 for the latter method. Many approaches in embedding spaces use the latter method because generation from an embedding space can be challenging, and we describe some in Subsubsection 2.2.1. However, such retrieval approaches are limited to the available vocabulary and are unable to perform analogical innovation, despite it being a key mechanism in the evolution of languages <ref type="bibr" target="#b28">[29,</ref><ref type="bibr" target="#b29">30]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Retrieval</head><p>Analogy solving on word embeddings has been around since early works on Latent Semantic Analysis <ref type="bibr" target="#b30">[31]</ref> and word embeddings <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b22">23]</ref>, in which examples like 𝑘𝑖𝑛𝑔 − 𝑚𝑎𝑛 + 𝑤𝑜𝑚𝑎𝑛 = 𝑞𝑢𝑒𝑒𝑛 have been used do demonstrate the ability to encode semantic features in the word representation. These examples can be formulated as analogical equations 𝑚𝑎𝑛 : 𝑤𝑜𝑚𝑎𝑛 :: 𝑘𝑖𝑛𝑔 : 𝑋, for which the solution is retrieved among a vocabulary of candidate words. In <ref type="bibr" target="#b22">[23]</ref>, the authors use morphological <ref type="foot" target="#foot_0">1</ref> analogies to demonstrate that some word embedding models encode a degree of morphological information. Two of the most used methods for solving analogies in embedding spaces by retrieval are 3CosAdd <ref type="bibr" target="#b19">[20]</ref> and 3CosMul <ref type="bibr" target="#b31">[32]</ref>. In 3CosAdd, the solution 𝑋 is retrieved from the vocabulary by minimizing the cosine distance 𝑐𝑜𝑠(𝑒 𝑤𝑜𝑟𝑑 , 𝑒 𝑋 ), with 𝑒 𝑋 = 𝑒 𝐶 − 𝑒 𝐴 + 𝑒 𝐵 and 𝑒 𝐴 , 𝑒 𝐵 , 𝑒 𝐶 , and 𝑒 𝑋 the embeddings of 𝐴, 𝐵, 𝐶, and 𝑋. 3CosMul follows a similar intuition but we refer the reader to <ref type="bibr" target="#b31">[32]</ref> for a detailed description. However, the quality of the solution produced by the methods described above have been criticized by <ref type="bibr" target="#b18">[19]</ref> for being far from human performance in some cases. Nonetheless, frameworks based on analogy datasets like those mentioned in <ref type="bibr" target="#b7">[8]</ref> appear to bridge this gap in performance. By replacing the arbitrary formula by a learned estimator, Lim et al. significantly improved performance on solving semantic word analogies. This latter approach was adapted to morphological word analogies in <ref type="bibr" target="#b24">[25]</ref> and outperforms the generative methods described in Subsubsection 2.2.2. Those two approaches rely on the postulates of analogical proportions, and achieve high analogy solving performance.</p><p>While the approach by Marquer et al. has state of the art performance on solving analogical equations in morphology, it suffers from the limitations of retrieval approaches: the solutions are retrieved from a de facto finite vocabulary and analogical innovation is impossible. By using a generative deep learning model, the present work aims to maintain state of the art performance while solving the limitation of retrieval approaches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Generation</head><p>In <ref type="bibr" target="#b32">[33]</ref>, the author uses the postulates of <ref type="bibr" target="#b25">[26]</ref> to address multiple characteristics of words, such as their length, the occurrence of letters and of patterns. Based on these features, Lepage proposes an algorithm to solve analogies between character strings. Following the results of <ref type="bibr" target="#b33">[34]</ref> about closed form solutions, the Alea algorithm <ref type="bibr" target="#b5">[6]</ref> proposes a Monte-Carlo estimation of the solutions of an analogical equation by sampling among multiple sub-transformations. Those sub-transformations are obtained by considering the words as bags of characters and generating permutations of characters that are present in 𝐵 but not in 𝐴 on one side, and characters of 𝐶 on the other. Intuitively, if we consider 𝑏𝑎𝑔(𝐴) the bag of characters in 𝐴, Alea considers 𝑏𝑎𝑔(𝐷) = (𝑏𝑎𝑔(𝐵) − 𝑏𝑎𝑔(𝐴)) + 𝑏𝑎𝑔(𝐶) and thus 𝐷 is a permutation of the characters of 𝑏𝑎𝑔(𝐷). Recently, a more empirical approach was proposed by <ref type="bibr" target="#b8">[9]</ref>, which does not rely on the axioms of analogical proportions. The generation model proposed by the authors considers some transformation 𝑓 such that 𝐵 = 𝑓 (𝐴) and 𝑓 (𝐶) is computable. The simplest transformation 𝑓 is usually the one human use to solve analogies <ref type="bibr" target="#b8">[9]</ref>, and is found by minimizing the Kolmogorov complexity of 𝑓 . This complexity is estimated by first expressing 𝑓 using a language of operations (insertion, deletion, etc.), and computing the length of the resulting program. Unlike Alea, Kolmo is able to handle mechanisms like reduplication (repeating part of a word).</p><p>Recently, <ref type="bibr" target="#b21">[22]</ref> proposed a generation framework to solving sentence analogies. They use an autoencoder model (named ConRNN) trained to reconstruct sentences, and perform simple First, a sentence (as a sequence of words) is used as input to an encoder RNN, and the last hidden state of the RNN is used as the sentence embedding. The latter is then fed to a decoder RNN that tries to predict the words of the input sentence. The use of a generative model achieves significantly better results than previous retrieval approaches on the same embedding space. The current work aims to extend the one of <ref type="bibr" target="#b24">[25]</ref> by replacing the retrieval by the generation of the solution of morphological analogical equations. To do so, it is necessary to generate words at the character level from fixed-size embeddings, however in the literature there is to our best knowledge no approach proposed to tackle this specific issue. Inspired by the success of <ref type="bibr" target="#b21">[22]</ref>, we propose a character-level autoencoder for words and display its performance in solving morphological analogies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Our Approach</head><p>In this section we present the approach we use, illustrated in Figure <ref type="figure" target="#fig_0">1</ref>. The architecture for our model is a character-level sequence-to-sequence autoencoder model, based on the model described in <ref type="bibr" target="#b34">[35]</ref>. In order to properly decode the final vector solution, the model is trained to encode words and then decode the resulting vector back into the same word. Each character in a word 𝑤 is encoded into a one-hot vector and is then fed into the encoder, which uses a Bidirectional Long Short Term Memory (BiLSTM) layer. This layer outputs four the last hidden state ℎ 𝑓 and cell state 𝑐 𝑓 in the forward direction, and similarly ℎ 𝑏 and 𝑐 𝑏 for the backward direction. The concatenation of these vectors 𝑒 𝑤 = concat(ℎ 𝑓 , ℎ 𝑏 , 𝑐 𝑓 , 𝑐 𝑏 ) is the embedding of the word. The decoder is a regular LSTM layer, followed by a dense layer with softmax activation. The input for the first step of the decoder is the above-mentioned embedding, split into two states ℎ = concat(ℎ 𝑓 , ℎ 𝑏 ) and 𝑐 = concat(𝑐 𝑓 , 𝑐 𝑏 ). During training, we use teacher forcing: (i) the characters of the word 𝑤 to predict are used as input, with an added beginning-of-word (BOW) character at the begining; (ii) the prediction targets are the characters of 𝑤, but ahead by one time-step and with an end-of-word (EOW) character at the end.</p><p>To compute the solution of an analogy 𝐴 : 𝐵 :: 𝐶 : 𝑋, the embeddings 𝑒 𝐴 , 𝑒 𝐵 , and 𝑒 𝐶 are computed by the encoder and used to compute 𝑒 𝑋 = 𝑒 𝐵 − 𝑒 𝐴 + 𝑒 𝐶 . Then, 𝑒 𝑋 is decoded into a word 𝑋 by the decoder. Beginning with the BOW character, at each time-step the sampled character with the highest probability of occurrence is added to the word until either the EOW character is predicted or the length of the word is the same as the longest word in the dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head><p>In this section we present our experimental setup. First, the dataset we use is described in Subsection 4.1. We then report the performance of our model in the autoencoder setting in Subsection 4.2. The analogy solving performance of our approach is compared with baselines in Subsection 4.3. Finally, we discuss the overall performance of the model in Subsection 4.4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Datasets</head><p>For our experiments, we used the analogies from 8 languages available in the Siganalogies dataset <ref type="bibr" target="#b35">[36]</ref>: Arabic, English, French, German, Hungarian, Portuguese, Russian, and Spanish extracted from the high resource languages of Sigmorphon2019 <ref type="bibr" target="#b36">[37]</ref>. These languages were chosen such that, in later stages of the work, the authors have enough linguistic knowledge to interpret the model outputs. In order to obtain train and test sets, non-overlaping random subsets of analogies from the entire Sigmorphon dataset for a given language are taken, to ensure that no analogy is seen in both the train and test sets.</p><p>The Siganalogies dataset also provides a method for data augmentation via permutating the four words in a given analogy. These permutations are obtained using the symmetry and central permutation postulates of analogy. From a base form 𝐴 : 𝐵 :: 𝐶 : 𝐷, we generate 7 permutations:</p><p>• 𝐴 : 𝐶 :: 𝐵 : 𝐷; • 𝐷 : 𝐵 :: 𝐶 : 𝐴; • 𝐶 : 𝐴 :: 𝐷 : 𝐵; • 𝐶 : 𝐷 :: 𝐴 : 𝐵; • 𝐵 : 𝐴 :: 𝐷 : 𝐶;</p><p>• 𝐷 : 𝐶 :: 𝐵 : 𝐴;</p><p>• 𝐵 : 𝐷 :: 𝐴 : 𝐶.</p><p>In Siganalogies, the base forms 𝐴 : 𝐵 :: 𝐶 : 𝐷 are such that 𝐵 is an inflected form of 𝐴 and 𝐷 is inflected from 𝐶. In addition to that, base forms 𝐴 : 𝐴 :: 𝐵 : 𝐵 derived from the identity postulate (𝐴 : 𝐴 :: 𝐵 : 𝐵 is true for all 𝐴 and 𝐵) are present. An example of analogy in English is dog : dogs :: cat : cats, another in French is revérifier : revérifiasse :: tormenter : tormentasse, and in German there is Donor : Donor :: Herstellungsverfahren : Herstellungsverfahren (identity, but also accusative singular declension of the noun).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Autoencoder performance</head><p>As shown in Table <ref type="table" target="#tab_0">1</ref>, our autoencoder achieves very high accuracy in decoding vectors back into words, meaning that any wrong solutions are a result of the operations performed on the analogy rather than the decoding process. In our experiments, the model encodes the words as 128-dimensional vectors and there is a 0.1 dropout on the decoder LSTM layer. The loss function used is categorical cross-entropy, since a probability for the likelihood of each character appearing is required at each time-step. An 80/20 train/validation split is used, and the validation loss was the metric used for the early stopping. If the early stopping is not triggered, the model is trained for 100 epochs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Analogy solving performance</head><p>Two metrics are used to determine the performance of our model. The first one is a variation of Levenshtein distance, which calculates the minimum number of edits required to change one sequence into another using insertions, deletions, and substitutions. In order to display how close the decoded analogy solutions are to the expected analogy solutions, the Levenshtein distance 𝐿 was normalized using the length of the manipulated words into a percentage 𝐿 𝑝 like so:</p><formula xml:id="formula_0">𝐿 𝑝 (𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑, 𝑑𝑒𝑐𝑜𝑑𝑒𝑑) = 1 − 𝐿(𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑, 𝑑𝑒𝑐𝑜𝑑𝑒𝑑) max (𝑙𝑒𝑛 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 , 𝑙𝑒𝑛 𝑑𝑒𝑐𝑜𝑑𝑒𝑑 ) Table 2</formula><p>Results for 8 languages for 10,000 base analogies and all of their permutations (80,000 analogies in total). We report 𝐿 𝑝 in % and the accuracy (Acc.) in %. Our autoencoder was trained for 100 epochs on 40,000 random words per language. Baselines Alea <ref type="bibr" target="#b5">[6]</ref> and Kolmo <ref type="bibr" target="#b8">[9]</ref> were tested in the same setting. The accuracy of the retrieval model ANNr <ref type="bibr" target="#b24">[25]</ref> is reported as mean ± standard deviation for 10 random initialization, but note that these results are not completely comparable with our approach as they were obtained in a closed setting. The resulting percentage measures the rate of correctly decoded characters per word -when it is 1 (or 100%), then the decoded solution matches the expected solution perfectly. The second metric, accuracy, was calculated by dividing the number of correctly decoded analogies by the total number of analogies for each language. We report the results of decoding a test set of 10,000 base analogies and their permutations in Table <ref type="table">2</ref>. We compare our approach with Alea <ref type="bibr" target="#b5">[6]</ref> and Kolmo <ref type="bibr" target="#b8">[9]</ref> described in Subsubsection 2.2.2. We also report the retrieval accuracy of ANNr <ref type="bibr" target="#b24">[25]</ref>, however as the model is a retrieval approach it is not directly comparable to our model and other baselines. Instead, it indicates the performance one can reach when bypassing the issue of generation. Our model reaches comparable performance to the generation baselines in terms of 𝐿 𝑝 for all languages, and comparable performance in terms of accuracy for half of the languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Language</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Discussion</head><p>The performance on Arabic of all generation models is very low, while the retrieval model does not appear to suffer from the same effect. Further analysis of the data reveals that the character encoding used for Arabic decomposes each character into multiple encoded characters, resulting in longer and more complex sequences of characters than expected. We suppose this makes generation harder and is the cause of this low performance.</p><p>There is a significant difference in the performance of the model depending on which permutations are used. Due to the high accuracy of the decoder and the nature of the parallelogram rule, the model performs very well on analogies where the solution 𝐷 is the same as another element in the analogical equation. Permutations of this form include strong reflexivity, strong inner reflexivity, and identity.</p><p>As Table <ref type="table">2</ref> shows, the raw accuracy is often lower than the baselines Alea and Kolmo, but the Levenshtein percentage is often on par or higher. This suggests that more individual characters are correctly decoded with our model on average when compared to the baselines, but that it does not decode entire words with 100% accuracy as often. This is to be expected as the model is not trained to solve analogies, but rather is trained to properly decode words after vector arithmetic is performed on the encoded vectors.</p><p>When applied to the encoded vectors, the parallelogram rule is highly accurate with regular morphology and with certain permutations, but it often struggles when the morphology is more irregular. Since the model decodes individual words with high accuracy, the problem lies with the operations performed on the three vectors in an analogical equation after encoding. Given the model's current performance without explicitly encoding any morphological features or features of analogical equations when training, we expect that better performance can be obtained if these features are included in future iterations of the trained autoencoder.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion and Perspectives</head><p>In this paper we proposed an autoencoder framework to solving morphological analogies by generating solutions. This partially addresses the limitations of previous works relying on case based approaches that prevent creative inference and innovation. Our adaptation to the morphology setting was illustrated in several languages with promising results, and that reveal new potential directions for future work.</p><p>However, this is a preliminary proposal that will profit from further training and the combination with state of the art retrieval approaches such as ANNr from <ref type="bibr" target="#b24">[25]</ref>. Moreover, we will also explore its tranferability potential and its generalization across multiple modalities and data contexts.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Character-based word auto encoder and vector arithmetic to solve analogies</figDesc><graphic coords="5,109.99,79.19,392.01,261.01" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Autoencoder accuracy at the word level for 8 languages, trained for 100 epochs on 40,000 random words.</figDesc><table><row><cell>Language</cell><cell>Accuracy (%)</cell></row><row><cell>Arabic</cell><cell>99.99</cell></row><row><cell>English</cell><cell>99.98</cell></row><row><cell>French</cell><cell>99.99</cell></row><row><cell>German</cell><cell>99.98</cell></row><row><cell>Hungarian</cell><cell>99.97</cell></row><row><cell>Portuguese</cell><cell>99.99</cell></row><row><cell>Russian</cell><cell>99.96</cell></row><row><cell>Spanish</cell><cell>99.98</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">In<ref type="bibr" target="#b22">[23]</ref> the authors refer to morphological transformations as syntactic transformations, because they refer to the syntactic role of the word (e.g., past participle) and not the arrangement of its morphemes (e.g., the addition of the suffix "-ed").</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research work was partially supported by TAILOR, a project funded by EU Horizon 2020 research and innovation program under GA No 952215, and the Inria Project Lab "Hybrid Approaches for Interpretable AI" (HyAIAI).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Learning to rank based on analogical reasoning</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Fahandar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hüllermeier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AAAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2951" to="2958" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Analogical embedding for analogy-based learning to rank</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Fahandar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hüllermeier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IDA</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">12695</biblScope>
			<biblScope unit="page" from="76" to="88" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Analogical proportion-based methods for recommendation -first investigations</title>
		<author>
			<persName><forename type="first">N</forename><surname>Hug</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Prade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Richard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Serrurier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Fuzzy Sets Systems</title>
		<imprint>
			<biblScope unit="volume">366</biblScope>
			<biblScope unit="page" from="110" to="132" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Abstraction and analogy-making in artificial intelligence</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mitchell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ann. N.Y. Acad. Sci</title>
		<imprint>
			<biblScope unit="volume">1505</biblScope>
			<biblScope unit="page" from="79" to="101" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Analogy-preserving functions: A way to extend boolean samples</title>
		<author>
			<persName><forename type="first">M</forename><surname>Couceiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hug</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Prade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Richard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">26th IJCAI</title>
				<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1575" to="1581" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Improvements in analogical learning: Application to translating multi-terms of the medical domain</title>
		<author>
			<persName><forename type="first">P</forename><surname>Langlais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Yvon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zweigenbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">12th EACL, ACL</title>
				<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="487" to="495" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Morphological predictability of unseen words using computational analogy</title>
		<author>
			<persName><forename type="first">R</forename><surname>Fam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">24th ICCBR workshops</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="51" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Solving word analogies: A machine learning perspective</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Prade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Richard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">15th ECSQARU</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">11726</biblScope>
			<biblScope unit="page" from="238" to="250" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Solving analogies on words based on minimal complexity transformation</title>
		<author>
			<persName><forename type="first">P.-A</forename><surname>Murena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Al-Ghossein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-L</forename><surname>Dessalles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cornuéjols</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">29th IJCAI</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1848" to="1854" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Visalogy: Answering visual analogy questions</title>
		<author>
			<persName><forename type="first">F</forename><surname>Sadeghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Zitnick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Farhadi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015</date>
			<publisher>NeurIPS</publisher>
			<biblScope unit="page" from="1882" to="1890" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Detecting unseen visual relations using analogies</title>
		<author>
			<persName><forename type="first">J</forename><surname>Peyre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Laptev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sivic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE ICCV</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1981" to="1990" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Analogy and formal languages</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">6th CFG and 7th CML</title>
				<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="180" to="191" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Analogical dissimilarity: Definition, algorithms and two experiments in machine learning</title>
		<author>
			<persName><forename type="first">L</forename><surname>Miclet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bayoudh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Delhay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JAIR</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="793" to="824" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
		<title level="m">De l&apos;analogie rendant compte de la commutation en linguistique</title>
				<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
		<respStmt>
			<orgName>Université Joseph-Fourier -Grenoble I</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Habilitation à diriger des recherches</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Antic</surname></persName>
		</author>
		<title level="m">Analogical proportions</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Analogy between concepts</title>
		<author>
			<persName><forename type="first">N</forename><surname>Barbot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Miclet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Prade</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">275</biblScope>
			<biblScope unit="page" from="487" to="539" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">When Revision-Based Case Adaptation Meets Analogical Extrapolation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lieber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Prade</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">29th ICCBR</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">12877</biblScope>
			<biblScope unit="page" from="156" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A model for analogical reasoning</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Rumelhart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Abrahamson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognitive Psychology</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="1" to="28" />
			<date type="published" when="1973">1973</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Drozd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gladkova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Matsuoka</surname></persName>
		</author>
		<title level="m">Word embeddings, analogies, and machine learning: Beyond king -man + woman = queen</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="3519" to="3530" />
		</imprint>
	</monogr>
	<note>26th COLING</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">1st ICLR, Workshop Track</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Evaluating vector-space models of analogy</title>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Peterson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Griffiths</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">39th CogSci, Cognitive Science Society</title>
				<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1746" to="1751" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Vector-to-sequence models for sentence analogies</title>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ICACSIS</title>
		<imprint>
			<biblScope unit="page" from="441" to="446" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Linguistic regularities in continuous space word representations</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-T</forename><surname>Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zweig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NAACL</title>
		<imprint>
			<biblScope unit="page" from="746" to="751" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A neural approach for detecting morphological analogies</title>
		<author>
			<persName><forename type="first">S</forename><surname>Alsaidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Decker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Marquer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-A</forename><surname>Murena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Couceiro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 8th DSAA</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">A Deep Learning Approach to Solving Morphological Analogies</title>
		<author>
			<persName><forename type="first">E</forename><surname>Marquer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Alsaidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Decker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-A</forename><surname>Murena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Couceiro</surname></persName>
		</author>
		<ptr target="https://hal.inria.fr/hal-03660625" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Saussurian analogy: a theoretical account and its application</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ando</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">16th COLING</title>
				<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Tools for the production of analogical grids and a resource of n-gram analogical grids in 11 languages</title>
		<author>
			<persName><forename type="first">R</forename><surname>Fam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th LREC</title>
				<imprint>
			<publisher>ELRA</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1060" to="1066" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Glove: Global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">EMNLP</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1532" to="1543" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">Analogy and morphological change</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Fertig</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<publisher>Edinburgh University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">Analogy in Word-formation</title>
		<author>
			<persName><forename type="first">E</forename><surname>Mattiello</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>De Gruyter Mouton</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Using latent semantic to improve access to textual information</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">T</forename><surname>Dumais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">W</forename><surname>Furnas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Landauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Deerwester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Harshman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIGCHI</title>
		<imprint>
			<biblScope unit="page" from="281" to="285" />
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Dependency-based word embeddings</title>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Goldberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Short Papers), ACL</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="302" to="308" />
		</imprint>
	</monogr>
	<note>52nd ACL</note>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Character-position arithmetic for analogy questions between word forms</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lepage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">25th ICCBR workshops</title>
				<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">2028</biblScope>
			<biblScope unit="page" from="23" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Yvon</surname></persName>
		</author>
		<title level="m">Finite-state transducers solving analogies on words</title>
				<imprint>
			<publisher>Rapport GET/ENST&amp;LTCI</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<title level="m" type="main">Character-level recurrent sequence-to-sequence model</title>
		<author>
			<persName><forename type="first">F</forename><surname>Chollet</surname></persName>
		</author>
		<ptr target="https://keras.io/examples/nlp/lstm_seq2seq/" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m" type="main">Siganalogies -morphological analogies from Sigmorphon</title>
		<author>
			<persName><forename type="first">E</forename><surname>Marquer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Couceiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Alsaidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Decker</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016. 2019. 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">The SIGMORPHON 2019 shared task: Morphological analysis in context and cross-lingual transfer for inflection</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Mccarthy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Vylomova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Malaviya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wolf-Sonkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Nicolai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Kirov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Silfverberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Mielke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Heinz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cotterell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hulden</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">16th CRPPM workshops, ACL</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="229" to="244" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
