<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Empathic Response Generation in Chatbots</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Timo</forename><surname>Spring</surname></persName>
							<email>timo.spring@students.unibe.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Bern</orgName>
								<address>
									<settlement>Bern</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jacky</forename><surname>Casas</surname></persName>
							<email>jacky.casas@hes-so.ch</email>
							<affiliation key="aff1">
								<orgName type="department">HES-SO</orgName>
								<orgName type="institution">University of Applied Sciences Western Switzerland</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Karl</forename><surname>Daher</surname></persName>
							<email>karl.daher@hes-so.ch</email>
							<affiliation key="aff2">
								<orgName type="department">HES-SO</orgName>
								<orgName type="institution">University of Applied Sciences Western Switzerland</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Elena</forename><surname>Mugellini</surname></persName>
							<email>elena.mugellini@hes-so.ch</email>
							<affiliation key="aff3">
								<orgName type="department">HES-SO</orgName>
								<orgName type="institution">University of Applied Sciences Western Switzerland</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Omar</forename><surname>Abou Khaled</surname></persName>
							<email>omar.aboukhaled@hes-so.ch</email>
							<affiliation key="aff4">
								<orgName type="department">HES-SO</orgName>
								<orgName type="institution">University of Applied Sciences Western Switzerland</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Empathic Response Generation in Chatbots</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E0BE6BAED95B96C7257BECFC52654939</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T17:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Recent years show an increasing popularity of chatbots, with latest efforts aiming to make them more empathic and humanlike, finding application for example in customer service or in treating mental illnesses. Thereby, emphatic chatbots can understand the user's emotional state and respond to it on an appropriate emotional level. This survey provides an overview of existing approaches used for emotion detection and empathic response generation. These approaches raise at least one of the following profound challenges: the lack of quality training data, balancing emotion and content level information, considering the full end-to-end experience and modelling emotions throughout conversations. Furthermore, only few approaches actually cover response generation. We state that these approaches are not yet empathic in that they either mirror the user's emotional state or leave it up to the user to decide the emotion category of the response. Empathic response generation should select appropriate emotional responses more dynamically and express them accordingly, for example using emojis.</p><p>1 https://www.gartner.com/en/newsroom/ press-releases/2018-02-19-gartnersays-25-percent-of-customer-serviceoperations-will-use-virtual-customerassistants-by-2020</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Chatbots are everywhere, from booking a flight online to checking the balance of a bank account.</p><p>Thereby, most of these interactions with chatbots are still of transactional nature, for example when ordering a pizza. Furthermore, the interactions with chatbots are usually short and therefore, not resembling normal human-like conversations. Hence, recent efforts aim to also create more personalised chatbots for deeper and emotionally charged conversations. This can also help boosting the usage of chatbots by making users feel better, instead of providing or offering certain services to them. Thus, making the overall interaction more natural and human-like. Empathic chatbots find application for example in customer service or for treating mental illnesses.</p><p>Chatbots for customer service is a growing trend and Gartner 1 predicts that by 2020 about 25% of customer service requests will be handled using chatbots. <ref type="bibr" target="#b24">Xu et al. (2017)</ref> have analysed one million service requests made over Twitter. The authors note that about 40% of the requests express emotions, attitudes or opinions rather than seek for specific information. In addition, the average response time for customer service requests is about 6.5 hours. However, 72% of users who file a request, expect a response within an hour. Thus, empathic chatbots could help improving customer support by reducing the response time, reacting to specific user emotions and reduce overall costs.</p><p>Empathic chatbots also indicate potential in di-agnosing and treating mental illnesses. According to the Swiss Health Observatory, one out of five Swiss suffers from at least a slight depression<ref type="foot" target="#foot_0">2</ref> . In the United States of America, nearly one in five adults suffers from some form of mental illness causing economic costs of around $210 billion annually<ref type="foot" target="#foot_1">3</ref> . A lack of mental professionals and psychiatrists makes it difficult to treat and detect affected individuals. Empathic chatbots can provide good accessibility and are scalable to a vast public with a low entrance-barrier and to help detecting and treating mental illness faster.</p><p>There are already some noteworthy advancements in the field of empathic chatbots to treat or detect mental illnesses. <ref type="bibr">Woebot (Fitzpatrick et al., 2017)</ref> for example, is a chatbot from the University of Stanford, using methods from Cognitive Behavioural Therapy (CBT) to provide a step-bystep guidance to users with anxieties or depressions. Another noteworthy chatbot is Replika in the form of a digital companion with the main goal of providing someone to talk to 24/7 and to tackle certain resolutions for example being more social <ref type="foot" target="#foot_2">4</ref> . The underlying code of Replika is open source.</p><p>One important aspect in designing empathic chatbots is understanding what empathy actually is. In this survey, we consider empathic chatbots to use affective empathy as defined by Liu and <ref type="bibr" target="#b11">Sundar (2018)</ref>. So, the chatbots detect and understand the user's emotions and respond to them on an appropriate emotional level. <ref type="bibr" target="#b11">Liu and Sundar (2018)</ref> observe in their study that the expression of either sympathy or empathy from a health advice chatbot is favoured over an unemotional response.</p><p>With chatbots becoming more and more humanlike, it gets difficult for people to distinguish online conversations with bots and humans. This fact has lately become problematic<ref type="foot" target="#foot_3">5</ref> , since bots are increasingly being misused for political propaganda and the manipulation of people. The Computational Propaganda Research Project (COM-PROP)<ref type="foot" target="#foot_4">6</ref> from the University of Oxford devotes it's time to investigating, how chatbots and other al-gorithms are used to manipulate the public and form opinions, yielding in multiple reports on the matter. <ref type="bibr" target="#b24">Woolley and Guilbeault (2017)</ref> analysed the usage of chatbots during the 2016 presidential election in the United States using a quantitative network analysis of over 17 million tweets. The authors state that chatbots in fact showed a measurable influence during the election by either manufacturing online popularity or by democratizing propaganda. Thus, governments of several countries start to introduce regulations to fight against these kinds of online manipulations <ref type="bibr" target="#b6">(Howard et al., 2018)</ref>. However, chatbots oftentimes remain a widely-accepted tool for propaganda <ref type="bibr" target="#b24">(Woolley and Guilbeault, 2017)</ref>.</p><p>The rest of this survey is structured as follows. Section 2 introduces the different stages in the interaction with empathic chatbots. For each stage, we present the most common and noteworthy approaches. In Section 2.3, we focus on the state of empathic response generation and outline shortcomings. Finally, we conclude and discuss the survey in Section 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The Four Stages of Empathic Chatbots</head><p>We partition the interaction with an empathic chatbot in four stages -the emotion expression by the user in text format, the emotion detection and response generation by the chatbot and the response or rather emotion expression from the chatbot back to the user in text format. An overview of the stages can be seen in Figure <ref type="figure">1</ref>. Each stage requires special attention to ensure a proper endto-end user experience. The following chapter presents each stage and points out common approaches, challenges and shortcomings.</p><p>Figure <ref type="figure">1</ref>: Four stages of interaction to consider when building empathic chatbots</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Emotion Expression</head><p>Emotions are a complex construct and the ability to detect emotions in text is heavily dependent on how these emotions are expressed. Even for humans, it can be tricky to guess the emotional state of a text message. There are three major challenges when it comes to emotions. First, they are context sensitive by nature, they can be multilayered within a sentence, and they can be implicit. Hence, emotions are perceived differently based on contextual and personal circumstances, such as the culture, age, sex, education, previous experiences and other individual parameters <ref type="bibr">(Ben-Zeev, 2000;</ref><ref type="bibr" target="#b14">Oatley et al., 2006)</ref>.</p><p>In normal face-to-face conversations, emotions are also expressed over the tonality of the speaker, body language, gestures and facial expressions. However, when focusing solely on the emotion expression in text, lots of potential information stemming from these non-verbal cues go lost. This might lead to mis-interpretations of the opponent's emotions, when communicating over text messages only.</p><p>Words holding a strong emotional charge such as kisses for love, tears for sadness, or wow for surprise can help interpreting the emotional state. Such word associations are also used in emotion lexicons and word embeddings.</p><p>The usage of emojis can help amplifying or transporting emotional meaning in text-based conversations, but can also pose additional interpretation challenges, for example, when multiple contradicting emojis are used. We further discuss emojis in the context of response expression in Section 2.4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Emotion Detection</head><p>In the emotion detection stage, we try to classify and map an utterance to an emotional category. It is important to note, that emotion detection is strongly tied with response generation as similar approaches are used for both stages.</p><p>One of the first challenges is setting the number of emotion categories to be used for classification. There is no universally accepted model of emotions and the number of emotions differs drastically depending on the underlying model. One of the most popular models in emotion detection is Ekman's six basic emotions -happiness, sadness, fear, anger, disgust, and surprise <ref type="bibr" target="#b2">(Ekman, 1992)</ref>. Other popular models include Plutchik's wheel of emotions <ref type="bibr" target="#b16">(Plutchik, 1991)</ref> or Parrot's Emotional Layers <ref type="bibr" target="#b15">(Parrott, 2001)</ref> consisting of thirty-one different emotions. However, the lat-ter two models are seldomly used for emotion detection, since more emotion categories mean additional complexity for the classification task. <ref type="bibr" target="#b18">Seyeditabari et al. (2018)</ref> review existing works and approaches in the field of emotion detection and provide a good overview of the state-of-theart of emotion detection in text. They list different resources used for detecting emotions in text such as labelled text, emotion lexicons, or word embeddings and elaborate on common approaches used for emotion detection. <ref type="bibr" target="#b18">Seyeditabari et al. (2018)</ref> conclude that there is still potential for improving emotion detection in text. Thereby, the complex nature of emotion expression, the shortage of quality data and inefficient models induce most challenges for future work.</p><p>In this survey, we distinguish three major approaches used for emotion detection in text -rule-based, non-neural machine learning and deep learning.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1">Rule-Based Approaches</head><p>Rule-based approaches mainly use emotion lexicons or word embeddings. Both approaches are based on keyword lookup from text to detect the underlying emotion. The rule-based approach is only as good, as is its parsing algorithm, and the quality of the lexical resource used for the lookup. Emotion lexicons list emotion-bearing words and classify them to single or multiple emotional categories. Word embeddings, on the other hand, also take into account frequently co-occurring words that are semantically similar.</p><p>Emotion lexicons can be built from scratch. However, there exist good off-the-shelf solutions. One of the most popular being WordNet-Affect <ref type="bibr" target="#b21">(Strapparava et al., 2004)</ref>. These off-theshelf solutions differ tremendously in terms of their number of entries. WordNet-Affect contains close to five thousand words, whereas De-pecheMood (Liu and Zhang, 2012), another popular lexicon, contains more than thirty-five thousand words. Nonetheless, the quality of the lexicon is not solely dependent on its size. The vocabulary used for the lexicon also impacts its quality. <ref type="bibr">Bandhakavi et al. (2017)</ref> argue that generalpurpose lexicons such as WordNet-Affect perform not as good as domain-specific emotion lexicons. Therefore, a smaller domain-specific lexicon might yield in better results than a larger general-purpose lexicon. LIWC-based lexicons (Linguistic Inquiry and Word Count) are also widely used, since these dictionaries list grammatical, psychological, and content word categories, and thus also emotion categories with thousands of entries <ref type="bibr" target="#b1">(Chung and Pennebaker, 2012)</ref>.</p><p>The idea behind word embeddings is similar to emotion lexicons. Each word is represented as a vector in the vector space. Thereby, frequently cooccurring words are considered semantically similar and therefore, close in the vector space <ref type="bibr" target="#b18">(Seyeditabari et al., 2018)</ref>. Among the most popular word embedding methods is word2Vec <ref type="bibr" target="#b13">(Mikolov et al., 2013)</ref>. Word embeddings are also often used to train machine learning models, like LSTM, which usually take word vectors as inputs.</p><p>Both rule-based approaches are straightforward. However, there are some drawbacks to them. The emotional meaning of keywords can be ambiguous and is context-sensitive. The sentences She hates me, and I hate her, could both be classified as anger based on the keyword hate. However, when looking at the sentence level information, the first utterance could also be perceived as sad. Ignoring the syntactic structure and semantics of the whole sentence, can therefore lead to misinterpretations. Furthermore, sentences without any emotional keywords cannot be classified. Even if they might contain an implicit expression of emotions, for example in the form of a metaphor <ref type="bibr" target="#b9">(Kao et al., 2009)</ref>. As a consequence, especially emotion lexicons often lack accuracy compared to more complex approaches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2">Non-Neural Machine-Learning</head><p>Unlike rule-based approaches, non-neural learning-based approaches are trying to detect emotions using trained classifiers, such as the Support Vector Machine (SVM) <ref type="bibr" target="#b23">(Teng et al., 2006)</ref>, Naive Bayes, or Decision Trees.</p><p>We distinguish between supervised and unsupervised learning. Unsupervised approaches are an evolution of the rule-based approaches and are learning from test data that is not annotated with emotional labels. Most commonly, these approaches use movie dialogues <ref type="bibr">(Banchs, 2017;</ref><ref type="bibr" target="#b5">Honghao et al., 2017)</ref> or children's fairy tales <ref type="bibr" target="#b11">(Kim et al., 2010)</ref> to build emotional lexicons and train their models.</p><p>Supervised approaches, on the other hand, learn from labelled data such as Twitter messages. Common labels are annotations, hashtags or emojis. There exist a few good sources for emotionally labelled text, one of the most prominent be-ing the Swiss Center for Affective Sciences<ref type="foot" target="#foot_5">7</ref> providing datasets like the International Survey On Emotion Antecedents And Reactions (ISEAR) and other useful tools for emotion detection. Other well-known datasets are EmotiNet <ref type="bibr">(Balahur et al., 2011)</ref> and SemEval-2007 <ref type="bibr" target="#b20">(Strapparava and Mihalcea, 2007)</ref>.</p><p>As stated by <ref type="bibr" target="#b18">Seyeditabari et al. (2018)</ref>, one of the major challenges for supervised approaches is the lack of quality training data. Oftentimes, these datasets are unbalanced in terms of emotion categories. Banchs (2017) analyse the large movie dialogue dataset MovieDiC and conclude that emotions such as love, or joy occur much more frequently than fear or surprise. The classifiers trained on such datasets will therefore underperform for these emotional categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.3">Deep Learning Approaches</head><p>Most recent advances that showed to be effective in the field of emotion detection, have been made using deep learning <ref type="bibr" target="#b24">(Xu et al., 2017)</ref>.</p><p>Oftentimes, deep learning approaches are covering both, the emotion detection and the response generation, for example when using an Encoder-Decoder architecture <ref type="bibr" target="#b17">(Serban et al., 2015)</ref>. This architecture consists of two stages -the encoding and decoding stage. In the encoding stage, the raw text input is turned into a feature representation, usually in the form of a vector. The vector is then used as an input for the decoding stage to generate a response by applying the same strategies as in the encoding stage, but in the opposite direction.</p><p>A well-known approach applying the encodingdecoding architecture is Long Short-Term Memory (LSTM) <ref type="bibr" target="#b8">(Jithesh et al., 2017)</ref>. LSTM is a Recurrent Neural Network (RNN) that allows to capture long-term dependencies and store sequential information over a longer time. It can retain and forget the previous state and memorise extracted information from the input data depending on its importance <ref type="bibr" target="#b24">(Xu et al., 2017;</ref><ref type="bibr" target="#b22">Sun et al., 2019)</ref>.</p><p>The commonly used Sequence to Sequence (Seq2Seq) model also uses LSTMs and the Encoding-Decoding architecture <ref type="bibr">(Sutskever et al., 2014)</ref>. There is one LSTM for the encoding stage, transforming the raw text input into a fixed-length vector representation, whereas another LSTM is used for the decoding to a variable-length text out-put <ref type="bibr" target="#b0">(Cho et al., 2014;</ref><ref type="bibr" target="#b24">Xu et al., 2017;</ref><ref type="bibr" target="#b0">Chan and Lui, 2018)</ref>.</p><p>To improve the model's efficiency, Chan and Lui (2018) investigate different approaches on embedding emotional information for Seq2Seq models. Different styles, positioning, and embeddings of emotional information are tested. The authors conclude that the positioning in general matters and impacts the emotion detection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Response Generation</head><p>One of the most difficult tasks for empathic chatbots is generating an empathic response. Firstly, because it faces similar challenges as the emotion detection stage. Secondly, because it not only has to ensure that the response is appropriate in terms of content level information, but also in terms of emotion level information. This balancing act is tremendously difficult, as one usually has to sacrifice accuracy for one of the information levels, when trying to optimise the other <ref type="bibr" target="#b24">(Xu et al., 2017;</ref><ref type="bibr" target="#b25">Zhou et al., 2017)</ref>.</p><p>In terms of empathic response generation, we distinguish between two strategies -retrievalbased approaches and dynamic generation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.1">Retrieval-Based approaches</head><p>These approaches look up common responses to the user's utterance in conversation datasets. However, this method is very limited in its applicability. Similar inputs yield in the same responses, making the conversation repetitive and less natural. Furthermore, huge datasets of emotional conversations are required for such systems in order to achieve acceptable results. As such datasets are scare, these types of approaches tend to yield in responses such as I don't know in cases where no candidate response can be found.</p><p>A more advanced version of retrieval-based systems uses word embeddings on the input text to find the closest candidate responses, thus yielding in slightly more diverse responses (Bartl and Spanakis, 2017). However, it still requires lots of emotionally charged sample conversations.</p><p>Empathic response generation in general requires similar datasets as emotion detection, but with a bigger focus on conversation and dialogue turns. Movie dialogues are a good source for emotionally charged conversations. However, they oftentimes do not resemble daily conversation and seem more artificial and theatrical (Chan and Lui, 2018). Furthermore, emotions in movie dialogues are after all still acted and not naturally occurring, which might also have an impact on the quality of the training data.</p><p>Other common datasets include chat conversations, for example from Twitter service requests <ref type="bibr" target="#b24">(Xu et al., 2017)</ref> that might yield in more natural conversations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.2">Dynamic Generation</head><p>These approaches are strongly tied with the deep-learning approaches used for emotion detection from Section 2.2.3 and usually based on the encoder-decoder architecture, such as the Sequence-to-Sequence model.</p><p>The input sentence is encoded on a word-byword basis by embedding each word separately, whilst taking into account already encoded words using hidden states. Thereby, the last word embedding will produce a vector representation of the whole input sequence, encapsulating all relevant sentence level information. Semantically similar sentences are therefore close to each other in a vector space <ref type="bibr">(Sutskever et al., 2014)</ref>.</p><p>The decoder will then use the sentence vector or rather sentence embedding to produce an output sentence using inverted encoding mechanisms on a word-by-word basis. This allows encodingdecoding architectures to generate variable length responses. Thereby, it will consider already decoded words to ensure that the generated response is also grammatically correct. To find appropriate responses to a given word from the encoding stage, vocabularies or word-embeddings are being used.</p><p>However, the longer the input sequence, the more challenging to capture the full meaning in a single sentence embedding. For an input sentence of 30 words, the decoder would have to consider, what was encoded 30 steps ago, just to decode the first word. This long-range dependency problem oftentimes results in poor responses, like I don't know <ref type="bibr" target="#b0">(Chan and Lui, 2018)</ref>.</p><p>Attention mechanisms are commonly used to tackle the issue of long-range dependencies <ref type="bibr">(Sutskever et al., 2014)</ref>. Using attention, the decoder has direct access to the hidden state of each encoded word and can weight each word correspondingly. This allows the decoder to attend and weight on relevant parts of an input sentence, when generating the output. This mechanism is also applied in Neural Machine Translation <ref type="bibr" target="#b0">(Bahdanau et al., 2014)</ref>.</p><p>However, these Recurrent Neural Networks (RNN) still suffer from the vanishing gradient problem, that causes issues with long-range dependencies <ref type="bibr" target="#b5">(Hochreiter, 1998)</ref>. LSTMs also apply attention and in addition allow to retain and forget information, therefore handling the long-range dependency problem better than other approaches.</p><p>All these mechanisms are essentially required to ensure an appropriate response in terms of content level information. When we also want to consider emotion level information, we add additional complexity to the model. Emotions either have to be additionally encoded during the encoding stage or fed directly to the decoding stage to generate emotional responses <ref type="bibr" target="#b25">(Zhou et al., 2017)</ref>.</p><p>The Emotional Chatting Machine as proposed by <ref type="bibr">Zhou et al. (</ref> <ref type="formula">2017</ref>) is a recent and noteworthy approach for assessing the emotional state of conversations and to generate appropriate emotional responses. Therefore, it belongs to the dynamic generation approaches. The ECM deep learning algorithm is trained with 22.300 Chinese blog posts that are manually annotated with Ekman's six basic emotions.</p><p>In terms of architecture, the ECM is based on the Seq2Seq model with an encoding and decoding phase. In addition, <ref type="bibr" target="#b25">Zhou et al. (2017)</ref> introduce an internal and external memory to the model to capture changes in the emotion state throughout the sentence and to map explicit emotion expressions to emotion categories. Figure <ref type="figure">2</ref> provides a good overview of the ECM architecture. As input, the ECM requires the user's text message and one of the Ekman's six emotion categories to condition the response. Based on the input message, the ECM will generate an appropriate response and condition it using the input emotion category. <ref type="bibr" target="#b25">Zhou et al. (2017)</ref> benchmark the ECM with other approaches, such as the traditional Seq2Seq model or lexicon-based approaches and show that ECM performs best across all emotion categories. However, it lacks behind slightly on the content level that the authors put down to an imbalance in the training set.</p><p>One drawback of the ECM is that the input emotion has to be set manually. This hardwiring of the output emotion can be useful, if the chatbot should always respond in the same emotional state, or express certain personality traits such as being angry all the time. However, if we want the chatbot to dynamically react to the user's emotions and make the interaction natural, then we have to change the emotion category based on the emotional state of the user's message automatically.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.3">Empathic Responses</head><p>As discussed in Section 1, empathy requires understanding of the user's emotion and replying to them on an appropriate emotional level. Using emotion detection, we can achieve good results in understanding the user's emotions. The difficult part is actually selecting the appropriate emotion to condition the response with.</p><p>One approach could be to simply mirror the user's emotion.</p><p>However, in humanconversations, empathy finds expression, when one tries to feel with the opponent and not necessarily similar to the opponent. One's own emotion must not be confused with the opponent's emotion. When resonating to the opponent's emotion, one is still aware that it might be different from the personal emotion <ref type="bibr" target="#b19">(Singer and Klimecki, 2014)</ref>. If someone is sad, you might understand this sadness and try to cheer them up, instead of responding in a sad way as well. However, it does not mean that you are necessarily feeling sad as well. Thus, simply mirroring the user's emotion does not necessarily yield in empathic responses. This is also an important aspect with regards to a chatbot's personality, since in these cases, one should think about the chatbot's own emotion, as well as how it might resonate on someone else's emotions.</p><p>Another crucial aspect is taking into account the user's emotional evolution throughout the whole conversation. When just considering the latest user utterance for emotion detection, misinterpretations or frequent switches in the emotions expressed by the chatbot's response might occur. For example, the user could genuinely be in a bad mood, but laugh at a joke one just made. If we would consider just the latest user utterance to detect the user's emotion and condition the response accordingly, the chatbot's expressed emotion would switch from negative to positive within a single sentence. Modelling the user's emotion over a longer period might also be important when applying empathic chatbots in treating mental illnesses, or when building personality profiles to monitor the emotional state of the user.</p><p>We observe that only little research actually focusses on the generation of empathic responses in Computer Science, compared to the efforts Figure <ref type="figure">2</ref>: Overview of the ECM architecture based on the encoder-decoder framework with the addition of an internal and external memory to further improve the emotional response <ref type="bibr" target="#b25">(Zhou et al., 2017)</ref>. done for emotion detection. There exist some approaches, such as the Encoder-Decoder architecture, that cover emotion detection as well as emotional response generation. Nonetheless, an emotional response is not necessarily an empathic response as elaborated before.</p><p>How humans are generating empathic responses is still an ongoing field of research in neuroscience (Shamay-Tsoory and Lamm, 2018). Similar to emotions, there is no universally accepted model for empathic responses, except that empathy is heavily context-dependent <ref type="bibr" target="#b19">(Singer and Klimecki, 2014)</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Response Expression</head><p>In a normal conversation, non-verbal cues such as facial expressions or gestures can help indicate a person's emotional state. However, with chatbots, we are missing such information and have to focus solely on the user's text, to detect the emotional state. Similar constraints also apply to the response expression by the chatbot. It is difficult to transport the intended emotion from the generated response back to the user in a text format. Some approaches try to simulate non-verbal cues by displaying the chatbot as a 3D simulation of a person <ref type="bibr" target="#b23">(Tatai et al., 2003)</ref>. We note that in general, chatbots do not express responses in any other way than text. Because such non-verbal cues are missing in traditional electronic messaging systems, people are using emojis to supply such cues. The usage of emojis has increased heavily over the previous years. In 2017, Facebook revealed that on an average day, over 5 Billion emojis are being sent over Messenger only<ref type="foot" target="#foot_6">8</ref> . <ref type="bibr" target="#b7">Hu et al. (2017)</ref> state that the main reason for the usage of emojis in messages is to express emotions or strengthen expressions.</p><p>Emojis could therefore also be considered when detecting the user's emotion. However, two challenges arise when using emojis as possible emotional labels. First, the emoji label could be contradictory to the perceived emotional state from the text, for example implying a sarcastic utterance. Figure <ref type="figure" target="#fig_0">3</ref> shows, how emojis can lead to such contradicting interpretations. Second, emojis are prone to cultural differences as stated by <ref type="bibr" target="#b12">Ljubešić and Fišer (2016)</ref>. Chatbots with a global scope should therefore take into account, that emojis might be used and perceived differently depending on the country. DeepMoji <ref type="bibr" target="#b3">(Felbo et al., 2017</ref>) is an impressive tool translating text into a set of emojis expressing a similar emotional state returning the five most likely emojis together with their probabilities. It is trained on 1.2 billion tweets containing emojis and uses LSTM to predict the most appropriate emojis. Generated emojis could be mapped to different emotion categories and used to express emotions in the response to the user. We leave the validation of this method for future work.</p><p>We state that the usage of emojis in conversational agents might be a clue to make them more human-like and to help expressing non-verbal cues that otherwise might go missing. Future work should therefore focus on validating this hypothesis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Discussion and Conclusion</head><p>We note that current state-of-the-art approaches face the following major challenges:</p><p>1. Shortage of quality training data -Machine-Learning algorithms for emotion detection and empathic response generation require an extensive amount of annotated training data. Existing datasets are scarce and are oftentimes unbalanced for different emotions. Hence, chatbots that were trained using such datasets will lead to poor performances on these emotions. Using annotated data from social media has proven to yield in good results, but also suffers from unbalanced emotion distribution. To generate human-like responses, natural conversations should be used for training as opposed to artificial and theatrical movie dialogues.</p><p>2. Emotion level and content level -To generate responses that are grammatically correct and that reflect the appropriate content level is very complex. Using domain specific training data can help improve the accuracy. If the answer should also reflect the emotional level and detect possibly implicit or multi-layered emotions, then the complexity increases even further. Improving one of the levels -emotion or content -without sacrificing accuracy for the other is very challenging. For response generation, we note that existing approaches are mainly focusing on content level information and consider emotions only as additional information during encoding.</p><p>3. Considering the full end-to-end experience -In order to achieve good results, one has to consider the impacts of all four stages -emotion expression by the user, emotion detection by the chatbot, response generation by the chatbot, and appropriate response expression back to the user.</p><p>Only by considering the full end-to-end experience can chatbots be improved to be more humanlike and empathic. Future work should investigate the integration of emojis into the full end-to-end experiencefrom emotion detection to response expression.</p><p>4. Modelling emotions throughout conversations -We state that when selecting emotions to condition the response, one should not only consider the detected emotion from the latest user message. Taking into account the evolution of the user's emotion throughout the whole conversation and possibly even over several previous conversations, prevents frequent changes of the chatbot's expressed emotions and helps model the user's long-term emotional state.</p><p>Furthermore, more efforts should be devoted to understanding empathy and how chatbots can generate empathic responses instead of just emotional responses.</p><p>In this survey, we have presented the state-ofthe-art of empathy and especially empathic response generation in chatbots and pointed out several noteworthy approaches. We pointed out the four stages of the interaction with the chatbot and underlined the importance to take all the stages into account when creating empathic chatbots.</p><p>We note that there exist many different approaches to tackle the problem of emotion detection, but only few for empathic response generation. Overall, deep learning algorithms, such as the Emotional Chatting Machine (ECM) tend to yield in the best results. Even though, there is still potential for improvement as the ECM only generates emotional responses, but not empathic ones. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>References</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Two examples of challenging cases where the emojis are contradictory to the perceived emotional state of the text message, or the multiple contradicting emojis are used.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 https://arxiv.org/abs/1409.0473. Alexandra Balahur, Jesús M. Hermida, Andrés Montoyo, and Rafael Muñoz. 2011. EmotiNet: A knowledge base for emotion detection in text built on the appraisal theories. In Natural Language Processing and Information Systems, Springer Berlin Heidelberg, pages 27-39. https://doi.org/10.1007/978-3-642-22327-3 4 . Rafael E. Banchs. 2017. On the construction of more human-like chatbots: Affect and emotion analysis of movie dialogue data. In 2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC). IEEE. https://doi.org/10.1109/apsipa.2017.8282245.</figDesc><table><row><cell>Anil Bandhakavi, Nirmalie Wiratunga, Stewart Massie, and Deepak Padmanabhan. 2017. Lex-icon generation for emotion detection from text. IEEE Intelligent Systems 32(1):102-108. https://doi.org/10.1109/mis.2017.22.</cell></row><row><cell>A. Bartl and G. Spanakis. 2017. based dialogue system utilizing utterance and A retrieval-context embeddings. In 2017 16th IEEE International Conference on Machine Learning</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://www.obsan.admin.ch/de/ publikationen/psychische-gesundheit</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://www.nimh.nih.gov/health/ statistics/mental-illness.shtml</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://replika.ai</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">https://www.nytimes.com/2018/12/04/ opinion/chatbots-ai-democracy-freespeech.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">6 https://comprop.oii.ox.ac.uk</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">https://www.unige.ch/cisa/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6">https://www.adweek.com/digital/ facebook-world-emoji-day-stats-theemoji-movie-stickers/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Learning phrase representations using RNN encoder-decoder for statistical machine translation</title>
		<author>
			<persName><forename type="first">;</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Hei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrew</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fai</forename><surname>Kwok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Lui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bart</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Caglar</forename><surname>Van Merrienboer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dzmitry</forename><surname>Gulcehre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fethi</forename><surname>Bahdanau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Holger</forename><surname>Bougares</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yoshua</forename><surname>Schwenk</surname></persName>
		</author>
		<author>
			<persName><surname>Bengio</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICMLA.2017.00011.AaronBen-Zeev</idno>
		<ptr target="https://doi.org/10.3115/v1/d14-1179" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics</title>
				<meeting>the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics</meeting>
		<imprint>
			<publisher>The MIT Press</publisher>
			<date type="published" when="2000">2000. 2018. 2018. 2014</date>
			<biblScope unit="page" from="1120" to="1125" />
		</imprint>
	</monogr>
	<note>International Conference on Artificial Intelligence and Big Data (ICAIBD)</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Linguistic inquiry and word count (LIWC)</title>
		<author>
			<persName><forename type="first">Cindy</forename><forename type="middle">K</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">James</forename><forename type="middle">W</forename><surname>Pennebaker</surname></persName>
		</author>
		<idno type="DOI">10.4018/978-1-60960-741-8.ch012</idno>
		<ptr target="https://doi.org/10.4018/978-1-60960-741-8.ch012" />
	</analytic>
	<monogr>
		<title level="m">Applied Natural Language Processing</title>
				<imprint>
			<publisher>IGI Global</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="206" to="229" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">An argument for basic emotions</title>
		<author>
			<persName><forename type="first">Paul</forename><surname>Ekman</surname></persName>
		</author>
		<idno type="DOI">10.1080/02699939208411068</idno>
		<ptr target="https://doi.org/10.1080/02699939208411068" />
	</analytic>
	<monogr>
		<title level="j">Cognition and Emotion</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">3-4</biblScope>
			<biblScope unit="page" from="169" to="200" />
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm</title>
		<author>
			<persName><forename type="first">Bjarke</forename><surname>Felbo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alan</forename><surname>Mislove</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anders</forename><surname>Søgaard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Iyad</forename><surname>Rahwan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sune</forename><surname>Lehmann</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/d17-1169</idno>
		<ptr target="https://doi.org/10.18653/v1/d17-1169" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics</title>
				<meeting>the 2017 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (woebot): A randomized controlled trial</title>
		<author>
			<persName><forename type="first">Kathleen</forename><surname>Kara Fitzpatrick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alison</forename><surname>Darcy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Molly</forename><surname>Vierhile</surname></persName>
		</author>
		<idno type="DOI">10.2196/mental.7785</idno>
		<ptr target="https://doi.org/10.2196/mental.7785" />
	</analytic>
	<monogr>
		<title level="j">JMIR Mental Health</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">e19</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The vanishing gradient problem during learning recurrent neural nets and problem solutions</title>
		<author>
			<persName><forename type="first">Sepp</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Wei ; Honghao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yiwei</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Junjie</forename><surname>Ke</surname></persName>
		</author>
		<idno type="DOI">10.1142/S0218488598000094</idno>
		<ptr target="http://web.stanford.edu/class/cs224s/reports" />
	</analytic>
	<monogr>
		<title level="m">Building chatbot with emotions</title>
				<imprint>
			<date type="published" when="1998">1998. 2017</date>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="107" to="116" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">Philip</forename><forename type="middle">N</forename><surname>Howard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bence</forename><surname>Kollanyi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Samantha</forename><surname>Bradshaw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lisa-Maria</forename><surname>Neudert</surname></persName>
		</author>
		<idno>CoRR abs/1802.03573</idno>
		<ptr target="http://arxiv.org/abs/1802.03573" />
		<title level="m">Social media, news and political information during the US election: Was polarizing content concentrated in swing states</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Spice up your chat: The intentions and sentiment effects of using emoji</title>
		<author>
			<persName><forename type="first">Tianran</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Han</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hao</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thuy-Vy Thi</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiebo</forename><surname>Luo</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1703.02860</idno>
		<ptr target="https://arxiv.org/abs/1703.02860" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">LSTM recurrent neural networks for high resolution range profile based radar target classification</title>
		<author>
			<persName><forename type="first">Justin</forename><surname>Jithesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sagayaraj</surname></persName>
		</author>
		<author>
			<persName><surname>Srinivasa</surname></persName>
		</author>
		<idno type="DOI">10.1109/ciact.2017.7977298</idno>
		<ptr target="https://doi.org/10.1109/ciact.2017.7977298" />
	</analytic>
	<monogr>
		<title level="m">International Conference on Computational Intelligence &amp; Communication Technology (CICT)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Towards text-based emotion detection a survey and possible improvements</title>
		<author>
			<persName><forename type="first">Edward</forename><surname>Chao-Chun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chun-Chieh</forename><surname>Kao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ting-Hao</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chang-Tai</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Von-Wun</forename><surname>Hsieh</surname></persName>
		</author>
		<author>
			<persName><surname>Soo</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<idno type="DOI">10.1109/icime.2009.113</idno>
		<ptr target="https://doi.org/10.1109/icime.2009.113" />
		<title level="m">International Conference on Information Management and Engineering</title>
				<imprint>
			<publisher>IEEE</publisher>
			<biblScope unit="page" from="70" to="74" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Should machines express sympathy and empathy? experiments with a health advice chatbot</title>
		<author>
			<persName><forename type="first">Mac</forename><surname>Sunghwan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rafael</forename><forename type="middle">A</forename><surname>Valitutti</surname></persName>
		</author>
		<author>
			<persName><surname>Calvo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Bingjie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><forename type="middle">S Shyam</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><surname>Sundar</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-1-4614-3223-413</idno>
		<ptr target="https://doi.org/10.1089/cyber.2018.0110" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the NAACL HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text</title>
				<editor>
			<persName><forename type="first">Lei</forename><surname>Zhang</surname></persName>
		</editor>
		<meeting>the NAACL HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text<address><addrLine>Stroudsburg, PA, USA, CAAGET &apos;10</addrLine></address></meeting>
		<imprint>
			<publisher>Springer US</publisher>
			<date type="published" when="2010">2010. 2012. 2018</date>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="625" to="636" />
		</imprint>
	</monogr>
	<note>Evaluation of unsupervised emotion models to textual affect recognition</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A global analysis of emoji usage</title>
		<author>
			<persName><forename type="first">Nikola</forename><surname>Ljubešić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Darja</forename><surname>Fišer</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/w16-2610</idno>
		<ptr target="https://doi.org/10.18653/v1/w16-2610" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th Web as Corpus Workshop. Association for Computational Linguistics</title>
				<meeting>the 10th Web as Corpus Workshop. Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kai</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Greg</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeffrey</forename><surname>Dean</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1301.3781</idno>
		<ptr target="https://arxiv.org/abs/1301.3781" />
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Understanding emotions</title>
		<author>
			<persName><forename type="first">Keith</forename><surname>Oatley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dacher</forename><surname>Keltner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jennifer</forename><forename type="middle">M</forename><surname>Jenkins</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
	<note>Blackwell publishing</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">Gerrod</forename><surname>Parrott</surname></persName>
		</author>
		<title level="m">Emotions in social psychology: Essential readings</title>
				<imprint>
			<publisher>Psychology Press</publisher>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">The emotions</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Plutchik</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1991">1991</date>
			<publisher>University Press of America</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Hierarchical neural network generative models for movie dialogues</title>
		<author>
			<persName><forename type="first">Iulian</forename><surname>Vlad Serban</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Sordoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yoshua</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aaron</forename><forename type="middle">C</forename><surname>Courville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joelle</forename><surname>Pineau</surname></persName>
		</author>
		<idno>CoRR abs/1507.04808</idno>
		<ptr target="http://arxiv.org/abs/1507.04808" />
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">The neuroscience of empathy -from past to present and future</title>
		<author>
			<persName><forename type="first">Armin</forename><surname>Seyeditabari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Narges</forename><surname>Tabari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wlodek</forename><surname>Zadrozny</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Simone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Claus</forename><surname>Lamm</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.neuropsychologia.2018.04.034</idno>
		<idno type="arXiv">arXiv:1806.00674</idno>
		<ptr target="https://doi.org/10.1016/j.neuropsychologia.2018.04.034" />
	</analytic>
	<monogr>
		<title level="m">The Neuroscience of Empathy</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="volume">116</biblScope>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Emotion detection in text: a review</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Empathy and compassion</title>
		<author>
			<persName><forename type="first">Tania</forename><surname>Singer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Olga</forename><forename type="middle">M</forename><surname>Klimecki</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.cub.2014.06.054</idno>
		<ptr target="https://doi.org/10.1016/j.cub.2014.06.054" />
	</analytic>
	<monogr>
		<title level="j">Current Biology</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">18</biblScope>
			<biblScope unit="page" from="R875" to="R878" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">SemEval-2007 task 14</title>
		<author>
			<persName><forename type="first">Carlo</forename><surname>Strapparava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rada</forename><surname>Mihalcea</surname></persName>
		</author>
		<idno type="DOI">10.3115/1621474.1621487</idno>
		<ptr target="https://doi.org/10.3115/1621474.1621487" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th International Workshop on Semantic Evaluations -SemEval 07. Association for Computational Linguistics</title>
				<meeting>the 4th International Workshop on Semantic Evaluations -SemEval 07. Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Wordnet affect: an affective extension of wordnet</title>
		<author>
			<persName><forename type="first">Carlo</forename><surname>Strapparava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Valitutti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Lrec. Citeseer</title>
				<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="1083" to="1086" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Dynamic emotion modelling and anomaly detection in conversation based on emotional transition tensor</title>
		<author>
			<persName><forename type="first">Xiao</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chen</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lian</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">; Oriol</forename><surname>Vinyals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Quoc</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.inffus.2018.04.001</idno>
		<idno>CoRR abs/1409.3215</idno>
		<ptr target="http://arxiv.org/abs/1409.3215" />
	</analytic>
	<monogr>
		<title level="m">Sequence to sequence learning with neural networks</title>
				<imprint>
			<date type="published" when="2014">2019. 2014</date>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="11" to="22" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Retracted: Recognition of emotion with SVMs</title>
		<author>
			<persName><forename type="first">Gábor</forename><surname>Tatai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Annamária</forename><surname>Csordás</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Árpád</forename><surname>Kiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Attila</forename><surname>Szaló</surname></persName>
		</author>
		<author>
			<persName><forename type="first">László</forename><surname>Laufer ; Zhi ; Teng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fuji</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shingo</forename><surname>Kuroiwa</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-540-39396-2</idno>
		<ptr target="https://doi.org/10.1007/978-3-540-37275-287" />
	</analytic>
	<monogr>
		<title level="m">Intelligent Virtual Agents</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2003">2003. 2006</date>
			<biblScope unit="page" from="701" to="710" />
		</imprint>
	</monogr>
	<note>Happy chatbot, happy user</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Computational propaganda in the united states of america: Manufacturing consensus online</title>
		<author>
			<persName><forename type="first">C</forename><surname>Samuel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Douglas</forename><forename type="middle">R</forename><surname>Woolley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Guilbeault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhe</forename><surname>Usa ; Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yufan</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vibha</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rama</forename><surname>Sinha</surname></persName>
		</author>
		<author>
			<persName><surname>Akkiraju</surname></persName>
		</author>
		<idno type="DOI">10.1145/3025453.3025496</idno>
		<ptr target="https://doi.org/10.1145/3025453.3025496" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems -CHI 17</title>
				<meeting>the 2017 CHI Conference on Human Factors in Computing Systems -CHI 17</meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
	<note>Computational Propaganda Research Project</note>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Emotional chatting machine: Emotional conversation generation with internal and external memory</title>
		<author>
			<persName><forename type="first">Hao</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Minlie</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tianyang</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaoyan</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bing</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1704.01074</idno>
		<ptr target="https://arxiv.org/abs/1704.01074" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
