<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Detecting and tracking ongoing topics in psychotherapeutic conversations</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Ilyas</forename><surname>Chaoua</surname></persName>
							<email>ilyaschaoua@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">Mathematics and Computer Science Department</orgName>
								<orgName type="institution">University of Cagliari</orgName>
								<address>
									<addrLine>Via Ospedale 72</addrLine>
									<postCode>09124</postCode>
									<settlement>Cagliari</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Philips Research</orgName>
								<address>
									<addrLine>High Tech Campus 34</addrLine>
									<postCode>5656 AE</postCode>
									<settlement>Eindhoven</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Diego</forename><forename type="middle">Reforgiato</forename><surname>Recupero</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Mathematics and Computer Science Department</orgName>
								<orgName type="institution">University of Cagliari</orgName>
								<address>
									<addrLine>Via Ospedale 72</addrLine>
									<postCode>09124</postCode>
									<settlement>Cagliari</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sergio</forename><surname>Consoli</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Philips Research</orgName>
								<address>
									<addrLine>High Tech Campus 34</addrLine>
									<postCode>5656 AE</postCode>
									<settlement>Eindhoven</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Aki</forename><surname>Härmä</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Philips Research</orgName>
								<address>
									<addrLine>High Tech Campus 34</addrLine>
									<postCode>5656 AE</postCode>
									<settlement>Eindhoven</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Rim</forename><surname>Helaoui</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Philips Research</orgName>
								<address>
									<addrLine>High Tech Campus 34</addrLine>
									<postCode>5656 AE</postCode>
									<settlement>Eindhoven</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Detecting and tracking ongoing topics in psychotherapeutic conversations</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">A7F91EFBEE39F27801C712B2B8229B3B</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T22:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Conversational AI</term>
					<term>Psychotherapeutic conversations</term>
					<term>Topics detection and modeling</term>
					<term>Partially Labeled Dirichlet Allocation</term>
					<term>Transitions matrices</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>One of the key aspects in a psychotherapeutic conversation is the understanding of topics dynamics driving the dialogue. This may provide insights on the therapeutic strategy adopted by the counselor for the specific patient, providing the opportunity of building up artificial intelligence (AI) based methods for recommending the most appropriate therapy for a new patient. In this paper, we propose a method able to detect and track topics in real-life psychotherapeutic conversations based on Partially Labeled Dirichlet Allocation. Topics detection helps in summarizing the semantic themes used during the therapeutic conversations, and in predicting a specific topic for each talk-turn. The conversation is modeled by means of a distribution of ongoing topics propagating through each talk sequence. Tracking of topics aims at exploring the dynamics of the conversation and at offering insights into the underlying conversation logic and strategy. We present an alternative way to look at face-to-face conversations in conjunction with a new approach that combines topic modeling and transitions matrices to elicit valuable knowledge.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The theoretical and technological advances in several disciplines of linguistics, computer science, and healthcare have made possible the recent investigation of therapeutic conversation analysis as a growing field of research <ref type="bibr" target="#b11">[12]</ref>. Computational learning techniques have been leveraged to extract useful information from humans interactions through the identification and exploration of unusual patterns <ref type="bibr" target="#b4">[5]</ref>.</p><p>Therapeutic conversations methods such as Cognitive Behavior Therapy (CBT) refer to a range of therapies that can help treating mental health problems, emotional challenges, and some psychiatric disorders of patients by changing their ways of thinking and behave. Accordingly, these therapeutic methods create a new way of looking at severe psychological issues to help patients to move towards a solution and to gain a better understanding of themselves. The treatment is usually a face-to-face conversation where the counselor interacts directly with the patient to understand his feelings, e.g. confident, anxious, or depressed, as well as the causes of his feelings. During the conversation, counselor and patient create a sequence of spoken sentences, each characterized by a certain topic, creating in this way a thematic structure to the whole therapeutic conversation. By adopting a set of techniques and conversational strategies coming from clinical practice, the counselor aims at solving behavioral and psychological problems of the patient by properly reacting to the patient and redirecting the conversation towards certain thematics.</p><p>As a result, investigating and modeling the human-to-human dialogues in this kind of context may serve as a guide for the development of AI-based dialoguing systems able, e.g., at recommending the most appropriate therapeutic strategy to adopt by the practitioner for a new patient <ref type="bibr" target="#b4">[5]</ref>. In this context, topic detection and tracking (TDT) has been the point of intensive studies since the beginning of natural language processing (NLP) <ref type="bibr" target="#b13">[14]</ref> and artificial intelligence (AI) research. One aim of TDT is to identify the appearance of new topics and following their reappearance and evaluation <ref type="bibr" target="#b10">[11]</ref>.</p><p>In this paper we investigate patient-therapist interactions by modeling the propagation of the topics addressed during a given therapeutic conversation by using Partially Labeled Latent Dirichlet Allocation (PLDA) <ref type="bibr" target="#b18">[19]</ref>. Traditional Latent Dirichlet Allocation (LDA) is one of the most popular topic model in the literature and is based on a bag-of-words approach. Since data we will deal are partially labeled, then PLDA is of interest because it is likely to produce better results than classic LDA. PLDA is often very useful for applications with a human in the loop, since induced classes correspond well with human categorization, as represented by the provided label space.</p><p>The study has been conducted over a dataset of 1729 real-life transcribed psychotherapeutic conversations, each made of different talk-turns, which we will describe further in Section 4. In our method we first identify the most common topics used within the psychology corpus. The PLDA model takes as input the given conversations and detects significant words for each topic. Secondly, the trained PLDA model is able to determine the potential topic addressed in each talk-turn. The talk-turns flow is then transformed into a sequence of potential topics within each conversation. Finally, the semi-supervised PLDA topic model is evaluated by computing its coherence over the most significant words for each topic. Our final aim is to find the quintessential patterns in therapeutic conversations and to understand the topic switches according to the adopted dialogue strategy and topics propagation dynamics. In our method we distinguish the topic changes driven by the counselor and the ones prompted by the patient, and two topic transition matrices are constructed accordingly. These matrices are able to characterize the conversation and to provide important hints towards the understanding of the topics propagation dynamics.</p><p>The remainder of the paper is organized as follows. In Section 2 we discuss the related works in the literature on automatic topic detection methods and therapeutic dialogue analysis, that help to establish the basis for the present work. In Section 3 we describe how topic modeling algorithms work, specifying their main characteristics. Afterwards we describe the data used for the experiments and the followed preprocessing steps in Section 4. Section 5 shows our approach on TDT by illustrating the developed methodology. Lastly, Section 6 discusses how to evaluate our model, while Section 7 ends the paper with conclusions and directions for future research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Current trends in therapeutic conversations research focus on the digitalization of spoken interactions and on the recommendation of the most appropriate treatments, employing computational techniques such as NLP which offer the potential to extract knowledge from consultation transcripts. Authors in <ref type="bibr" target="#b2">[3]</ref> combined robust communication theory used in healthcare and a visualization text analytic technique called Discursis, to analyze the conversational behavior in consultations. Discursis<ref type="foot" target="#foot_0">3</ref> is a visual text analytic tool for analyzing human communications, that automatically builds an inherent language model from a given transcribed conversation, mines its conceptual content for each talk-turns, and creates a visual brief. The resulting report can identify communication patterns present during the discussion, with appropriate results of engagement between interlocutors to understand the conversation dynamics. In medical consultations, the classification of conversations suffers from critical weaknesses, including intense performance requirements, time-consuming, and non-standardized annotating systems. To overcome these shortcomings, authors in <ref type="bibr" target="#b12">[13]</ref> built an automated annotating system employing a Labeled LDA <ref type="bibr" target="#b17">[18]</ref> model to learn the relationships between a transcribed conversation and its associated annotations. Those annotations refer to the subjects and patient symptoms discussed during the therapeutic conversations. The resulting system automatically identifies and restricts those annotations in separate talkturns within a given conversation. Differently, contributors in <ref type="bibr" target="#b14">[15]</ref> examined the use of a LDA <ref type="bibr" target="#b6">[7]</ref> topic model as an automatic annotator tool to explore topics and to predict the therapy outcomes of the conversation. The authors assumed that the automated detection of topics does not aim at predicting the symptoms, but it can be used to predict some essential factors such as patient satisfaction and ratings of therapy quality. The examinations from both approaches show that identification and tracking of topics can give useful information for clinicians, enabling them to assist better the identification of patients who may be at risk of loss of treatment. Analyzing human communications, the authors in <ref type="bibr" target="#b1">[2]</ref> converted transcribed conversations to time series by developing a discourse visualization system, a text analysis model and a set of quantitative metrics to identify significant features. The system was able to understand the topics used by specific participants, and to generate reports within a single conversation. The method can be used to observe the structure and patterns of interactions and to reconstruct the dynamics of the communication, including the consistency levels of topics discussed between participants and the timing of topic changes. Contributors in <ref type="bibr" target="#b21">[22]</ref> proposed a conceptual dynamic latent Dirichlet allocation (CDLDA) model for TDT in conversational text content. Differently to the traditional LDA model, which detects topics only through a bag-of-words technique, CDLDA considers essential information including speech acts, semantic concepts, and hypernym definitions in E-HowNet<ref type="foot" target="#foot_1">4</ref>  <ref type="bibr" target="#b9">[10]</ref>. It basically extracts the dependencies between speech acts and topics, where hypernym infor-mation makes the topic structure more complete and extends the abundance of original words. Experimental results revealed that the proposed approach outperforms the conventional Dynamic Topic Models <ref type="bibr" target="#b5">[6]</ref>, LDA, and support vector machine models, to achieve excellent performance for TDT in conversations. Authors in <ref type="bibr" target="#b0">[1]</ref> presented On-toLDA for the task of topic labeling utilizing an ontology-based topic model, along with a graph-based topic labeling method (i.e., the topic labeling method based on the ontological meaning of the concepts included in the discovered topics). The model was able to show each topics as a multinomial distribution of concepts, and each concept as a distribution over words. Comparing to the classical LDA, the model scaled up better the topic coherence score by combining ontological concepts with probabilistic topic models towards a combined framework applied to different kind of text collections. Contributors in <ref type="bibr" target="#b7">[8]</ref> showed an approach to improve human-agent dialogs using automatic identification and tracking of dialogue topics, exploiting the basis of contextual knowledge provided by Wikipedia category system. This approach was constructed by mapping the different utterances to Wikipedia articles and by defining their relevant Wikipedia categories as a likely of topics. As a result, the detection method was able to recognize a topic without holding a priori knowledge of its belonging subject category.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Topic Modeling</head><p>Topic models are a family of probabilistic approaches that aim at discovering latent semantic structures in large documents. Based on the presumption that meanings are relational, they interpret topics or themes within a set of documents originally constructed from a probability distribution over words. As a result, a document is viewed as a combination of topics, while a topic is viewed as a blend of words.</p><p>One of the most widely used statistical language modeling for this end is Latent Dirichlet Allocation (LDA) introduced by Blei et al. <ref type="bibr" target="#b6">[7]</ref>. LDA is a generative approach. It assumes that documents in a given corpus are generated by repeatedly picking up a topic, then a word from that topic according to the distribution of all observed words in the corpus given that topic. LDA aims at learning these distributions and inferring the (hidden) topics given the (observed) words of the documents <ref type="bibr" target="#b15">[16]</ref>. Given the nature of our data which includes partial annotations, we employ the following two variants of LDA.</p><p>Labeled Latent Dirichlet Allocation (LLDA) <ref type="bibr" target="#b17">[18]</ref> is a supervised version of LDA that constraints it by defining a one-to-one correspondence between topics and humanprovided labels. This allows Labeled LDA to learn word-label correspondences. Figure <ref type="figure" target="#fig_0">1</ref> illustrates the probabilistic graphical model of LLDA.</p><p>Partially Labeled Latent Dirichlet Allocation(PLDA) <ref type="bibr" target="#b18">[19]</ref> is a semi-supervised version of LDA which extends it with constraints that align some learned topics with a humanprovided label. The model exploits the unsupervised learning of topic models to explore the unseen themes with each label, as well as unlabeled themes in the large collection of data. As illustrated in Figure <ref type="figure">2</ref>, PLDA assumes that the document's words are drawn from a document-specific mixture of latent topics, where each topic is represented as Fig. <ref type="figure">2</ref>. Probabilistic graphical model for PLDA: each document's word ω and label Λ are observed, with the per-doc label distribution ψ, per-doc-label topic distributions θ, and per-topic word distributions Φ hidden variables. Because each document's label-set λ d is observed, its sparse vector prior γ is unused; included for completeness. <ref type="bibr" target="#b18">[19]</ref> a distribution over words, and each document can use only those topics that are in a topic class associated with one or more of the document's labels. This approach enables PLDA to detect extensive patterns in language usage correlated with each label.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experimental dataset</head><p>For our experiments, we use a dataset consisting of a collection of psychotherapeutic transcripts available for research. The conversations have been transcribed and collected according to the guidelines of the American Psychological Association (APA) <ref type="foot" target="#foot_2">5</ref> . An approval to use the collection was granted by an Internal Committee of Biomedical Experiments (ICBE) of Philips after a review of the agreements, the consent procedures, and data handling plan by legal and privacy experts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data description</head><p>Our dataset consists of 1729 transcripts of 1:1 conversation with a total of 340,455 talk turns, 75,732 unique terms, and more than 9 million words. Each transcript has on average 200 talk-turns and eight words for talk-turn. They are also extended with meta-data consisting of the corresponding school of psychotherapy, counselors-patients information such as gender, age range, and sexual orientation as well as a table of topics discussed during the therapeutic conversation. The table of topics contains two different kinds of information:</p><p>-Subjects: they are organized into three hierarchical levels. The top level is the most general whereas the other two are more precise. For example, the word Family could correspond to a top level topic, while Family violence and Child abuse would be associated to the second and third levels respectively. Up to 575 subjects have been used in the three levels in total. -Symptoms: there are 79 symptoms (e.g. Depression, Anger, Fear) as defined in the DSM-IV<ref type="foot" target="#foot_3">6</ref> manual.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Preprocessing of the meta-data</head><p>Given the high number of items in the table of topics, we applied the following steps to merge similar topics and reduce the number of subjects and symptoms to 18 and 16 respectively:</p><p>1. Eliminate all the subjects and symptoms that occur in less than 3% of the dataset; 2. Group together all the subjects belonging to the same Wikipedia category <ref type="foot" target="#foot_4">7</ref> regardless their position in the given hierarchical structure.</p><p>3. Assign a label to the new subject according to the psychology topics table from APA. For example Parent-child relationship and Family are mapped to a new subject from APA known as Parenting.</p><p>4. Reduce the number of symptoms by using the DSM-IV manual with the expert support of a counselor. In particular, we group symptoms with high-level correlation into a representative one. For example, Sadness and Hopelessness are merged into the symptom: Depression. The final set is depicted in Figure <ref type="figure" target="#fig_1">3</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Preprocessing the conversation text</head><p>Using the NLTK platform 8 we applied a number of NLP pre-processing steps to the dataset <ref type="bibr" target="#b20">[21]</ref>. The resulting corpus consists of 2,849,457 tokens (14,274 unique ones) and a total of 268,478 talk turns. The performed steps include:</p><p>tokenization, which transforms texts into a sequence of tokens removal of all punctuations, stop words, numbers, words that frequently appeared in the text with less content information (e.g., "mm-hmm") and words that occurred in less than five documents and keeping nouns, verbs and adjectives only. This was achieved by using unigram part-of-speech tagger <ref type="bibr" target="#b16">[17]</ref> to identify the types of words in each talk turn removal of the 100 most common words as well as talk turns with one word only as well as words shorter than three characters.</p><p>The proposed automatic TDT method has three phases; the detection of topics, assignment of topic labels to talk turns, and finally tracking of the propagation of topics over the conversation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Detection of Topics</head><p>The topic detection was performed using a PLDA implementation based on the Stanford Topic Modeling Toolbox<ref type="foot" target="#foot_5">9</ref> (TMT). The model requires the definition of parameters such as the number of hidden topics to discover, the hyperparameters α and η (see Figure <ref type="figure">2</ref>), and a vast amount of short text as input for training purpose. To enlarge the number of corpora, we defined each talk-turn as a document, and we associated each document (talk-turn) with the corresponding topics from the table of topics of the corresponding transcript because PLDA is useful in general only when each document has more than one label associated to it. As a result, we obtained a broader set of documents with higher word co-occurrences. More in detail, we needed to specify how many new topics (different from those in table of topics) the model had to discover. Experimentally this number was set to 20. As a further input, we fed the PLDA with the list of 34 topics and the values of α and η were set to 0.01. In total 268,478 talk turns were obtained after the preprocessing step. Moreover, we used the CVB0 algorithm <ref type="bibr" target="#b3">[4]</ref> where the overall number of iterations was 150. After training the model we obtained a list of topics and the associated learned words as shown in Table <ref type="table">1</ref>. The method also provided a perdocument topic distribution for each talk turn. An example is illustrated in Figure <ref type="figure" target="#fig_2">4</ref>, where the five topics with the highest likelihood in a conversation are shown (i.e. Stress and Job; Suicide and Death; Sexuality; Depression; Fear), each with the corresponding talk-turns.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Assignment of topics</head><p>The PLDA returns the topics and the associated terms for each talk turn (document).</p><p>Based on the terms of each document, we can then determine how likely each document was associated with a topic. Table <ref type="table">1</ref> shows a completion of the terms learned from the trained PLDA topic model. There we list the top ten terms for each topic. In the first column, one can see the discovered topic and its associated words whereas the second and third columns indicate two of the 34 topics already known and their related words. The terms in the same topic tend to be similar, particularly for subjects and symptoms. For example, Parenting carries the member of the family, such as mom, mother, dad, etc.. Moreover, Addiction includes the terms close to alcohol and drugs (drinking, smoke, etc.). On the other hand, the domain of the discovered topics has inherent interpretations and holds words that are not covered by the annotations in psychotherapy corpus. For example, the Topic-5 includes similar terms but their meaning (work) is far from any annotations in APA. For this reason, for the tracking task, we only used the 34 elements present in our table of topics.  <ref type="table">1</ref>. An illustrative example of the three kinds of topics and their most likely associated terms. Topic-5 shows an example of the discovered topic, Parenting presents an example of a known subject, and Addiction presents an example of a known symptom.</p><p>In addition to the 54 topics (34 known and 20 discovered by PLDA) and their relevant terms, we aimed to know the likelihood of each topic in each talk-turn. As such, another output of the PLDA topic model consisted in the partition of the documents into a set of 54 topic proportions, called per-document topic distributions, where each talk-turn was represented as a combination of topics with different proportions. As already explained earlier, Figure <ref type="figure" target="#fig_2">4</ref> shows an example of the potential ongoing topics in each talk-turn within a therapeutic conversation from the dataset. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Tracking Topics</head><p>The goal was to understand how the known topics propagate, are localized and change for each talker in conversations. We already used PLDA to identify a potential topic for each talk-turn from the table of topics converting the face-to-face conversation to a sequence of topics for each speaker. For the tracking topics task we added a new topic annotated as Meaningless talk which we associated to talk-turns that provide poor semantics contents or language, or non-verbal communication (e.g. "Yahh!!, Mm-hmm"). We built two topics transitions matrices (TTMs) to understand how the topics change from one talk-turn to another one. Topic changes can be seen as a dynamic mechanism that frequently occurs inside a conversation where speakers move from one topic to another and can be contested by either speaker. We recognize three kinds of changes: 1. The counselor keeps talking about the same topic to the patient from the previous talk-turn, and vice versa; 2. The counselor moves to a new topic after the talk-turn of the patient; 3. The patient moves to a new topic after the talk-turn of the counselor. More in detail we constructed patient-to-counselor TTM CP k that describes all the topics changes within the conversation k. In particular, CP k [i, j] is the number of times that topic i changes into j in the conversation k. We merged the CP k matrices together by summing up corresponding elements obtaining our final matrix CP . Similarly, we built a counselor-to-patient matrix P C by using the topics-change defined earlier by switching counselor and patient. The difference between the two matrices is illustrated in Figure <ref type="figure" target="#fig_3">5</ref>, which, shows the engagement patterns between counselors and patients, providing a new way to describe one-to-one conversations. There are three possibilities depending on whether the resulting value is lower than -10 (black), between -10 and +10 (gray) and greater than 10 (white). The diagonal of the matrix in Figure <ref type="figure" target="#fig_3">5</ref> gives an idea about the first type of topics changes which corresponds to the "resistance level" on the same topic; it has 17 gray values which proves the fact that the speakers like continuing to talk on the same topic. It has also twelve black values and six white values. The former means that the counselor switches topics twice as the patient does. A possible explanation is that the counselor aims at searching for other correlated symptoms or subjects that would lead to a mental disease. The other values of the matrix describe the second and third type of topic changes; the number of white and black values are approximately equal, which means that the conversations, in general, are discussed without perceived tactics. Nevertheless, some rows and columns are mostly negative or positive, like Parenting, and indicate the use of some strategies. The counselor often switches the topic if the previous one was Mania, Medication or Patient-Counselor Relations. Instead, he/she frequently starts a new topic if the patient's talk restrains less semantic contents (Meaningless Talk). Contrary, the patient often switches topics if the previous topic was related to Parenting, Friendship, Sexual dysfunction, Crying, or Stress-and-Work. Conclusively, TTM guides to a bright understanding of how and when topics are changing thus giving important insights to the counselor for CBT.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Evaluation</head><p>The evaluation of the performance of a topic model is not an easy task. In most cases, topics need to be manually evaluated by humans, which may express different opinions and annotations. The most common quantitative way to assess a probabilistic model is to measure the log-likelihood of a held-out test set performing perplexity. However, the authors in <ref type="bibr" target="#b8">[9]</ref> have shown that, surprisingly, perplexity and human judgment are often not correlated, and may infer less semantically meaningful topics. A potential solution to this problem is provided by the topic coherences, that is a typical way to assess qualitatively topic models by examining the most likely words in each topic. For such a purpose, we employed Palmetto<ref type="foot" target="#foot_6">10</ref> , a tool to compute topic coherence of a given word set with six different methods. The one that we selected for our purposes was the C V method <ref type="bibr" target="#b19">[20]</ref>, which uses word co-occurrences from the English Wikipedia, and that has been proven to highly correlate with human ratings. C V is depended on a one-set segmentation of the top words and a measure that uses normalized pointwise mutual information. The one-set segmentation computes the cosine similarity between each top words vector and the amount of all top words vectors. The coherence value is then the arithmetic average of these similarities and represents an intuitive measure of the goodness of the topics produced by PLDA. In this work, we evaluated our PLDA topic model for topic detection using C V coherence. In particular, we gave the top five terms (according to the weight of PLDA shown in Table <ref type="table">1</ref>) for each of the 34 topics as the input, obtaining as output a satisfactory coherence amongst all the detected topics. Indeed on average a topics coherence value larger than 50% was obtained, which is recognized in the research community already as a well-acceptable coherence score for a TDT model. This further substantiates the validity and potentials of our method.</p><p>Constituting a crucial aspect for the analysis and modeling of counselor-patient conversations, the automatic TDT task in a psychotherapeutic conversation poses a significant challenge. In this paper we implemented a topic detector which efficiently understands therapeutic discussions in consultations. We exploited PLDA and state-of-the-art NLP techniques and topic coherence evaluation systems. Furthermore, we computed TTM to capture the dynamics of each ongoing topic in the conversations understanding how much each interlocutor is affected in the dialogue and when he/she prefers switching topics. Knowing how topics change and their propagations can be used by counselors to drive the discussion and to detect a patient emotional state during the therapeutic conversation. These aspects of interaction are critical for all mental health specialists as they are involved in patient's health concerns. We conclude that PLDA and TTM may be of benefit to the therapeutic conversational speech analysis community, by having high potentials into real-life applications in psychotherapy.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig.1. Probabilistic graphical model of LLDA: unlike standard LDA, both the label set Λ as well as the topic prior α influence the topic mixture θ.<ref type="bibr" target="#b17">[18]</ref> </figDesc><graphic coords="5,201.38,115.84,212.60,86.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. The resulting 18 subjects (up) and 16 symptoms (bottom)</figDesc><graphic coords="7,170.20,115.84,274.97,345.22" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Example of the per-document topic distribution in each talk-turn on a conversation.</figDesc><graphic coords="9,173.03,419.73,269.30,121.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. The difference matrix between CP and P C.</figDesc><graphic coords="10,134.77,258.85,354.33,255.58" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">http://www.discursis.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">http://ckip.iis.sinica.edu.tw/taxonomy</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">http://www.apa.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">https://dsm.psychiatryonline.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4">https://en.wikipedia.org/wiki/Category:Main_topic_ classifications</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_5">https://nlp.stanford.edu/software/tmt/tmt-0.4/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_6">http://aksw.org/Projects/Palmetto.html</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Automatic Topic Labeling Using Ontology-Based Topic Models</title>
		<author>
			<persName><forename type="first">M</forename><surname>Allahyari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kochut</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 14th International Conference on Machine Learning and Applications (ICMLA)</title>
				<imprint>
			<date type="published" when="2015-12">2015. December 2015</date>
			<biblScope unit="page" from="259" to="264" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Human Communication as Coupled Time Series: Quantifying Multi-Participant Recurrence</title>
		<author>
			<persName><forename type="first">D</forename><surname>Angus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wiles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Audio, Speech, and Language Processing</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="1795" to="1807" />
			<date type="published" when="2012-08">August 2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Visualising conversation structure across time: insights into effective doctor-patient consultations</title>
		<author>
			<persName><forename type="first">D</forename><surname>Angus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Watson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gallois</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wiles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PloS One</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">e38014</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">On Smoothing and Inference for Topic Models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Asuncion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Welling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Smyth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">W</forename><surname>Teh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">UAI &apos;09</title>
				<meeting><address><addrLine>Arlington, Virginia, United States</addrLine></address></meeting>
		<imprint>
			<publisher>AUAI Press</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="27" to="34" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Learning the Structure of Task-Driven Human-Human Dialogs</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bangalore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">Di</forename><surname>Fabbrizio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stent</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Audio, Speech, and Language Processing</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="1249" to="1259" />
			<date type="published" when="2008-09">September 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Dynamic Topic Models</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Lafferty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICML &apos;06</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="113" to="120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Latent Dirichlet Allocation</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Jordan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="993" to="1022" />
			<date type="published" when="2003-03">March 2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Talking topically to artificial dialog partners: Emulating humanlike topic awareness in a virtual agent</title>
		<author>
			<persName><forename type="first">A</forename><surname>Breuing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Wachsmuth</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="volume">358</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Reading Tea Leaves: How Humans Interpret Topic Models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Boyd-Graber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gerrish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS&apos;09</title>
				<meeting><address><addrLine>USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="288" to="296" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">E-HowNet and Automatic Construction of a Lexical Ontology</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">C</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">J</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Association for Computational Linguistics</title>
				<meeting><address><addrLine>Stroudsburg, PA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="45" to="48" />
		</imprint>
	</monogr>
	<note>COLING &apos;10</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Development and research of Topic Detection and Tracking</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">7th IEEE International Conference on Software Engineering and Service Science (ICSESS)</title>
				<imprint>
			<date type="published" when="2016-08">2016. August 2016</date>
			<biblScope unit="page" from="170" to="173" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Conversation analysis: a method for research into interactions between patients and health-care professionals</title>
		<author>
			<persName><forename type="first">P</forename><surname>Drew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chatwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Collins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Health Expectations: An International Journal of Public Participation in Health Care and Health Policy</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="58" to="70" />
			<date type="published" when="2001-03">March 2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Content Coding of Psychotherapy Transcripts Using Labeled Topic Models</title>
		<author>
			<persName><forename type="first">G</forename><surname>Gaut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Steyvers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">E</forename><surname>Imel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Atkins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Smyth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Journal of Biomedical and Health Informatics</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="476" to="487" />
			<date type="published" when="2017-03">March 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Natural language processing</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gelbukh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Fifth International Conference on Hybrid Intelligent Systems (HIS&apos;05)</title>
				<imprint>
			<date type="published" when="2005-11">November 2005</date>
			<biblScope unit="page">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Investigating Topic Modelling for Therapy Dialogue Analysis</title>
		<author>
			<persName><forename type="first">C</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Purver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mccabe</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013-03">March 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Introduction-topic models: What they are and why they matter</title>
		<author>
			<persName><forename type="first">W</forename><surname>Mohr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bogdanov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Topic Models and the Cultural Sciences</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="page" from="545" to="569" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Analyzing Tagging Accuracy of Part-of-Speech Taggers</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">P P</forename><surname>Khin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">N</forename><surname>Aung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Intelligent Systems and Computing</title>
				<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2015-08">August 2015</date>
			<biblScope unit="page" from="347" to="354" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Labeled LDA: A Supervised Topic Model for Credit Attribution in Multi-labeled Corpora</title>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nallapati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Association for Computational Linguistics</title>
				<meeting><address><addrLine>Stroudsburg, PA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="248" to="256" />
		</imprint>
	</monogr>
	<note>EMNLP &apos;09</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Partially Labeled Topic Models for Interpretable Text Mining</title>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dumais</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD &apos;11</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="457" to="465" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Exploring the Space of Topic Coherence Measures</title>
		<author>
			<persName><forename type="first">M</forename><surname>Röder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Both</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hinneburg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WSDM &apos;15</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="399" to="408" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">The Impact of Preprocessing on Text Classification</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Uysal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Inf. Process. Manage</title>
		<imprint>
			<biblScope unit="volume">50</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="104" to="112" />
			<date type="published" when="2014-01">January 2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Topic detection and tracking for conversational content by using conceptual dynamic latent Dirichlet allocation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Yeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">216</biblScope>
			<biblScope unit="page" from="310" to="318" />
			<date type="published" when="2016-12">December 2016</date>
		</imprint>
	</monogr>
	<note>Supplement C</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
