<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="it">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Topic Modelling Games</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Rocco</forename><surname>Tripodi</surname></persName>
							<email>tripodi@di.uniroma1.it</email>
							<affiliation key="aff0">
								<orgName type="department">Sapienza NLP Group Department of Computer Science</orgName>
								<orgName type="institution">Sapienza University of Rome</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Topic Modelling Games</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">5B4592C8D12183DFDE8C002DD510FB3C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-19T15:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>English. This paper presents a new topic modelling framework inspired by game theoretic principles. It is formulated as a normal form game in which words are represented as players and topics as strategies that the players select. The strategies of each player are modelled with a probability distribution guided by a utility function that the players try to maximize. This function induces players to select strategies similar to those selected by similar players and to choice strategies not shared with those selected by dissimilar players. The proposed framework is compared with state-of-the-art models demonstrating good performances on standard benchmarks.</p><p>Italiano. Questo articolo presenta un approccio di modellazione dei topic ispirato alla teoria dei giochi. La modellazione dei topic è vista come un gioco in forma normale in cui le parole rappresentano i giocatori e i topic le strategie che i giocatori possono scegliere. Ogni giocatore sceglie le strategie da impiegare tramite una distribuzione di probabilità che viene influenzata da una funzione di utilità che i giocatori cercano di massimizzare. Questa funzione incentiva i giocatori a scegliere strategie simili a quelle impiegate da giocatori simili e disincentiva la scelta di strategie condivise con giocatori dissimili. Il confronto con modelli allo stato dell'arte dismostra buone prestazioni su diversi dataset di valutazione.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="it">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Topic modeling is a technique that discovers the underlying topics contained in a collection of doc-uments <ref type="bibr" target="#b2">(Blei, 2012;</ref><ref type="bibr" target="#b9">Griffiths and Steyvers, 2004)</ref>. It can be used in different tasks of text classification, document retrieval, and sentiment analysis, providing together vector representations of words and documents. State-of-the-art systems are based on probabilistic <ref type="bibr" target="#b1">(Blei et al., 2003;</ref><ref type="bibr">Mcauliffe and Blei, 2008;</ref><ref type="bibr" target="#b5">Chong et al., 2009)</ref> and neural networks models <ref type="bibr" target="#b0">(Bengio et al., 2003;</ref><ref type="bibr" target="#b10">Hinton and Salakhutdinov, 2009;</ref><ref type="bibr" target="#b13">Larochelle and Lauly, 2012;</ref><ref type="bibr" target="#b2">Cao et al., 2015)</ref>. A different perspective based on game theory is proposed in this article.</p><p>The use of game-theoretic principles in machine learning <ref type="bibr" target="#b8">(Goodfellow et al., 2014)</ref>, pattern recognition <ref type="bibr" target="#b17">(Pavan and Pelillo, 2007)</ref> and natural language processing <ref type="bibr" target="#b22">(Tripodi et al., 2016;</ref><ref type="bibr">Tripodi and Navigli, 2019)</ref> problems is developing a promising field of research with the development of original models. The main difference between computational models based on optimization techniques and game-theoretic models is that the former tries to maximize (minimize) a function (that in many cases is non-convex) and the latter tries to find the equilibrium state of a dynamical system. The equilibrium concept is useful because it represents a state in which all the constraints of a given system are satisfied and no object of the system has an incentive to deviate from it, because a different configuration will immediately lead to a worse situation in terms of payoff and fitness, at object and system level. Furthermore, it is guaranteed that the system converges to a mixed strategy Nash equilibrium <ref type="bibr" target="#b16">(Nash, 1951)</ref>. So far, game-theoretic models have been used in classification and clustering tasks <ref type="bibr" target="#b17">(Pavan and Pelillo, 2007;</ref><ref type="bibr" target="#b20">Tripodi and Pelillo, 2017)</ref>. In this work, it is proposed a gametheoretic model for inferring a low dimensional representation of words that can capture their latent semantic representation.</p><p>In this work, topic modeling is interpreted as a symmetric non-cooperative game <ref type="bibr" target="#b24">(Weibull, 1997)</ref> in which, the words are the players and the topics are the strategies that the players can select. Two players are matched to play the games together according to the co-occurrence patterns found in the corpus under study. The players use a probability distribution over their strategies to play the games and obtain a payoff for each strategy. This reward helps them to adjust their strategy selection in future games, considering what strategy has been effective in previous games. It allows concentrating more mass on the strategies that get high reward.</p><p>The underlying idea to model the payoff function is to create two influence dynamics, the first one forces similar players (words that appear in similar contexts) to select similar strategies; the second one forces dissimilar players (words that do not share any context) to select different strategies. The games are played repeatedly until the system converges, that is, the difference among the strategy distributions of the players at time t and at time t − 1 is under a small threshold. The convergence of the system corresponds to an equilibrium, a situation in which there is an optimal association of words and topics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Hofmann (1999) proposed one of the earliest topic models, probabilistic Latent Semantic Indexing (pLSI). It represents each word in a document as a sample from a mixture model, where topics are represented as multinomial random variables and documents as a mixture of topics. Latent Dirichlet Allocation (LDA) <ref type="bibr" target="#b1">(Blei et al., 2003)</ref>, the most widely used topic model, is a generalization of pLSI that introduces Dirichlet priors for both the word multinomial distributions over topics and topic multinomial distributions over documents. This line of research has been developed building on top of LDA different features to infer correlations among topics <ref type="bibr" target="#b12">(Lafferty and Blei, 2006)</ref> or to model jointly words and labels in a supervised way <ref type="bibr">(Mcauliffe and Blei, 2008)</ref>.</p><p>Topic models based on neural network principles have been introduced with the neural network language model proposed in <ref type="bibr" target="#b0">(Bengio et al., 2003)</ref>. This paradigm is very popular in NLP and many topic models are based on it because with these techniques it is possible to obtain a lowdimensional representation of the data. In particular, auto-encoders <ref type="bibr" target="#b18">(Ranzato and Szummer, 2008)</ref>, Boltzmann machines <ref type="bibr" target="#b10">(Hinton and Salakhutdinov, 2009)</ref> and autoregressive distributions <ref type="bibr" target="#b13">(Larochelle and Lauly, 2012)</ref> have been used to model documents with layer-wise neural network tools. <ref type="bibr">Neural Topic Model (NTM;</ref><ref type="bibr" target="#b2">(Cao et al., 2015)</ref>) tries to overcome some limitations of classical topic models, such as the initialization problem and the generalization to n-grams. It exploits word embedding to represent n-grams and uses backpropagation to adjust the weights of the network between the embedding and the word-topic and documenttopic layers. A general framework for topic modeling based also on neural networks is Sparse Contextual Hidden and Observed Language Autoen-codeR (SCHOLAR; <ref type="bibr" target="#b3">(Card et al., 2018)</ref>). It allows using covariates to influence the topic distributions and labels to include supervision. As Sparse Additive GEnerative models (SAGE; <ref type="bibr" target="#b7">(Eisenstein et al., 2011)</ref>)it can produce sparse topic representations but differently from it and Structural Topic Model (STM; <ref type="bibr" target="#b19">(Roberts et al., 2014)</ref>) it can easily consider a larger set of metadata. A graphical topic model was proposed by <ref type="bibr" target="#b7">Gerlach et al. (2018)</ref>. In this framework, the task of finding topical structures is interpreted as the task of finding communities in complex networks. It is particularly interesting because it shows analogies with traditional topic models and overcomes some of their limitations such as the bound with a Bayesian prior and the need to specify the number of topics in advance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Topic Modelling Games</head><p>Normal-form games consist of a finite set of players N = (1, .., n), a finite set of pure strategies, S i = {1, ..., m i } for each player i ∈ N and a payoff (utility) function u i : S → R, that associates a payoff to each combination of strategies S = S 1 × S 2 × ... × S n . The payoff function does not depend only on the strategy chosen by a single player but by the combination of strategies played at the same time by the players. Each player tries to maximize the value of u i . Furthermore, in noncooperative games the players choose their strategies independently, considering what other players can play and trying to find the best response to the strategy of the co-players. Nash equilibria <ref type="bibr" target="#b16">(Nash, 1951)</ref> represent the key concept of game theory and can be defined as those strategy combinations in which each strategy is a best response to the strategy of the co-player and no player has the incentive to unilaterally deviate from them because there is no way to do better. In addition to play pure strategies, that correspond to select-ing just one strategy from those available in S i , a player i can also use mixed strategies, which are probability distributions over pure strategies. A mixed strategy over S i is defined as a vector x i = (x 1 , . . . , x m i ), such that x j ≥ 0 and</p><p>x j = 1. In a two-player game, a strategy profile can be defined as a pair (x i , x j ). The expected payoff for this strategy profile is computed as:</p><formula xml:id="formula_0">u(x i , x j ) = x T i • A ij x j</formula><p>where A ij is the m i × m j payoff matrix between player i and j.</p><p>Evolutionary game theory <ref type="bibr" target="#b24">(Weibull, 1997)</ref> has introduced two important modifications: 1. the games are played repeatedly, and 2. the players update their mixed strategy over time until it is not possible to improve the payoff. The players, with these two modifications, can develop an inductive learning process, that allows them to learn their strategy distribution according to what other players are selecting. The payoff corresponding to the h-th pure strategy is computed as:</p><formula xml:id="formula_1">u(x h i ) = x h i • n i j=1 (A ij x j ) h<label>(1)</label></formula><p>The average payoff of player i is calculated as:</p><formula xml:id="formula_2">u(x i ) = m i h=1 u(x h i )<label>(2)</label></formula><p>To find the Nash equilibrium of the game, it is common to use the replicator dynamics equation <ref type="bibr" target="#b24">(Weibull, 1997)</ref>. It allows better than average strategies to grow at each iteration. It can be considered as an inductive learning process, in which the players learn from past experiences how to play their best strategy. It is important to notice that each player optimizes its individual strategy space, but this operation is done according to what other players simultaneously are doing so the local optimization is the result of a global process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Data Preparation</head><p>The players of the topic modelling games are the words v = (1, . . . , n) in the vocabulary V of the corpus under analysis and the strategies S = (1, . . . , m) are the topics to extract from the same corpus. The strategy space x i of each player i is represented as a probability distribution that can be interpreted as the mixture of topics typically used in topic modeling. The interactions among the players are modeled using the n × n adjacency matrix (W ) of an undirected weighted graph. Each entry w ij encodes the similarity between two words. The strategy space of the games can be represented as a n × m matrix X, where each row represents the probability distribution of a player over its m strategies (topics that have to be extracted from the corpus).</p><p>Payoff Function and System Dynamics The payoff function of the game is constructed exploiting the information stored in W . This matrix gives us the structural information of the corpus. It allows us to select the players with whom each player is playing the games, indicated with the presence of an edge between two nodes (players), and to quantify the level of influence that each player has on the other, indicated with the weight on each edge. The absence of an edge in this graph indicates that two words are distributional dissimilar. Using these three sources of information we model a payoff function that forces similar players to choose similar strategies (topics) and dissimilar players to choose different ones. The payoff of a player is calculated as,</p><formula xml:id="formula_3">u(x h i ) = x h i ( n i j=1 (A ij x j ) h − neg i g=1 ( x g ) h ) (3)</formula><p>where the first summation is over all the n i direct neighbors of player i that are the players with whom i share some similarity and the second summation is over the neg i negative players of player i, that are players with whom player i does not share any similarity. With the first summation player i will negotiate with its neighbors a correlated strategy (topic), with the second he will deviate from the strategies chosen by negative players, this is done by subtracting the payoff that i would have gained if these negative players would have been his neighbors. The negative players are sampled from V according to frequency, in the same way, negative samples are selected in word embeddings models <ref type="bibr" target="#b15">(Mikolov et al., 2013;</ref><ref type="bibr" target="#b21">Tripodi and Pira, 2017)</ref>. The equation that gives us the probability of selecting a word as negative is:</p><formula xml:id="formula_4">P (w i ) = f (w i ) 3/4 n j=0 f (w j ) 3/4 ,<label>(4)</label></formula><p>where f (w i ) is the frequency of word w i . Since the similarity with negative players is 0 we introduced the parameter to weight their influence and set it to (A &gt; 0). The number of negative players, neg i , is set to n i (number of neighbours of player i ).</p><p>Once the players have played all the games with their neighbors and negative players, the average payoff of each player can be calculated with Equation (2). The payoff is higher when two words are highly correlated and have a similar mixed strategy. For this reason the replicator dynamics equation <ref type="bibr" target="#b24">(Weibull, 1997)</ref> is used to compute the dynamics of the system. It pushes the players to be influenced by the mixed strategy of the co-players. This influence is proportional to the similarity between two players (A ij ). Once the influence dynamics do not affect the players the Nash equilibrium of the system is reached. The stopping criteria of the dynamics and are: 1. the maximum number of iterations (10<ref type="foot" target="#foot_2">5</ref> ); and 2. the minimum difference between two different iterations (10 −<ref type="foot" target="#foot_0">3</ref> ) that is calculated as n i=1 x i (t − 1) − x i (t).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experimental Results</head><p>In this section, we evaluate TMG and compare it with state-of-the-art systems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data and Setting</head><p>The datasets used to evaluate TMG are 20 Newsgroups 1 (20NG) and NIPS 2 . 20NG is a collection of about 20, 000 documents organized into 20 different classes. NIPS is composed of about 1, 700 NIPS conference papers published between 1987 and 1999 with no class information. Each text was tokenized and lowercased. The stop-words were removed and the vocabulary was constructed considering the 1000 and 2000 most frequent words in 20NG and NIPS, respectively. This choice is in line with previous work <ref type="bibr" target="#b3">(Card et al., 2018)</ref>. To keep the model as simple as possible, the tf-idf weighting was used to construct the feature vectors of the words and the cosine similarity was employed to create the adjacency matrix A. It is important to notice here that other sources of information can be easily included at this stage, derived from pre-trained word embeddings, syntactic structures or document metadata. Then A is sparsified taking only the r nearest neighbours of each node. r is calculated as r = log(n) this operation reduces the computational cost of the algorithm and guarantees that the graph remains connected (Von Luxburg, 2007). The strategy space of the players was initialized using a normal distribution to reduce the parameters of the framework 3 . The last two parameters of the systems concern the stopping criteria of the dynamics and are: 1. the maximum number of iterations (10 5 ); and 2. the minimum difference between two different iterations (10 −3 ) that is calculated as n i=1 x i (t − 1) − x i (t). TMG has been compared with SCHOLAR<ref type="foot" target="#foot_1">4</ref> , LDA 5 and NVDM<ref type="foot" target="#foot_3">6</ref> . We configured the NVDM network with two encoder layers (500-dimensional) and ReLu non-linearities. SCHOLAR has been configured using a more complex setting that consists in a single layer encoder and a 4-layer generator. LDA has been run with the following parameters: α = 50, iterations = 1000 and topic threshold = 0.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Evaluation</head><p>In this section, we compared the generalization performances of TMG and compared them with the models presented in the previous section. For the evaluation we used perplexity (PPL), even if it is has been shown to not correlate with human interpretation of topics <ref type="bibr" target="#b4">(Chang et al., 2009)</ref>. We computed perplexity on unobserved documents (C), as.</p><formula xml:id="formula_5">P P L(C) = exp(− 1 N N n=1 logP (C n ) N n=1 D n ) (5)</formula><p>where N is the number of documents in the collection C. Low perplexity suggests less uncertainties about the documents. Held out documents represent the 15% of each dataset. Perplexity is computed for 10 topics for the NIPS dataset and 20 topics for the 20 Newsgroups dataset. These numbers correspond to the real number of classes of each dataset.</p><p>Table <ref type="table" target="#tab_0">1</ref> shows the comparison of perplexity. As reported in previous work <ref type="bibr" target="#b3">(Card et al., 2018)</ref>, it is difficult to achieve a lower perplexity than LDA. The results in these experiments follow the same pattern, with LDA that has the lowest perplexity, TMG, and SCHOLAR that have similar results, and NVDM that performs slightly worse on both datasets.   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Topic Coherence and Interpretability</head><p>It has been shown that perplexity does not necessarily correlate well with topic coherence <ref type="bibr" target="#b4">(Chang et al., 2009;</ref><ref type="bibr" target="#b19">Srivastava and Sutton, 2017)</ref>. For this reason, we evaluated the performances of our system also on coherence <ref type="bibr" target="#b4">(Chang et al., 2009;</ref><ref type="bibr" target="#b6">Das et al., 2015)</ref>. The coherence is calculated by computing the relatedness between topic words using the pointwise mutual information (PMI). We used Wikipedia (2018.05.01 dump) as corpus to compute co-occurrence statistics using a sliding window of 5 words on the left and on the right of each target word. For each topic, we selected the 10 words with the highest mass. Then we calculated the PMI among all the words pair and finally compute the coherence as the arithmetic mean of all these values. This metric has been shown to correlate well with human judgments <ref type="bibr" target="#b14">(Lau et al., 2017)</ref>. We used two different sources of information for the computation of the PMI: one is internal and corresponds to the dataset under analysis; the other one is external and is represented by the English Wikipedia corpus.</p><p>Internal PMI Figure <ref type="figure" target="#fig_0">1</ref> presents the PMI values of the different models computed on the two corpora. As it is possible to see from figure <ref type="figure" target="#fig_0">1a</ref>, TMG has a low PMI compared to all other systems on the 20 Newsgroups dataset when there are few topics to extract (i.e.: 2 and 5). The situation changes drastically when the number of topics increases. In fact, it has the highest performances on this dataset when extracts 10, 20, 50, 100 topics. The performances of NDVM and SCHOLAR are similar and follow a decreasing pattern, with very high values at the beginning. On the contrary, the performances of LDA follow an opposite pattern this model seems to work better when the number of topics to extract is high. On NIPS (Figure <ref type="figure" target="#fig_0">1b</ref>) the performances of the systems are similar to those on 20 Newsgroups. The only exception is that TMG has always the highest PMI and seems to behave better also when the number of topics to extract is high. This probably because the number of words in NIPS is higher and for this, it is reasonable to have also a higher number of topics. This is also confirmed from a qualitative analysis of the topics in Section 4.4, where it is demonstrated that with low values of k it is possible to extract general topics and increasing its value it is possible to extract more specific ones.</p><p>In general, we can find three different patterns in these experiments: 1. NDVM and SCHOLAR work well on extracting a low number of topics; 2. LDA works well when it has to extract a large number of topics; 3. TMG works well on extracting a number of topics that is close to the real number of classes in the datasets. Another aspect to take into account is the fact that even if TMG has the highest performances, its results have also a high standard deviation. This is due to the stochastic nature of negative sampling.  Sparsity We compared the sparsity of the wordtopics matrices, X, in Figure <ref type="figure" target="#fig_2">3a</ref> and 3b, computed as s</p><formula xml:id="formula_6">= |X&gt;10 −3 | |X| .</formula><p>From both figures, we can see that TMG can produce highly sparse representations especially when the number of topics to extract is low. This is a nice feature since it provides more interpretable results. Only SCHOLAR produces more sparse representations when the number of topics to extract is high. Experimentally we also noticed that we can control the sparsity of X, in TMG, increasing the number of iterations of the game dynamics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Qualitative Evaluation</head><p>Examples of topics extracted from 20NG and NIPS are presented in Table <ref type="table" target="#tab_2">2 and 3, respectively 7</ref> . The first difference that emerges from these results are the external PMI values. This is due to the fact that the texts in NIPS have a very specific language and for this reason the PMI values are very high. We can also see that TMG groups highly coherent set of words in each topic. We can easily identify in Table <ref type="table" target="#tab_1">2</ref> the topics in which the dataset is organized and especially: talk.politics. midleast,  alt.atheism, comp.graphics, soc.religion.christian,  talk.politics.misc, rec.motorcycles, sci.crypt,  talk.politics.guns, rec.sport.hockey, sci.space,  talk.politics.misc.   7 for space limitation we presented only 15 topics for 20NG</p><p>We can also easily identify from Table <ref type="table" target="#tab_2">3</ref> highly coherent topics, related to optic, signal analysis, optimization, crowdsourcing, audio, graph theory and logics. We noticed from these topics that they are general and that it is possible to discover more specific topics increasing the number of topics to extract. For example, we discovered topics related to topic modelling and generative adversarial networks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion and Future Work</head><p>In this paper, it is presented a new topic modeling framework based on game-theoretic principles. The results of its evaluation show that the model performs well compared to state-of-the-art systems and that it can extract topically and semantically related groups of words. In this work, the model was left as simple as possible to assess if a game-theoretic framework itself is suited for topic modeling. In future work, it will be interesting to introduce the topic-document distribution and to test it on classification tasks and covariates to extract topics using different dimensions, such as time, authorship, or opinion. The framework is open and flexible and in future work, it will be tested with different initializations of the strategy space, graph structures, and payoff functions. It will be particularly interesting to test it using word embedding and syntactic information.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Internal PMI mean and std values.</figDesc><graphic coords="5,72.00,163.14,109.12,81.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: External PMI mean and std values.</figDesc><graphic coords="5,72.00,311.91,109.12,81.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Sparsity mean and std values.</figDesc><graphic coords="5,72.00,460.68,109.12,81.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Comparison of the models as perplexity.</figDesc><table><row><cell cols="5">Dataset TMG SCHOLAR NVDM LDA</cell></row><row><cell>20NG</cell><cell>824</cell><cell>819</cell><cell>927</cell><cell>791</cell></row><row><cell>NIPS</cell><cell>1311</cell><cell>1370</cell><cell>1564</cell><cell>1017</cell></row></table><note>1 http://qwone.com/ jason/20Newsgroups/ 2 http://www.cs.nyu.edu/ roweis/data.html</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Best topics (each topic is represented on the columns) extracted from 20 Newsgroup using TMG (setting k = 20) ordered using external PMI (bottom row).</figDesc><table><row><cell>turks</cell><cell>schneider</cell><cell>drive</cell><cell></cell><cell>vms</cell><cell>god</cell><cell></cell><cell>intellect</cell><cell>bike</cell><cell>providing</cell><cell>fbi</cell><cell>gun</cell><cell>team</cell><cell>space</cell><cell>male</cell><cell>tim</cell><cell>amateur</cell></row><row><cell>soviet</cell><cell>allan</cell><cell>ide</cell><cell></cell><cell>disclaimer</cell><cell>jesus</cell><cell></cell><cell>banks</cell><cell>ride</cell><cell cols="3">encryption compound firearms</cell><cell>game</cell><cell>orbit</cell><cell>gay</cell><cell>israel</cell><cell>georgia</cell></row><row><cell>turkish</cell><cell>morality</cell><cell>scsi</cell><cell></cell><cell>vnews</cell><cell>christians</cell><cell></cell><cell>gordon</cell><cell>riding</cell><cell>clipper</cell><cell>batf</cell><cell>guns</cell><cell>play</cell><cell>shuttle</cell><cell>men</cell><cell>israeli intelligence</cell></row><row><cell>armenian</cell><cell>keith</cell><cell cols="2">controller</cell><cell>vax</cell><cell>christ</cell><cell></cell><cell>surrender</cell><cell>dod</cell><cell>key</cell><cell>fire</cell><cell cols="3">criminals season launch</cell><cell>sexual</cell><cell>arab</cell><cell>ai</cell></row><row><cell>armenia</cell><cell>atheists</cell><cell>drives</cell><cell></cell><cell cols="3">necessarily christianity</cell><cell>univ</cell><cell>bikes</cell><cell>escrow</cell><cell>waco</cell><cell>crime</cell><cell>hockey</cell><cell>earth</cell><cell cols="2">percentage jews</cell><cell>programs</cell></row><row><cell>passes</cell><cell>moral</cell><cell>mb</cell><cell></cell><cell>represents</cell><cell>bible</cell><cell></cell><cell cols="2">pittsburgh motorcycle</cell><cell>crypto</cell><cell>children</cell><cell cols="3">weapons league mission</cell><cell>study</cell><cell>arabs</cell><cell>michael</cell></row><row><cell>roads</cell><cell>political</cell><cell>disk</cell><cell></cell><cell>views</cell><cell>christian</cell><cell></cell><cell>significant</cell><cell>bmw</cell><cell>keys</cell><cell>koresh</cell><cell>criminal</cell><cell>nhl</cell><cell>flight</cell><cell>sex</cell><cell>policy</cell><cell>radio</cell></row><row><cell cols="2">armenians pasadena</cell><cell>isa</cell><cell></cell><cell>expressed</cell><cell>faith</cell><cell></cell><cell>hospital</cell><cell>honda</cell><cell>chip</cell><cell>gas</cell><cell>violent</cell><cell>players</cell><cell>nasa</cell><cell>apparent</cell><cell>war</cell><cell>adams</cell></row><row><cell>argic</cell><cell>objective</cell><cell>bus</cell><cell></cell><cell>news</cell><cell>church</cell><cell></cell><cell>level</cell><cell>road</cell><cell>secure</cell><cell>branch</cell><cell>weapon</cell><cell>cup</cell><cell>moon</cell><cell>showing</cell><cell>land</cell><cell>ignore</cell></row><row><cell cols="2">proceeded animals</cell><cell>floppy</cell><cell></cell><cell>poster</cell><cell>belief</cell><cell></cell><cell>blood</cell><cell>advice</cell><cell>wiretap</cell><cell>started</cell><cell>armed</cell><cell>stanley</cell><cell>solar</cell><cell>women</cell><cell>north</cell><cell>occur</cell></row><row><cell>29.71</cell><cell>15.27</cell><cell>12.7</cell><cell></cell><cell>11.72</cell><cell>10.79</cell><cell></cell><cell>10.18</cell><cell>8.94</cell><cell>8.93</cell><cell>8.55</cell><cell>7.52</cell><cell>7.45</cell><cell>7.14</cell><cell>6.92</cell><cell>6.21</cell><cell>6.13</cell></row><row><cell>ocular</cell><cell cols="2">dendrites</cell><cell></cell><cell>oscillatory</cell><cell cols="4">crowdsourcing kaiming</cell><cell cols="2">retina</cell><cell>auditory</cell><cell cols="2">graph</cell><cell cols="2">disturbances</cell><cell>lifted</cell></row><row><cell>eye</cell><cell cols="2">dendritic</cell><cell></cell><cell>oscillations</cell><cell></cell><cell cols="2">crowds</cell><cell cols="3">shaoqing photoreceptor</cell><cell>sound</cell><cell cols="2">edges</cell><cell>plant</cell><cell>propositional</cell></row><row><cell>fovea</cell><cell cols="2">soma</cell><cell></cell><cell>oscillators</cell><cell cols="3">workers</cell><cell>xiangyu</cell><cell cols="2">retinal</cell><cell>sounds</cell><cell cols="2">graphs</cell><cell>controllers</cell><cell>predicate</cell></row><row><cell cols="3">dominance dendrite</cell><cell></cell><cell>oscillator</cell><cell></cell><cell cols="2">worker</cell><cell>jian</cell><cell cols="2">vertebrate</cell><cell cols="3">cochlear optimisation</cell><cell>controller</cell><cell>grounding</cell></row><row><cell>saccades</cell><cell cols="2">axonal</cell><cell></cell><cell>oscillation</cell><cell cols="3">labelers</cell><cell>yangqing</cell><cell cols="2">schulten</cell><cell>ear</cell><cell>edge</cell><cell></cell><cell>disturbance</cell><cell>predicates</cell></row><row><cell>saccadic</cell><cell cols="2">axons</cell><cell cols="3">synchronization</cell><cell cols="2">crowd</cell><cell>karen</cell><cell cols="3">photoreceptors hearing</cell><cell cols="2">vertices</cell><cell>plants</cell><cell>domingos</cell></row><row><cell>fixation</cell><cell cols="2">nmda</cell><cell></cell><cell>decoding</cell><cell></cell><cell></cell><cell>turk</cell><cell>sergey</cell><cell cols="2">ganglion</cell><cell>ears</cell><cell cols="2">optimise</cell><cell>activate</cell><cell>clauses</cell></row><row><cell>foveal</cell><cell cols="2">pyramidal</cell><cell></cell><cell>locking</cell><cell cols="3">wisdom</cell><cell>trevor</cell><cell cols="2">kohonen</cell><cell>acoust</cell><cell cols="2">optimising</cell><cell>activated</cell><cell>compilation</cell></row><row><cell>eyes</cell><cell cols="2">somatic</cell><cell></cell><cell>synchronize</cell><cell cols="3">expertise</cell><cell>sergio</cell><cell cols="2">bipolar</cell><cell>tone</cell><cell cols="2">optimised</cell><cell>activating</cell><cell>formulas</cell></row><row><cell>saccade</cell><cell cols="2">axon</cell><cell></cell><cell>synchronized</cell><cell></cell><cell></cell><cell>dawid</cell><cell>jitendra</cell><cell cols="2">visualizing</cell><cell>cochlea</cell><cell cols="2">vertex</cell><cell>activates</cell><cell>logical</cell></row><row><cell>304.85</cell><cell cols="2">283.66</cell><cell></cell><cell>276.39</cell><cell></cell><cell></cell><cell>230.5</cell><cell>218.51</cell><cell cols="2">196.86</cell><cell>176.75</cell><cell cols="2">146.3</cell><cell>146.25</cell><cell>145.84</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Topics extracted from NIPS using TMG (setting k = 10) ordered using external PMI (bottom row).</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">Experimentally it was also observed that using a Dirichlet distribution to initialize the strategy space with different α parameters did not affect much the performances of the model.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">4 https://github.com/dallascard/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">scholar 5 http://mallet.cs.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">umass.edu 6 https://github.com/ysmiao/nvdm</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A neural probabilistic language model</title>
		<author>
			<persName><surname>Bengio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of machine learning research</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1137" to="1155" />
			<date type="published" when="2003-02">2003. 2003. Feb</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Latent dirichlet allocation</title>
		<author>
			<persName><surname>Blei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of machine Learning research</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="993" to="1022" />
			<date type="published" when="2003-01">2003. 2003. Jan</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A novel neural topic model and its supervised extension</title>
		<author>
			<persName><forename type="first">M</forename><surname>David</surname></persName>
		</author>
		<author>
			<persName><surname>Blei</surname></persName>
		</author>
		<author>
			<persName><surname>Cao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AAAI</title>
				<imprint>
			<date type="published" when="2012-04">2012. April. 2015</date>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="2210" to="2216" />
		</imprint>
	</monogr>
	<note>Probabilistic topic models</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Neural models for documents with metadata</title>
		<author>
			<persName><surname>Card</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the ACL</title>
				<meeting>the 56th Annual Meeting of the ACL</meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="2031" to="2040" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Reading tea leaves: How humans interpret topic models</title>
		<author>
			<persName><forename type="first">Chang</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="288" to="296" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Simultaneous image classification and annotation</title>
		<author>
			<persName><forename type="first">Chong</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Conference on</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2009">2009. 2009. 2009</date>
			<biblScope unit="page" from="1903" to="1910" />
		</imprint>
	</monogr>
	<note>CVPR 2009</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Gaussian lda for topic models with word embeddings</title>
		<author>
			<persName><surname>Das</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 53rd Annual Meeting of the ACL</title>
				<meeting>the 53rd Annual Meeting of the ACL</meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="795" to="804" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Sparse additive generative models of text</title>
		<author>
			<persName><surname>Eisenstein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Science Advances</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">7</biblScope>
			<date type="published" when="2011">2011. 2011. 2018. 2018</date>
		</imprint>
	</monogr>
	<note>A network approach to topic models</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Generative adversarial nets</title>
		<author>
			<persName><surname>Goodfellow</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="2014">2014. 2014</date>
			<biblScope unit="page" from="2672" to="2680" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Finding scientific topics</title>
		<author>
			<persName><forename type="first">Steyvers2004</forename><forename type="middle">;</forename><surname>Griffiths</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Griffiths</surname></persName>
		</author>
		<author>
			<persName><surname>Steyvers</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the National academy of Sciences</title>
		<imprint>
			<biblScope unit="volume">101</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5228" to="5235" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
	<note>suppl</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Replicated softmax: an undirected topic model</title>
		<author>
			<persName><forename type="first">Salakhutdinov2009</forename><surname>Hinton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Geoffrey</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ruslan</forename><forename type="middle">R</forename><surname>Salakhutdinov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="1607" to="1614" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Probabilistic latent semantic indexing</title>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Hofmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd annual international ACM SIGIR conference</title>
				<meeting>the 22nd annual international ACM SIGIR conference</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="50" to="57" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Correlated topic models</title>
		<author>
			<persName><forename type="first">Blei2006</forename><forename type="middle">;</forename><surname>Lafferty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>John</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><forename type="middle">M</forename><surname>Lafferty</surname></persName>
		</author>
		<author>
			<persName><surname>Blei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="147" to="154" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A neural autoregressive topic model</title>
		<author>
			<persName><forename type="first">Lauly2012</forename><surname>Larochelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hugo</forename><surname>Larochelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stanislas</forename><surname>Lauly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="2708" to="2716" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Topically driven neural language model</title>
		<author>
			<persName><surname>Lau</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 55th Annual Meeting of the ACL</title>
				<editor>
			<persName><forename type="first">Jon</forename><forename type="middle">D</forename><surname>Mcauliffe</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">David</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</editor>
		<meeting>the 55th Annual Meeting of the ACL</meeting>
		<imprint>
			<date type="published" when="2008">2017. 2017. 2008</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="121" to="128" />
		</imprint>
	</monogr>
	<note>Supervised topic models. In NIPS</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><surname>Mikolov</surname></persName>
		</author>
		<idno>CoRR, abs/1301.3781</idno>
		<imprint>
			<date type="published" when="2013">2013. 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Non-cooperative games</title>
		<author>
			<persName><forename type="first">John</forename><surname>Nash</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annals of mathematics</title>
		<imprint>
			<biblScope unit="page" from="286" to="295" />
			<date type="published" when="1951">1951</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Dominant sets and pairwise clustering</title>
		<author>
			<persName><forename type="first">Pelillo2007</forename><surname>Pavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Massimiliano</forename><surname>Pavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marcello</forename><surname>Pelillo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE transactions on pattern analysis and machine intelligence</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Semi-supervised learning of compact document representations with deep networks</title>
		<author>
			<persName><forename type="first">Szummer2008</forename><forename type="middle">;</forename><surname>Ranzato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aurelio</forename><surname>Marc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Ranzato</surname></persName>
		</author>
		<author>
			<persName><surname>Szummer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th international conference on Machine learning</title>
				<meeting>the 25th international conference on Machine learning</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="792" to="799" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Game theory meets embeddings: a unified framework for word sense disambiguation</title>
		<author>
			<persName><surname>Roberts</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2014">2014. 2014. 2017. 2019. November</date>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="88" to="99" />
		</imprint>
	</monogr>
	<note>International Conference on Learning Representations (ICLR)</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A game-theoretic approach to word sense disambiguation</title>
		<author>
			<persName><forename type="first">Pelillo2017</forename><surname>Tripodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rocco</forename><surname>Tripodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marcello</forename><surname>Pelillo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="31" to="70" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Analysis of italian word embeddings</title>
		<author>
			<persName><forename type="first">Pira2017</forename><surname>Tripodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rocco</forename><surname>Tripodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stefano</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pira</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fourth Italian Conference on Computational Linguistics (CLiC-it 2017)</title>
				<meeting>the Fourth Italian Conference on Computational Linguistics (CLiC-it 2017)<address><addrLine>Rome, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017-12-11">2017. December 11-13, 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Context aware nonnegative matrix factorization clustering</title>
		<author>
			<persName><surname>Tripodi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">23rd International Conference on Pattern Recognition, ICPR 2016</title>
				<meeting><address><addrLine>Cancún, Mexico</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-12-04">2016. 2016. December 4-8, 2016</date>
			<biblScope unit="page" from="1719" to="1724" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A tutorial on spectral clustering</title>
		<author>
			<persName><forename type="first">Von</forename><surname>Luxburg2007</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Ulrike</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Von</forename><surname>Luxburg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistics and computing</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="395" to="416" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Weibull</surname></persName>
		</author>
		<title level="m">Evolutionary game theory</title>
				<imprint>
			<publisher>MIT press</publisher>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
