=Paper=
{{Paper
|id=Vol-2028/paper22
|storemode=property
|title=Knowledge Transfer in Artificial Learning
|pdfUrl=https://ceur-ws.org/Vol-2028/paper22.pdf
|volume=Vol-2028
|dblpUrl=https://dblp.org/rec/conf/iccbr/Murena17
}}
==Knowledge Transfer in Artificial Learning==
204 Research Summary: Knowledge Transfer in Artificial Learning Pierre-Alexandre Murena Télécom ParisTech - Université Paris Saclay, 46 rue Barrault, 75013 Paris, France, murena@telecom-paristech.fr 1 Introduction This document presents the main research problems addressed during my PhD studies. All these researches are led inside the two teams DBWeb in Télécom Paris- Tech and LInK (Learning and Integration of Knowledge) in AgroParisTech, both located in Paris, and supervised by Pr. Jean-Louis Dessalles and Pr. Antoine Cornuéjols. My researches focus on learning theory both in the perspective of symbolic machine learning and of learning in continuous domains. I aim at finding an information-theoretic principle guiding information transfer in learning. The start point of these researches is the idea that most machine learning takes a strong stationary hypothesis for granted. The general framework of sta- tistical learning (mainly supervised and semi-supervised learning) considers two data sets: a learning data set (from which the concepts have to be learned) and a test data set (on which quality of the learned concepts is evaluated). The key idea of current learning methods and theories is to assume that training data and test data are independent and identically distributed (i.i.d.). However this strong hy- pothesis does not hold in many cases: either the data generation process evolves over time (aging effect, trending effect...) or the data belong to a different do- main. Because similar questions of transfer and domain adaptation had already been addressed in analogical reasoning, we proposed to use an approach based on Kolmogorov complexity instead of probabilities. Kolmogorov complexity is a measure of the information contained inside an object. The use of Kolmogorov complexity in machine learning is accepted by the community, but mainly in a stationary point of view (when the key concept does not vary); we proposed to extend its use to non-stationary environments, in the same way as done in analogical reasoning. A presentation of these issues is given in section 2. The strong similarity between transfer learning and analogical reasoning led me to consider this issue in my researches. Analogical reasoning consists in situa- tions of the form “‘b’ is to ‘a’ as ‘d’ is to ‘c’ ”. Because its value has already been demonstrated, I focus on Hofstadter’s micro-world, made up of strings of alpha- betical characters that can be described with simple concepts like ‘predecessor’, ‘successor’ or ‘repetition’. The use of Kolmogorov complexity for analogical rea- soning had already been considered, but our approach is slightly different. We Copyright © 2017 for this paper by its authors. Copying permitted for private and academic purpose. In Proceedings of the ICCBR 2017 Workshops. Trondheim, Norway 205 developed a small descriptive language for Hofstadter’s problems and convert it into a binary code, the length of which corresponds to Kolmogorov complexity. Our work on analogy is presented in section 3. Finally, because of its very global perspective, our research topic leads naturally to collaborations on various topics related to learning. This side aspect of my research is presented in section 4. 2 A global approach of learning Statistical machine learning in its current form is often considered to be based mainly on three inductive principles: Empirical Risk Minimization (ERM), Bayesian- ism and Minimum Description Length (MDL). The validity of ERM has been demonstrated under the strong i.i.d. assumption using several frameworks, all inspired by the Probably Approximately Correct learning framework. Besides strong links have been stated between Bayesianism and MDL. When the i.i.d. hypothesis is not verified, these three principles are not valid anymore and have to be replaced by new principles. Exploring this direction, we considered the most straightforward transfer learning problem and the classifi- cation problem, the purpose of which is to associate data to labels. Given source data XS associated to their classes YS and a target data XT , we aim at finding the corresponding classes YT . The idea is to find a classification function βS such that βS (XS ) = YS and to transfer this function into a classification function βT available on the target data XT . Because this problem is the same as analogical reasoning, we used a simplification of the general MDL principle in the context of analogy [2]. The mathematical tool used to measure the description length in MDL is Kol- mogorov complexity [4]. Kolmogorov complexity (also called algorithmic com- plexity) of an object x is an information theoretic measure defined as the minimal length of a program defined on a Universal Turing Machine (UTM) and the out- put of which is object x. This quantity can be shown to be machine independent, but non calculable. The idea we developed is to consider an upper-bound of this complexity based on a restricted Turing machine. The choice of a restricted Turing ma- chine corresponds to an inductive bias, inevitable in any inductive reasoning (see for example the no-free-lunch theorem). This choice also raises the problem of machine dependency which seems crucial in human learning. Our first contribution is a direct use of analogical MDL in the context of transfer learning [5]: I presented a two-part MDL equation based on a data rep- resentation called model. A model is any object which may be used to compress data. In transfer learning, our purpose is to infer a source model MS and a target model MT : min C(MS ) + C(XS |MS ) + C(YS |MS , XS ) + C(MT |MS ) + C(XT |MT ) (1) MS ,MT where C(.) designates Kolmogorov complexity, X the data, YS the source labels and M the model. This equation applies both for continuous data (X is a matrix, 206 the rows of which correspond to a vector data point) and for symbolic data (X is a sequence of symbols, for example a sequence of letters). The different terms in the equation present a strong similarity with usual machine learning terms: C(M ) corresponds to a model penalization based on its complexity; C(X|M ) corresponds to a likelihood term, ie. a fitness of the model toward data; and the term C(Y |M, X) corresponds to an empirical risk. In the paper, we also proposed experimental validations on two toy data sets with a simple prototype- based model. They present good results and high performance on these data sets. A direct variant of this formula has been proposed for incremental learn- ing [6]. In incremental learning, the system faces a sequence of questions X1 , X2 , . . . and has to find the solution to each problem one by one. The model used to de- scribe data can be updated if it is outdated and does not correspond to the current data anymore. We propose the following simplified MDL objective: X min C(Mt |M∆−1 ({1}) ) + C(Xt |Mt ) + C(βt |Mt , Xt ) + C(Yt |Mt , Xt , βt ) (2) M t t where ∆t is a model association function such that ∆t (u) = 1 if model Mt can be described with model Mu , and ∆t (u) = 0 otherwise. The consistency of our framework with existing heuristics state-of-the-art methods has been established, as well as the validity of a naive algorithm based on the same neural model as presented for transfer learning. The successful results obtained with MDL so far encourage future research tracks. Several problems emerge from the developed framework. First, the trans- fer objective 1 is valid for one target only. In practice, several target problems may occur, hence a multi-target variant of transfer has to be given. In partic- ular, i.i.d. hypothesis consists in assuming infinitely-many targets with specific distributions. I am currently exploring an approach based on Pareto-optimality, implying a prior over the future and thus a new learning concept: concern for future question. Another question of interest is the theoretical validity of such approaches. Unlike statistical learning which has a clear measure of quality (given by the risk), an approach based on MDL does not present any natural quality measure. Such a function has to be found before an equivalent of PAC learning can be developed. Additionally, incremental learning methods do not have access to the whole objective function 2 at all steps: only local optimizations are possible. A measure of the impact of this algorithmic simplification appears as a direct consequence. Finally, we aim at finding a geometric interpretation of these equa- tions. An interesting track is offered by the domain of information geometry and probability distribution manifolds. Under some specific conditions, Kolmogorov complexity may be associated to a probability distributed, hence considered in the perspective of information geometry. 207 3 Approaches of analogy Because of its crucial role in my researches on learning, I attach great impor- tance to studying analogical reasoning. For now on, my researches focused on Hofstadter’s micro-world [3] which presents highly general characteristics of ana- logical thought. I will work on this research track with Dr. Laurent Orseau. Preliminary works have already given insightful results and promising per- spectives. I chose to work on the development of a small prototype language generating Hofstadter’s analogies (of the form ABC : ABD :: IJK : IJL, which has the be read “ABD is to ABC what IJL is to IJK”). Among other specifications, the proposed language had to be generative, ie. describe a dy- namic generation rather than a static description (as opposed to the description in [2] for instance). For example, the string ABC will be generated by the pro- gram alphabet,sequence,3, which can be read as “Consider the alphabet and take the sequence of first three letters”. Once such a language is defined, it is turned into a prefix-delimited binary code, the length of which measures an upper-bound of Kolmogorov complexity. A more elaborated and general version of the language has been recently proposed. This memory-based language offers a flexibility in the management of prior knowledge of the user and offers a simple way to set priority to operations: its grammar enables any possible operator, as long as the operator can be put in long-term memory. The complexity of an element in memory is defined as the complexity of its depth in memory. A more rigorous presentation of this new framework including the considerations on memory will be presented at ICCBR 2017 Computational Analogy Workshop [7]. We propose a validation of our approach with a human experiment. In an online survey, we submitted a few Hofstadter’s problems and asked participants for their most intuitive solution. We shew that the majority answers correspond to local minima of Kolmogorov complexity. These results are not yet published. The proposed approach offers a tool to compare two results of an analogical problem when the generative instructions are given to the system. A first logical direction is to provide automatic instruction generation, hence software able to produce an optimal generative instruction for any complete analogy. Once this will done, I have to find a solution to an analogical problem (eg. find the best solution to ABC:ABD::IJK:?): because the space of solutions is infinite, it cannot be explored naively and research biases have to be found. In order to address these issues, I am currently working on a Python interpreter for the developed language. 4 Collaborations and side problems In the context of my research, I have the opportunity to collaborate on several projects related to non-stationarity and transfer. A first project is led jointly with Dr. Jérémie Sublime (ISEP) and Dr. Basarab Matei (Université Paris 13) and concerns collaborative clustering. Classic clus- tering consists in associating similar data together in clusters. Collaborative and 208 multi-view clustering is a framework in which several clustering algorithms are involved and try to influence each other. The algorithms do not produce the same number of clusters, the same underlying model nor the same final solution. This problem is closely related to transfer because it involves sharing informa- tion between several different domains. A first contribution has been proposed using operational research tools to select relevant collaborators in an existing probabilistic framework [9]. A brand new approach, based on complexity, is be- ing developed: we expressed the problem of collaborative clustering in terms of data compression and worked on minimal assumptions to obtain a tractable model. This new perspective offers a theoretical background for a wide range of heuristic state-of-the-art approaches and inspires a new algorithm [8]. A new related collaboration has been engaged recently with Dr. Cristina Manfredotti, Pr. Juliette Dibie (AgroParisTech) and Dr. Fatiha Sais (LRI), exploring common approaches in transfer learning and structural mappings in knowledge bases (ontology alignment). Finally, I am exploring the problem of Transfer Learning using boosting al- gorithms with Pr. Antoine Cornuéjols and Sema Akkoyunlu. The use of boosting may offer new perspectives on transfer in general and be beneficial to my under- standing of transfer mechanisms. A first contribution has been submitted [1]. References 1. Cornuéjols, A., Akkoyunlu, S., Murena, P.A., Olivier, R.: Transfer learning by boost- ing projections from the target domain to the source domain, submitted to NIPS 2017 2. Cornuéjols, A., Ales-Bianchetti, J.: Analogy and induction : which (missing) link? In: Workshop “Advances in Analogy Research : Integration of Theory and Data from Cognitive, Computational and Neural Sciences”. Sofia, Bulgaria (1998) 3. Hofstadter, D.: The copycat project: An experiment in nondeterminism and creative analogies. AI Memo 755, Artificial Intelligence Laboratory, Massachusetts Institute of Technology (1984) 4. Li, M., Vitanyi, P.M.: An Introduction to Kolmogorov Complexity and Its Appli- cations. Springer Publishing Company, Incorporated, 3 edn. (2008) 5. Murena, P., Cornuéjols, A.: Minimum description length principle applied to struc- ture adaptation for classification under concept drift. In: 2016 International Joint Conference on Neural Networks, IJCNN 2016, Vancouver, BC, Canada, July 24-29, 2016. pp. 2842–2849 (2016) 6. Murena, P.A., Cornuéjols, A., Dessalles, J.L.: Incremental learning with the min- imum description length principle, accepted in International Joint Conference on Neural Networks 2017 7. Murena, P.A., Dessalles, J.L., Cornuéjols, A.: A complexity based approach for solv- ing hofstadter’s analogies, accepted at Computational Analogy Workshop, ICCBR 2017 8. Murena, P.A., Sublime, J., Matei, B., Cornuéjols, A.: Multi-view clustering with the principle of minimum description length, submitted to ECML 2017 9. Sublime, J., Matei, B., Murena, P.A.: Analysis of the influence of diversity in col- laborative and multi-view clustering, accepted in International Joint Conference on Neural Networks 2017