=Paper=
{{Paper
|id=Vol-2540/article_1
|storemode=property
|title=Towards a visual framework for the incorporation of knowledge in the phases of machine learning
|pdfUrl=https://ceur-ws.org/Vol-2540/FAIR2019_paper_32.pdf
|volume=Vol-2540
|authors=CJ Swanepoel,KM Malan
|dblpUrl=https://dblp.org/rec/conf/fair2/SwanepoelM19
}}
==Towards a visual framework for the incorporation of knowledge in the phases of machine learning==
Towards a visual framework for the incorporation of knowledge in the phases of machine learning CJ Swanepoel and KM Malan Department of Decision Sciences, University of South Africa. swanecj@unisa.ac.za Abstract. Incorporating domain knowledge into machine learning al- gorithms to some extent is almost unavoidable. Doing it well and ex- plicitly can avoid unnecessary bias, improve efficiency and accuracy, and increase transparency. To increase an awareness of the relative contri- butions of domain knowledge and machine learning expertise, as well as an indication of the direction of information flow, a tentative qualitative visualisation framework is suggested, and two examples are given. It is hoped that such a mechanism will encourage reflection on the (sometimes implicit and innate) inclusion of domain knowledge in machine learning systems. Keywords: Domain knowledge · Machine learning · Characterisation framework · Visualisation. 1 Introduction Machine learning as a component of artificial intelligence, and especially deep learning, has experienced phenomenal growth over the last couple of years. (For example, in 1998 there were two papers with primary subcategory ‘Machine learning’ on arXiv, and in 2017 there were 2 332 papers [19].) This growth is the result of the advances in computing power, especially through the use of graphics processing units and more recently tensor processing units, the ubiquity of available data, the connectivity afforded by the internet, and major advances in machine learning algorithms by Bengio, Hinton and LeCun, amongst others [14]. In well defined domains where data is plentiful or cheap to generate and where the context is stable, machine learning can be an extremely useful tool. However, many machine learning techniques (and especially deep learning) are fragile, greedy, shallow and not transparent [8, 12, 27]. The techniques are fragile because transfer to a slightly different domain or context usually breaks them; greedy because they require huge amounts of (labelled) data; shallow because they depend on superficial features and do not possess an underlying model based on the physical reality; and not transparent because the internal structure is often too complex to analyse and connect to the features that determine the output. This necessitates careful attention to subtle aspects of the machine learning development process. Copyright © 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0) 2 CJ Swanepoel, KM Malan This paper focusses on the inclusion of domain knowledge in machine learning, because this is one factor that can affect the fragility, greediness, shallowness and lack of transparency of machine learning. The aim is to provide a frame- work to assist practitioners to make explicit the different ways in which domain knowledge and machine learning expertise are incorporated in the stages of ma- chine learning. Being aware of subtle inclusions of domain knowledge and innate knowledge endowed to the machine learning system might assist in identifying opportunities to improve its performance, accuracy or even transparency, and can contribute to more accurate reporting of machine learning system designs. 2 The inclusion of domain knowledge The recent successes of machine learning that depend on data only, and use ‘no domain specific’ data (e.g. AlphaZero) create the impression that domain specific knowledge is not really necessary in machine learning, and might even reduce the ‘generality’ of the resulting system. The practical difficulties in finding useful representations of expert knowledge, and the fact that domain knowledge is often not complete or perfect, reinforces this view [27]. Two seemingly diametrically opposed views are expressed in recent literature [15, 13]: on the one hand, the view that all problems can be solved by scaling the model up and rely on the data only (AlphaZero, for example), and on the other hand, the belief that using a combination of data and domain knowledge will eventually prove to be the best approach (an idea already propounded by Alan Turing in his 1950 paper [23]). The ‘data only’ camp demonstrated spectacular results, particularly in the domain of game play and text generation, although the fear exists that it will not generalise easily (mainly due to data constraints in most domains) and that the performance will hit a ceiling. Even when domain knowledge is not explicitly injected (the ‘data only’ ap- proach), implicit domain knowledge almost always features in machine learning [9]. This implicit or innate domain knowledge can include the choice and struc- ture of the algorithm, representational formats, and innate knowledge or expe- rience [13]. For example, the choice of data encoding method depends partly on an understanding of the problem domain, and can greatly influence the effective- ness of the algorithm (see, e.g. [7, 18]). Feature selection is often task dependent and sometimes even based on intuition [5, 26]. The type of algorithm used also depends heavily on the nature of the problem domain. Marcus [13] quotes Pe- dro Domingos: “[Machine learning] paradigms differ in what assumptions they encode, and what form of additional knowledge they make it easy to encode.” The structure of a deep learning neural network is influenced by the nature of the problem. In the description of the neural network architecture for AlphaGo Zero, where emphasis was placed on using as little as possible explicit domain knowledge, it is stated that “History features Xt , Yt are necessary, because Go is not fully observable solely from the current stones. . . ” [21]. Further examples of Towards a visual framework for the incorporation of knowledge in ML 3 the extensive embedding of human domain specific knowledge in the construction of AlphaGo Zero and AlphaZero are given in [13]. In a genetic algorithm the choice of cross-over and mutation mechanisms will to some extent depend on implicit domain knowledge [10]. Risk or loss functions include implicit domain knowledge, but can also encode selected prior knowledge [15, 22]. Constraining the output of a machine learning system based on heuristics or rules from the problem domain is common practice. (For example, all Go playing algorithms before AlphaGo Zero routinely removed stupid (but legal) moves [21].) In most (non-game) domains there are additional considerations for including domain knowledge, such as the fact that data can be difficult to get, or expensive, or might include bias, or be unbalanced (for example, the lack of edge cases). It is often necessary to explicitly include additional information to get to a feasible, unbiased or useful solution [4, 6]. The explicit inclusion of domain knowledge can potentially also improve the transparency of the model [28]. Conversely, including less explicit domain knowledge might lead to more general algorithms. There are some obstacles to the inclusion of domain knowledge, though. These include the difficulty of finding workable encodings and injection mechanisms, the fact that most domain experts are not data science experts and vice versa [27]. The restrictions introduced by injecting domain knowledge can potentially also prevent the discovery of valid but unexpected solutions [3]. The explicit integration of knowledge into machine learning is called ‘informed machine learning’ by von Rueden et al. [25]. They developed a taxonomy for the explicit integration of knowledge. Their proposed taxonomy contains three components: the type of knowledge, the method used to integrate the knowledge into the machine learning system, and lastly where in the machine learning pipeline the integration happens. It is a useful tool to classify papers in the assisted or informed machine learning domain. However, the implicit inclusion of domain knowledge in the form of innate knowl- edge and convention or experience is often not recognised and neglected in the reporting on and meta-analysis of machine learning algorithms. Many of the choices made throughout all phases of the machine learning process are based on some understanding of aspects of the problem or task, augmented by ex- pertise and experience in the machine learning domain. This paper attempts to provide a tentative high level framework to visually characterise the inclusion of any domain knowledge into machine learning. 3 Proposed framework In Figure 1 the generic phases of the machine learning pipeline are given as coloured blocks labelled with a capital letter. A very brief description of the phases is given in Table 1, and a more comprehensive account is given in Sec- tion 4. 4 CJ Swanepoel, KM Malan Although the phases are listed more or less in the order in which they occur in the machine learning pipeline, it is an idealised representation that does not necessarily reflect the actual workflow of a specific system – in practice the order in which the phases are executed can be convoluted, and might include several iterative loops. The colours group the phases into six clusters: problem for- mulation (green), data preparation (yellow), machine learning activities (blue), output constraints (red), interpretation and explanation (brown), and external comparison (green). A subjective qualitative estimate of the magnitude of the contribution from re- spectively domain knowledge and machine learning is represented by the relative thicknesses of the arrows in the figure. In cases where the emphasis is on the exploitation of data only, many of the arrows from domain knowledge will be thin or have zero thickness, for example. The direction of the arrows indicates the direction of information flow. Such a representation will be unique to every particular instance of a machine learning system, and can give a quick visual overview of the nature of the interactions in that instance. The representation is subjective, and as such cannot provide accurate or quanti- tative data. However, this simplified model allows a quick high-level evaluation of the relative contributions from the two knowledge domains. A: Problem identification and formula- H: Machine learning algorithm structure tion determination B: Data sourcing/labelling I: Learning process mechanisms C: Data cleaning/validation/quality eval- J: Hyperparameter tuning uation K: Constraining outputs D: Data augmentation L: Interpretation/validation E: Data encoding M: Explanation F: Machine learning algorithm selection N: Comparative evaluation G: Feature engineering Table 1. Phases of machine learning Domain Knowledge A B C D E F G H I J K L M N Machine Learning Expertise Fig. 1. The contribution of domain knowledge and machine learning expertise to a typical machine learning workflow Towards a visual framework for the incorporation of knowledge in ML 5 4 Phases of the machine learning pipeline A: Problem identification and formulation Thorough knowledge of the do- main is necessary for this step. The context will determine what data can be collected, what the objective(s) is (are), and what information might be available that is not included in the data. B: Data sourcing/labelling Although the machine learning process often starts with available data, in some cases data will have to be sourced or labelled. Domain knowledge in the form of deep knowledge of the relationships be- tween different features, the nature of the data, the difficulty and cost of labelling as well as an understanding of machine learning algorithms and how the data will be used can contribute to the quality of the data that is eventually used, and hence influence the outcome or success of the machine learning process. C: Data cleaning/validation/quality evaluation Knowledge of the prop- erties of the domain and the data collection methodology can assist in iden- tifying outliers or invalid data points, and allow for an evaluation of the quality of the data. It will also give insight into the coverage of the data space (whether there are areas where data is too sparse to be useful). D: Data augmentation Both knowledge of the domain and the machine learn- ing algorithm is required for the successful augmentation of data. The impu- tation of missing data points, for example, can take various forms, and some of the techniques might be counterproductive in the training of a model. Another example: the perturbation of images by shifting a few pixels hori- zontally or vertically to provide additional training data assumes an under- standing that such a translation will indeed provide novel information to the machine learning system, while it remains valid as a representation and does not influence the labelling of the image. E: Data encoding Mainly machine learning expertise is required. However, a deep understanding of the domain knowledge environment is assumed. For example, one-hot encoding can be difficult or impossible to work with when the feature space is very large. Piech et al. [17] address this data encoding challenge by utilising a random low-dimensional representation of a one-hot high-dimensional vector. This encoding is motivated by the idea of compressed sensing introduced by Baraniuk in 2007 [1] as an effective method to capture and represent compressible signals at a rate significantly below the Nyquist rate. Another beautiful example is given in the paper by Lusci et al. [11] where molecules are represented as ensembles of directed acyclic graphs as input to a recursive neural network to predict the solubility of these molecules. F: Machine learning algorithm selection This deals with the choice of a suitable machine learning algorithm. A deep understanding of the working of different machine learning techniques is required, as well as an understanding 6 CJ Swanepoel, KM Malan of the nature of the problem domain. For example, if the problem has a temporal component, a recurrent neural network might be a good fit, or if filtering is required, an auto-encoder should be considered. Also see the paper by Olson et al. [16] where thirteen machine learning algorithms are compared over 165 publicly available classification problems. G: Feature engineering Feature engineering includes binning, transformation of features, scaling, grouping operations and feature selection. Domain knowl- edge in the form of an understanding of the relationships between features and the information content of different features are required. H: Machine learning algorithm structure determination This is sometimes described as more of an art than a science. Typically the structure of the machine learning algorithm is determined empirically through experimenta- tion. Here experience with similar or related problems, which is a form of domain knowledge, plays a huge role. For example, Chandrasekaran et al. [2] in a recent paper designed a machine learning predictor for density of state and charge density of a material or molecule. For the charge density, the modelling is done with a simple fully connected neural network with one output neuron. The local density of states spectrum, on the other hand, is modelled with a recurrent neural network, where the local density of state at every energy window is represented as a single output neuron (linked via a recurrent layer to other neighbouring energy windows). Domain knowledge therefore played a huge part in determining the structure of the machine learning system. I: Learning process mechanisms (transfer function / learning mechanism / mutation operator, etc. selection) This is probably the area where there is the biggest opportunity of innovation. Novel functions for cross-over, or innovative transfer functions can hugely influence the way in which the search space is traversed. J: Hyperparameter tuning As with stage H, this is an area that is mostly approached empirically. The performance metric used in the hyperparameter optimisation process is influenced by domain knowledge. Setting ranges for grid or random searches to optimise hyperparameters, as well as determining of which of the hyperparameters should be included in the search cannot be analytically determined, but knowledge of the hyperparameter behaviour in related problems can provide a good starting point for an empirical search strategy. K: Constraining outputs This is one of the most important mechanisms to include domain knowledge in the machine learning process. Techniques used include the augmentation or restriction of the loss or risk function, and filtering or transforming intermediate outputs or the final output. Stewart and Ermon [22], for example, use laws from physics to constrain the output space in training a convolutional neural network to track objects without using any labelled examples. Towards a visual framework for the incorporation of knowledge in ML 7 L: Interpretation / validation A good understanding of the underlying knowl- edge domain will allow an evaluation of the feasibility or quality of the out- comes. In simple classification problems this is not really an issue, but for many decision support applications this is essential. The output of a machine learning algorithm might also increase our understanding of the domain, hence the possibility of a two-way arrow in the proposed framework. M: Explanation One of the biggest criticisms against many machine learning techniques is the lack of transparency, or the ability to explain the output of the machine learning system. Legal and moral requirements dictate that in certain environments it should be possible to justify the outcomes in the light of the inputs. Here an understanding of the domain complexities as well as the machine learning mechanism is required – although this might not be sufficient in many cases. In the proposed framework the direction of the arrow towards the domain knowledge bar indicates the flow of information that might add to our understanding of the domain. In addition, an analysis of how the machine learning process obtained the outputs could add to machine learning expertise. N: Comparative evaluation Comparing the performance of a machine learn- ing algorithm against previously applied approaches (benchmarking) is often necessary to evaluate the performance of a new approach. This also con- tributes to the knowledge base of machine learning. 5 Case studies/examples In this section two recent contributions to the machine learning environment are described briefly, and the proposed visual representations of knowledge are given for both. 5.1 Combination of domain knowledge and deep learning for sentiment analysis, by Vo et al. In the paper ‘Combination of domain knowledge and deep learning for sentiment analysis’, published in 2017, Vo et al. [24] found that existing approaches in the application of machine learning to sentiment analysis suffer from two major drawbacks, the first of which is that until then nobody has paid attention to the different types of sentiment terms. Different domains use different terms to express positive and negative sentiments, and some words carry a higher emotive content than others. Secondly, the loss functions used previously did not include a measure of the magnitude of sentiment misclassification, and did not distinguish between different types of misclassification. To address these two issues, they proposed using sentiment scores (learnt by quadratic programming) to augment training data; and introduced a penalty 8 CJ Swanepoel, KM Malan matrix to enhance the loss function. The enhancements were applied to a stan- dard sentiment analysis workflow using a convolutional neural network. To eval- uate the success of their approach, they compared the performance of the new system with a baseline convolution neural network sentiment analyser, as well as with a traditional support vector machine based sentiment analyser. In the com- parative analysis the new approach performs better than the previous versions, showing that the inclusion of the two enhancements are useful in this domain. This knowledge and the description of the novel enhancements led to the dia- gram in Figure 2, where the incorporation of additional domain knowledge in the ‘data augmentation’ (D) and ‘constraining outputs’ (K) phases is emphasised. Domain Knowledge A B C D E F G H I J K L M N Machine Learning Expertise Fig. 2. The contribution of domain knowledge and machine learning expertise to the sentiment analyser of [24] 5.2 Mastering the game of Go without human knowledge, by Silver et al. During 2016 and 2017 DeepMind released three versions of their Go playing software. The first system, now called AlphaGo Lee, defeated the world champion Lee Seedol 4–1 in a five game match in March 2016. AlphaGo Lee used two neural networks – a ‘policy’ and a ‘value’ network, as well as a Monte Carlo tree search algorithm. It was trained using historic game information, and improved through self play. A second version, AlphaGo Zero, was introduced in a paper in Nature on 19 October 2017 [21]. The title of the paper: ‘Mastering the game of Go without human knowledge’ expresses the major claim of this version – that it used no human or domain knowledge except for the rules of the game. A third version of the software (AlphaZero) was introduced in a paper published on arXiv in December 2017 [20]. This version added chess and shogi to the repertoire of games mastered by the system. It was claimed that in this progression the newer version each time learnt quicker and exceeded the performance of its predecessor. The contribution of knowledge into the second version, AlphaGo Zero, is discussed in this section. Towards a visual framework for the incorporation of knowledge in ML 9 In [13] Marcus extensively discusses different inclusions (some not mentioned in the original paper) of domain knowledge into the AlphaGo Zero machine learning system. He argues that the use of carefully constructed Monte Carlo tree search machinery, the artful placement of convolutional layers that allow the system to recognise that many patterns on the board are translation invariant, and the application of a sampling algorithm for dealing with reflections and rotations constitute the injection of domain knowledge into the system. Based mostly on Marcus’s assessment, rough qualitative judgements on the in- clusion of domain knowledge into AlphaGo Zero were made that are reflected in Figure 3. Domain Knowledge A B C D E F G H I J K L M N Machine Learning Expertise Fig. 3. The contribution of domain knowledge and machine learning expertise to Al- phaGo Zero The lack of explicit contributions from domain knowledge in the phase of ‘con- straining outputs’ (K) is indicative of the fact that this popular method of inject- ing domain knowledge was not used. There are, however, substantial contribu- tions in phases F and H, the selection of the machine learning algorithm (a combi- nation of Monte Carlo search trees and reinforcement learning) and the structure determination of the algorithm (taking into account symmetries, for example). The direction of information flow in the three phases ‘interpretation/validation’ (L), ‘explanation’ (M) and ‘comparative evaluation’ (N) should also be noted. Domain knowledge and machine learning expertise are expanded to varying de- grees with a successful implementation of a machine learning system. In the case of AlphaGo Zero new (‘alien’) gameplay strategies were discovered (do- main knowledge), and some insight gained into the comparative performance of different approaches (machine learning expertise). 10 CJ Swanepoel, KM Malan 6 Conclusion A visualisation scheme for the inclusion of domain knowledge and machine learn- ing expertise into machine learning systems is proposed. The hope is that it will aid in greater awareness of the innate, implicit and explicit use of domain knowl- edge in the machine learning workflow. References 1. Baraniuk, R.: Compressive sensing [lecture notes]. IEEE Signal Processing Maga- zine 24(4), 118–121 (Jul 2007). https://doi.org/10.1109/msp.2007.4286571 2. Chandrasekaran, A., Kamal, D., Batra, R., Kim, C., Chen, L., Ramprasad, R.: Solving the electronic structure problem with machine learning. npj Computational Materials 5(1) (Feb 2019). https://doi.org/10.1038/s41524-019-0162-7 3. Childs, C.M., Washburn, N.R.: Embedding domain knowledge for machine learn- ing of complex material systems. MRS Communications pp. 1–15 (Jul 2019). https://doi.org/10.1557/mrc.2019.90 4. Choo, J., Liu, S.: Visual analytics for explainable deep learning. IEEE Computer Graphics and Applications 38(4), 84–92 (Jul 2018). https://doi.org/10.1109/mcg.2018.042731661 5. Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.P.: Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12, 2493– 2537 (2011), http://dl.acm.org/citation.cfm?id=2078186 6. Du, M., Liu, N., Hu, X.: Techniques for interpretable machine learning. arXiv:1808.00033 (2018) 7. Geman, S., Bienenstock, E., Doursat, R.: Neural networks and the bias/variance dilemma. Neural Computation 4(1), 1–58 (Jan 1992). https://doi.org/10.1162/neco.1992.4.1.1 8. Ghorbani, A., Abid, A., Zou, J.: Interpretation of Neural Networks is Fragile. arXiv:1710.10547 arXiv:1710.10547 (Oct 2017) 9. Hessel, M., van Hasselt, H., Modayil, J., Silver, D.: On inductive biases in deep reinforcement learning. arXiv:1907.02908 (2019) 10. Johns, M.B., Mahmoud, H.A., Walker, D.J., Ross, N.D.F., Keedwell, E.C., Savic, D.A.: Augmented evolutionary intelligence. In: Proceedings of the Genetic and Evolutionary Computation Conference on – GECCO '19. ACM Press (2019). https://doi.org/10.1145/3321707.3321814 11. Lusci, A., Pollastri, G., Baldi, P.: Deep architectures and deep learning in chemoinformatics: The prediction of aqueous solubility for drug-like molecules. Journal of Chemical Information and Modeling 53(7), 1563–1575 (Jul 2013). https://doi.org/10.1021/ci400187y 12. Marcus, G.: Deep learning: A critical appraisal. arXiv:1801.00631 (2018) 13. Marcus, G.: Innateness, AlphaZero, and artificial intelligence. arXiv:1801.05667 (2018) 14. Marx, V.: Machine learning, practically speaking. Nature Methods 16(6), 463–467 (May 2019). https://doi.org/10.1038/s41592-019-0432-9 15. Muralidhar, N., Islam, M.R., Marwah, M., Karpatne, A., Ramakrishnan, N.: Incorporating prior domain knowledge into deep neural networks. In: 2018 IEEE International Conference on Big Data (Big Data). pp. 36–45 (Dec 2018). https://doi.org/10.1109/BigData.2018.8621955 Towards a visual framework for the incorporation of knowledge in ML 11 16. Olson, R.S., La Cava, W., Mustahsan, Z., Varik, A., Moore, J.H.: Data-driven Ad- vice for Applying Machine Learning to Bioinformatics Problems. arXiv:1708.05070 arXiv:1708.05070 (Aug 2017) 17. Piech, C., Spencer, J., Huang, J., Ganguli, S., Sahami, M., Guibas, L.J., Sohl- Dickstein, J.: Deep knowledge tracing. arXiv:1506.05908 (2015) 18. Potdar, K., Pardawala, T.S., Pai, C.D.: A comparative study of cat- egorical variable encoding techniques for neural network classifiers. In- ternational Journal of Computer Applications 175(4), 7–9 (Oct 2017). https://doi.org/10.5120/ijca2017915495 19. Shoham, Y., Perrault, R., Brynjolfsson, E., Clark, J., Manyika, J., Niebles, J.C., Lyons, T., Etchemendy, J., Grosz, B., Bauer, Z.: The AI Index 2018 Annual Report. AI Index Steering Committee, Human-Centered AI Initiative, Stanford University, Stanford, CA (Dec 2018) 20. Silver, D., Hubert, T., Schrittwieser, J., Antonoglou, I., Lai, M., Guez, A., Lanctot, M., Sifre, L., Kumaran, D., Graepel, T., Lillicrap, T.P., Simonyan, K., Hassabis, D.: Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv:1712.01815 (2017) 21. Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., Chen, Y., Lillicrap, T., Hui, F., Sifre, L., van den Driessche, G., Graepel, T., Hassabis, D.: Mastering the game of Go without human knowledge. Nature 550(7676), 354–359 (Oct 2017). https://doi.org/10.1038/nature24270 22. Stewart, R., Ermon, S.: Label-free supervision of neural networks with physics and domain knowledge. arXiv:1609.05566 (2016) 23. Turing, A.M.: I.—Computing Machinery and Intelligence. Mind LIX(236), 433– 460 (Oct 1950). https://doi.org/10.1093/mind/lix.236.433 24. Vo, K., Pham, D., Nguyen, M., Mai, T., Quan, T.: Combination of do- main knowledge and deep learning for sentiment analysis. In: Lecture Notes in Computer Science, pp. 162–173. Springer International Publishing (2017). https://doi.org/10.1007/978-3-319-69456-6_14 25. von Rueden, L., Mayer, S., Garcke, J., Bauckhage, C., Schuecker, J.: Informed Machine Learning – Towards a Taxonomy of Explicit Integration of Knowledge into Machine Learning. arXiv:1903.12394 (Mar 2019) 26. Yang, L., Zheng, Z., Sun, J., Wang, D., Li, X.: A domain-assisted data driven model for thermal comfort prediction in buildings. In: Proceedings of the Ninth International Conference on Future Energy Systems. pp. 271–276. e-Energy ’18, ACM, New York, NY, USA (2018). https://doi.org/10.1145/3208903.3208914 27. Yu, T., Jan, T., Simoff, S., Debenham, J.: Incorporating prior domain knowledge into inductive machine learning. Technical report, International Institute of Fore- casters (IIF), University of Massachusetts, Amherst, USA (Oct 2006) 28. Yu, T., Simoff, S., Jan, T.: VQSVM: A case study for incorporating prior domain knowledge into inductive machine learning. Neurocomputing 73(13-15), 2614–2623 (Aug 2010). https://doi.org/10.1016/j.neucom.2010.05.007