=Paper= {{Paper |id=Vol-1436/Paper75 |storemode=property |title=Dynamic Music Emotion Recognition Using Kernel Bayes’ Filter |pdfUrl=https://ceur-ws.org/Vol-1436/Paper75.pdf |volume=Vol-1436 |dblpUrl=https://dblp.org/rec/conf/mediaeval/MarkovM15 }} ==Dynamic Music Emotion Recognition Using Kernel Bayes’ Filter== https://ceur-ws.org/Vol-1436/Paper75.pdf
 Dynamic Music Emotion Recognition Using Kernel Bayes’
                        Filter

                       Konstantin Markov                                           Tomoko Matsui
                    Human Interface Laboratory                           Department of Statistical Modeling
                      The University of Aizu                             Institute of Statistical Mathematics
                       Fukushima, Japan                                              Tokyo, Japan
                     markov@u-aizu.ac.jp                                         tmatsui@ism.ac.jp

ABSTRACT                                                          performs poorly when the data relationship is non-linear.
This paper describes the temporal music emotion recogni-             Non-parametric non-linear kernel models [4] are becom-
tion system developed at the University of Aizu for the Emo-      ing more and more popular in the Machine Learning com-
tion in Music task of the MediaEval 2015 benchmark eval-          munity for their ability to learn highly non-linear mappings
uation campaign. The arousal-valence trajectory prediction        between two continuous data spaces. They extend the con-
is cast as a time series filtering task and is performed using    ventional kernel data mapping into high dimensional spaces
state-space models. A simple and widely used example is           to embedding data distributions in such spaces. This allows
the Kalman Filter, however, it is a linear parametric model       for Bayesian reasoning and developing inference algorithms
and has serious limitations. On the other hand, non-linear        which, however, involve only Gram matrices manipulations.
and non-parametric approaches don’t have such drawbacks,          Although, the complexity is O(n3 ) because of matrix inver-
but often scale poorly with the number of training data and       sions, it does not depend on data dimensionality, which is a
their dimension. One such method proposed recently is the         big advantage compared to other non-linear methods based
Kernel Bayes’ Filter (KBF). It uses only data Gram matrices       on Monte Carlo sampling [3].
and thus works (almost) equally well with data of both low           The task and the database used in this evaluation are
and high dimension. In our experiments, we used the feature       described in detail in the task overview paper [1].
set provided by the organizers without any change. All the
development data were clustered in six clusters based on the      2.    KERNEL BAYES’ FILTER
genre information available from the meta-data. For perfor-         Details about the Kernel Baeys Filter can be found in [2].
mance comparison, we build three more emotion recogni-            Here we provide just the basic notation and the final update
tion systems based on the standard Multivariate Linear re-        rules. During the KBF training truth values of both obser-
gression (MLR), Support Vector machine regression (SVR)           vations X = {x1 , . . . , xT } and corresponding state values
and Kalman Filter (KF). The results obtained from a 4-fold        Y = {y1 , . . . , yT } are required. The prediction and con-
cross-validation on the development set show that all types       ditioning steps of the standard filtering algorithms can be
of models, except KF, achieved very similar performance,          reformulated with respect to the kernel embeddings. The
which suggests that they may have reached the upper bound         embedding of the predictive distribution p(xt |y1:t ) is de-
of the feature set discrimination power.                          noted as µxt |y1:t and is estimated as
                                                                                                           PT
                                                                                                             i=1 αi φ(xi ), where
                                                                  φ() is the feature map and αt is updated recursively using
1.   INTRODUCTION
                                                                       Dt+1    = diag((G + λI)−1 G̃αt ),
   Dynamic or continuous emotion estimation is more diffi-
cult task and there are several approaches to solve it. The            αt+1    = Dt+1 K((Dt+1 K)2 + βI)−1 )Dt+1 K:xt+1 (1)
simplest one is to assume that for a relatively short period of   Here, G and K are the training states and observations
time emotion is constant and apply static emotion recogni-        Gram matrices, G̃ is a Gram matrix with entries Gij =
tion methods. These include conventional regression meth-         k(xi , xj+1 ), and K:xt+1 = (k(x1 , xt+1 ), . . . , k(xT , xt+1 )). The
ods as well as a combination of classification and regression     regularization parameters λ and β are needed to avoid nu-
where data are clustered in advance and for each cluster a        merical problems during matrix inversion.
separate regression model is built. Testing involves initial         There are few kernel functions that can be used with the
classification step or model selection procedure. A better        KBF such as linear, rbf, and polynomial. Their parameters
approach is to consider emotion trajectory as a time vary-        as well as the regularization constants λ and β comprise the
ing process and try to track it or use time series modelling      set of hyper-parameters of a KBF system. Unfortunately,
techniques involving state-space models (SSM). A popular          there is no algorithm for learning those hyper-parameters
and simple SSM is the Kalman filter (KF). It is a linear          from data. They have to be set manually and as our experi-
system and is quite fast since it requires just matrix multi-     ments showed are critical for obtaining a good performance.
plications and its complexity is linear in the number of data.
However, the linearity assumption is a big drawback and KF
                                                                  3.    EXPERIMENTS
                                                                    Using the genre information available from the metadata,
Copyright is held by the author/owner(s).
MediaEval 2015 Workshop, September 14-15, 2015, Wurzen, Ger-      we divided all development clips into six clusters roughly
many                                                              corresponding to the following genres:Classical, Electronic,
                                                                  Table 1: Kernel Bayes’ filter results on the develop-
                                                                  ment set.
                                                                          Genre          Arousal          Valence
                                                                                       R      RMSE       R    RMSE
                                                                         Classical   0.282     0.355   0.132   0.390
                                                                        Electronic   0.306     0.347   0.265   0.355
                                                                        Jazz-Blues   0.367     0.357   0.192   0.372
                                                                        Rock-Pop     0.350     0.342   0.167   0.382
                                                                      International  0.219     0.365   0.207   0.371
                                                                    Hip-Hop, SoulRB 0.307      0.342   0.204   0.348
                                                                         Average     0.305     0.351   0.194   0.369


                                                                  Table 3. Due to time limitations, the KBF system uses re-
                                                                  duced (to one forth) training set data which apparently has
                                                                  negative effect on the performance. Since the reference and
                                                                  predicted AV values are scaled to [-1.0, 1.0], direct compar-
Figure 1: Mean A-V values distribution for the de-                ison of the RMSE scores with those from previous tables is
velopment data. Different colors represent differ-                possible when they are divided by 2.
ent genres. Circle sizes are proportional to the A-V
standard deviation.
                                                                  Table 2: Averaged results of all systems on the de-
                                                                  velopment set.
Jazz-Blues, Rock-Pop, International-Folk, HipHop-SoulRB.                  Genre        Arousal         Valence
The number of clusters was chosen such that the data dis-                             R     RMSE      R    RMSE
tribution becomes as uniform as possible.                                             Regression
   In order to visualize the relationship between clustered               Linear    0.269    0.341  0.184   0.357
clips and their emotional content, we calculated arousal and              SVM       0.283    0.340  0.214   0.351
valence statistics per clip and Figure 1 shows the distribution                         Filters
of mean AV vectors in the affect space. Different colors rep-            Kalman     0.113    0.390  0.068   0.393
resent different genres/clusters and the circle size is propor-       Kernel Bayes’ 0.305    0.351  0.194   0.369
tional to the AV standard deviation. As can be seen, there
are no clear grouping by genre, though some genres show
more compact clouds than others. Both filtering systems,
i.e. KF and KBF, were build using this clustering scheme               Table 3: Averaged results on the test set.
where one model was trained for one genre and tested with                Genre     Arousal         Valence
the test data from the same genre only. Linear regression                         R    RMSE      R     RMSE
and SVR based systems were trained with no regard to genre                SVR 0.490    0.446   -0.019   0.542
clusters.                                                                KBR 0.419     0.498   -0.035   0.620
   Since there is no validation data set available, we used 4-
fold cross-validation approach to tune systems’ parameters.
The SVR and KBF models have hyper-parameters such as
kernel function and regularization constants which cannot be      5.   CONCLUSIONS
learned from data. An unconstrained simplex search method            We described several systems developed at the University
was adopted to find optimum parameter setting, however, it        of Aizu for the MediaEval’2015 Emotion in Music evalua-
does not guarantee global maximum and in the case of KBF,         tion campaign. Our focus is on the machine learning part of
it turned out the initial point has a big impact on the final     this very challenging task and, thus, we built and evaluated
result.                                                           few systems based on conventional regression techniques as
                                                                  well as on a new non-parametric non-linear approach us-
                                                                  ing Kernel Bayes’ Filter state-space system. All used the
4.   RESULTS                                                      feature set provided by the challenge organizers. Although
  Before the calculation of the correlation and RMSE per-         the modelling techniques we utilized range from simple lin-
formance measures, predicted arousal and valence values as        ear regression to sophisticated state-space Bayesian filter,
well as the reference values were scaled to fit the range [-      there was a negligible difference in the performance. This
0.5,+0.5]. This similar to the way results were obtained          suggests that the feature set may not have enough discrimi-
during previous evaluations.                                      nating power to enable non-parametric non-linear models to
  Table 1 shows the performance of the KBF for each genre         show their advantages.
as well as the total average. For some genres, the results are
better, which may be due to differences in data distributions,    6.   ACKNOWLEDGEMENT
but also because of a better hyper-parameter settings. Total        Authors would like to thank Dr. Y.Nishiyama from the
averages from all the regression and state-space model based      University of Electro-Communications, Tokyo, for sharing
systems are summarised in Table 2.                                his Matlab Kernel mean toolbox (kmtb).
  The results using the official test data set are shown in
7.   REFERENCES
[1] A. Aljanaki, Y.-H. Yang, and M. Soleymani. Emotion in music
    task at mediaeval 2015. In Working Notes Proceedings of
    the MediaEval 2015 Workshop, September 2015.
[2] K. Fukumizu, L. Song, and A. Gretton. Kernel bayes’ rule:
    Bayesian inference with positive definite kernels. J. Mach.
    Learn. Res., 14(1):3753–3783, Dec. 2013.
[3] K. Markov, T. Matrui, F. Septier, and G. Peters. Dynamic
    speech emotion recognition with state-space models. In Proc.
    EUSIPCO’2015, pages 2122–2126, 2015.
[4] L. Song, K. Fukumizu, and A. Gretton. Kernel embeddings of
    conditional distributions: A unified kernel framework for
    nonparametric inference in graphical models. Signal
    Processing Magazine, IEEE, 30(4):98–111, July 2013.