<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Joint Conference (March</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Structural Change Point Detection Using A Large Random Matrix and Sparse Modeling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Katsuya Ito∗</string-name>
          <email>katsuya1ito@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Akira Kinoshita</string-name>
          <email>kino@preferred.jp</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Masashi Yoshikawa</string-name>
          <email>yoshikawa@preferred.jp</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Graduate School of Economics, The University of Tokyo</institution>
          ,
          <addr-line>Bunkyo, Tokyo</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Preferred Networks, Inc.</institution>
          ,
          <addr-line>Chiyoda, Tokyo</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>26</volume>
      <issue>2019</issue>
      <abstract>
        <p>This paper proposes a new structural change point detection method for time series using our new matrix decomposition method. We propose YMN decomposition which decomposes one time-series data matrix into a large random matrix and a sparse matrix. YMN decomposition can obtain the information about the latent structure from the sparse matrix. Thus our structural change point detection method using YMN decomposition can detect the change in higher-order moments of a mixing matrix as well as typical structural changes such as the change in mean, variance, and autocorrelation of time series. We also partly theorize our methods using existing theories of random matrix and statistics. Our experiment using artificial data demonstrates the efectiveness of our change point detection techniques and our experiment using real data demonstrates that our methods can detect structural changes in economics and finance.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Unveiling the model of time-series data is necessary to explain
the causal relationships and to forecast the future. In economic
and financial modeling, the models of time-series data are not
apparent in many cases; moreover, their models are time varying
[
        <xref ref-type="bibr" rid="ref2 ref28 ref32 ref34">2, 28, 32, 34</xref>
        ]. Under such obscurity and instability of the models,
explaining the causal relationships among data and forecasting
the future are dificult but central problems in these fields [
        <xref ref-type="bibr" rid="ref14 ref16">14, 16</xref>
        ].
One straightforward method to overcome the instability is using
change-point detection methods. Change-point detection methods
divide piece-wise stable time-series data into several stable
timeseries data [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ]. Moreover, detecting the change point of the
structure is an essential task in time-series modeling in economics
and finance [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Research with a similar purpose can be found in
many fields such as biology [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], neuroscience [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ], and computer
network analysis [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Structural change point detection is dificult because we do
not know the structure of the time-series before/after the change
point, and how the change occurs. However, factor modeling is
typically used to unveil the partial structures in economics [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
and finance [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ]. Economic or financial data such as GDPs and
stock prices are observed as high-dimensional time series data.
Assume that we observe a vector yt at time t . A factor model
assumes that a few typical “factors” generate the observed data.
It formally introduces two latent variables A and zt and assumes
that the linear equation holds:
      </p>
      <p>yt = Azt .
∗This work was performed when the author was an intern at Preferred Networks,
Inc.</p>
      <sec id="sec-1-1">
        <title>Field</title>
      </sec>
      <sec id="sec-1-2">
        <title>Robotics</title>
      </sec>
      <sec id="sec-1-3">
        <title>Neuro</title>
        <p>science
yt (observed)</p>
      </sec>
      <sec id="sec-1-4">
        <title>Sensor data</title>
      </sec>
      <sec id="sec-1-5">
        <title>EEG/fMRI</title>
      </sec>
      <sec id="sec-1-6">
        <title>Network</title>
      </sec>
      <sec id="sec-1-7">
        <title>Trafic data</title>
      </sec>
      <sec id="sec-1-8">
        <title>Economics</title>
      </sec>
      <sec id="sec-1-9">
        <title>Macro data</title>
      </sec>
      <sec id="sec-1-10">
        <title>Finance</title>
      </sec>
      <sec id="sec-1-11">
        <title>Stock prices</title>
        <p>zt (latent)
Motion
of parts
Waves
from
parts</p>
      </sec>
      <sec id="sec-1-12">
        <title>Packets</title>
      </sec>
      <sec id="sec-1-13">
        <title>Factors</title>
      </sec>
      <sec id="sec-1-14">
        <title>Factors</title>
      </sec>
      <sec id="sec-1-15">
        <title>A (structure)</title>
        <p>Structure
of data</p>
      </sec>
      <sec id="sec-1-16">
        <title>Structure of brain</title>
      </sec>
      <sec id="sec-1-17">
        <title>Network</title>
        <p>
          Structure
Economic
Structure
Structure
of Market
zt is a low-dimensional time-variant vector of factors, while A is
a time-invariant coeficient matrix. A is called “factor loadings”
in the context of econometrics and “mixing matrix” in the context
of computer science. A implies the latent “structure” behind the
observed data. However, the structure may be changed abruptly
in reality because of major economic or political events such
as Brexit or the Global Financial Crisis. Most of the work
performed by economists and quantitative financial analysts, such
as the prediction and explanation of time-series, strongly rely on
their models [
          <xref ref-type="bibr" rid="ref19 ref37">19, 37</xref>
          ]. It is, therefore, crucial for them to detect
changes in their structures. As summarized in Table 1, the
structures (i.e. the relationships between observed and latent data) are
important, and similar types of problems occur in many fields.
        </p>
        <p>
          The change exhibits several patterns. The simplest one is the
change in the mean and variance of A’s entities; it can be detected
by merely calculating the mean and variance of y. The change
in the numbers of factors and the change in the autocorrelation
of y are also the essential but straightforward ones [
          <xref ref-type="bibr" rid="ref12 ref20 ref9">9, 12, 20</xref>
          ].
The third- and fourth-order moments are also significant in the
literature of finance and econometrics, which attach special
importance to them [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. However, they are dificult to be detected
because in high-dimensional cases, they approach the normal
distribution by the central limit theorem1 and the change cannot
be observed in y. To the best of our knowledge, nonparametric
methods that detect changes of such skewness and kurtosis do
not exist, whereas many parametric methods do exist. [
          <xref ref-type="bibr" rid="ref30 ref31 ref4">4, 30, 31</xref>
          ].
        </p>
        <p>
          Many previous studies showed the efectiveness of linear
models in detecting the structural changes of high-dimensional
data. Well-known methods such as principal component analysis
(PCA), non-negative matrix factorization (NMF), and
independent component analysis (ICA) can be interpreted as methods to
1This interpretation of the central limit theorem is often used in ICA literature
[
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. In this literature, skew and kurtosis are interpreted as non-Gaussianity and by
using many samples, the non-Gaussianity of data will decrease by the central limit
theorem.
decompose one data matrix into two matrices that satisfy specific
properties. Moreover, a variant of these methods that specializes
in change point detection is developed such as OMWRPCA [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ],
ILRMA [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ], and TVICA [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ].
        </p>
        <p>
          We herein propose a new non-parametric change point
detection method of the structures of multidimensional time series.
Our method decomposes the observed data matrix into a large
random matrix and a sparse coeficient matrix. First, we use
randomly generated large matrices as latent time series. However,
when a large random time series row vectors is generated, most
of its row vectors are not related to the latent time series, and
some are close to the latent time series. Therefore, we select good
row vectors by applying a sparse matrix to this. Finally, we can
detect the structural changes by calculating the diference of the
sparse matrix. Some previous works reported similar frameworks
such as latent feature Lasso [
          <xref ref-type="bibr" rid="ref42">42</xref>
          ] or extreme learning machine
[
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]; however, studies using both the random matrix and the
sparse matrix have not been reported.
        </p>
        <p>Moreover, we defined a new 2 type of change point, i.e., “the
third or fourth moment of factor loading changes”. Our
experiment showed that our method can detect these change points. Our
framework does not exert power in other tasks such as prediction
and factor analysis. Meanwhile, in the context of non-parametric
change point detection, it has a high expressive power.</p>
        <p>The main contributions of this paper are as follows.
• We propose YMN decomposition, which decompose one
time-series data matrix into a product of a random matrix
and a sparse matrix.
• We propose a structural change point detection method
using YMN decomposition.
• We demonstrate the efectiveness of our structural change
point detection techniques using experimental data and
real data.</p>
        <p>The remainder of the paper is organized as follows. In the
next section, we present the related work. In Section 3, we clearly
define the problems to solve. In Section 4, we propose the method.
The method is evaluated in Section 5. We discuss the experimental
results, issues, and future work in Section 6. Finally, we conclude
the paper in Section 7.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        This paper proposes a new structural change point detection
method using nonparametric matrix decomposition. In this
section, we review the previous research that is related to ours in
terms of (1) method and (2) purpose. We do not review the works
that are not directly related to ours. Therefore, we refer to the
survey by Truong et al. [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ] for a general review of change point
detection methods.
2.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Non-parametric change point detection</title>
      <p>As mentioned in the introduction, change point detection can
be divided into two types: (1) parametric and (2) nonparametric
change point detection.</p>
      <p>Because parametric change point detection places strong
assumptions on the distribution of time series, it exhibits strong
capability within the assumption of the model users, but any
unexpected change point cannot be detected. Meanwhile,
nonparametric change point detection methods place only a few
assumptions on the time-series distribution of data. Therefore,
2This concept is not new. However, no research in econometrics has defined this
type of change yet.
the model is free and can present a robust result; however, their
results often cannot be interpreted, and they can be less accurate
than parametric methods in some situations, e.g., the observed
time-series data completely follows one stochastic model.</p>
      <p>
        Nonparametric change point detection methods are
primarily divided into two types. One type uses matrix decomposition
and the other uses kernel methods. The kernel methods [
        <xref ref-type="bibr" rid="ref21 ref26">21, 26</xref>
        ]
demonstrate a high ability when handling data with nonlinear
models; however, they are similar to parametric methods in that
the results strongly depend on the choice of the kernel functions.
On the other hand, methods that fit each purpose and each data’s
property are developed in each field, such as Robust PCA [
        <xref ref-type="bibr" rid="ref41">41</xref>
        ],
NMF [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], and ICA [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. However, because a unique method to
decompose one matrix into two matrices does not exist, it is
necessary to place some assumptions on both matrices. These
assumptions significantly restrict the degree of freedom of the
model. PCA assumes no covariance on latent data,ICA assumes
independence on latent data. Hence, the method of matrix
decomposition used in each field difers significantly.
2.2
      </p>
    </sec>
    <sec id="sec-4">
      <title>Structural change point detection</title>
      <p>
        Next, we will explain the structural change point detection
methods. In the parametric method, it is possible to assume a graph
structure or a time-series structure such as auto-regressive (AR)
and moving average [
        <xref ref-type="bibr" rid="ref25 ref40">25, 40</xref>
        ]. Meanwhile, in nonparametric
methods, such a strong assumption cannot be made originally, but
many empirical studies indicate the magnitude of the expressive
power of the linear model. Among them, research to verify the
magnitude of the changes in the linear model factor loadings is
not attracting much attention; however, it is valuable in many
ifelds including economics, finance, biology, neuroscience, and
computer network analysis. However, economics and finance
contain rich multidimensional time-series data from the beginning,
and they unveil both their economic structure and structural
changes. Changes in the number of factors are the most popular
research topics [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and the structural changes in AR models are
also popular [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ]. Recently, a method using wavelet transform
and PCA has been developed, and this method can be applied
to both structural changes above [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Meanwhile, in the finance
literature, a method that uses ICA has been developed [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>Finally, we clarify our position. First, our method is an
online change point detection method using nonparametric matrix
decomposition. Our method can be considered a matrix
decomposition method; meanwhile, it is close to the kernel method in that
it assumes that the latent time-series follows some model.
However, in the context of online change point detection, time-series
are generated within a small window width; therefore, the
randomly generated time-series variation is large, and our method
can be considered as model-free. Meanwhile, our purpose is to
detect the change point of the structure, and can be applied to
all the structural changes defined above. It is also applicable to
new types of changes such as changes in higher moments of the
mixing matrix. Under such circumstances, our method can be
regarded as a new nonparametric method to detect structural
changes.
3</p>
    </sec>
    <sec id="sec-5">
      <title>PROBLEM STATEMENT</title>
      <p>In this section, we mathematically and clearly explain the models
that we will investigate, define the structural change that we
want to detect and explain the change according to the real-world
applications.Table 2 summarizes the notations used herein.</p>
      <p>We begin from the generative model. We observe a vector yt ∈
RD at every time t . We assume that the vector yt is generated as
a mixture of latent time series</p>
      <p>yt = At zt ,
where At ∈ RD×K is a mixing matrix and zt ∈ RK is a latent
time series. We can observe only yt and that both At and zt are
latent.</p>
      <p>As is often the case in an econometric setting zt is assumed to
follow one stable deterministic or stochastic model. At is assumed
to be piece-wise constant with respect to time and shows the
structure behind data. Therefore, At may change abruptly because
major events occurred and the relation between the latent and
the observed changed. In this study, we denote t ∗ as the time of
a structural change point if the following holds:</p>
      <p>At =
( A(1)
A(2)
where A(1), A(2) ∈ RD×K are constants and A(1) , A(2).
Meanwhile, latent random signal zt is assumed to follow a Gaussian
distribution at any time, and it is independent and identically
distributed (i.i.d.) with respect to time:</p>
      <p>zt ∼ N (0, I K ).
where I K is an identity matrix in RK ×K . Finally, the problem
is to obtain the time of the structural change point t ∗, and the
estimation of At is of no interest.</p>
      <p>
        As for the type of changes that we will detect, we will follow
the literature of econometric and finance [
        <xref ref-type="bibr" rid="ref5 ref7">5, 7</xref>
        ]. The structural
changes defined in this literature can be divided into three types,
that is
(1) change in the n-th moment of At
(2) change in the dimension of At and zt
(3) change in the autocorrelation of yt
Finally, we will explain the real-world applications and
interpretation of these three changes.
      </p>
      <p>Change (1) is the simplest change. For example, the 1st-order
moment change in the At means the abrupt increase and
decrease of the values. The 2nd-order moment change in the At
means some of the latent factors are increasingly or decreasingly
correlated to the observed data. The 4th-order moment change in
the At means the increase or decrease in the number of non-zero
entries of At .</p>
      <p>Change (2) is also under intense investigation in econometrics
because a new connection between the latent and observed
timeseries are recognized and the explanation capability of the old
model decreased if the number of factors increased. Although
this is contrary to the setting, “The dimension of latent data K is
a constant,” we will address this type of change point similarly.</p>
      <p>Change (3) appears not related to our models, but by
substituting ⟨zt , yt −1, yt −2, · · · ⟩ for zt , we can also address the time-series
models. This change is also important in many literature because
the time-series becomes less predictable from the observed
timeseries if the autocorrelation of the y decreased and vice versa.
4</p>
    </sec>
    <sec id="sec-6">
      <title>METHOD</title>
      <p>
        In this section, we propose our structural change point
detection method. Our goal is to detect the change in the matrix At .
Although both coeficients At and factors zt are latent, the
theory of random matrix and statistics [
        <xref ref-type="bibr" rid="ref10 ref11 ref36">10, 11, 36</xref>
        ] suggests that we
can approximate zt by selecting some vectors in a large random
matrix, and that At can be determined by Lasso regression. We
ifrst decompose the matrix of observed vectors based on this
suggestion, and subsequently evaluate the change in the estimated
At .
4.1
      </p>
    </sec>
    <sec id="sec-7">
      <title>Theory of Random Matrix and Lasso</title>
      <p>In this section, we describe the methodology and the theory of
our method. First, we will explain the problem, and how and why
that problem can be solved.</p>
      <p>As described in the previous section, our goal is to detect
the change in At . Most of the existing methods such as PCA,
NMF, and ICA solve this problem by estimating the At and
subsequently detect the change in the estimated At . However, we will
directly calculate the change d(At , At −1) for an ideal distance of
matrix d, and we are not interested in estimating At .</p>
      <p>To calculate the change in structure d(At , At −1), our method
decomposes the observed data yt into a sparse matrix Mt and a
large random matrix Nt , and subsequently calculates the change
in Mt . How our method operates can be explained by the
following three basic ideas.</p>
      <p>
        (1) If we generate many random vectors {Nt(i)}iQ=1, the linear
combination of {N t(i)}iQ=1 may be similar to the true latent
variables zt .
(2) If we perform a Lasso regression analysis, setting Nt
(random vectors) as explanatory variables and yt (observed
time-series) as explained variables, then random vectors
in Nt that are similar to zt may be automatically chosen
and used to explain yt .
(3) If we perform such a Lasso regression analysis and
obtain Mt as a sparse coeficient matrix (i.e., yt = Mt Nt ),
the change in Mt may be related with the change in At
(structural change = true coeficient matrix’s change).
We explain our basic ideas briefly but they are mathematically
formulated and confirmed by the existing random matrix theory
and statistical theory in this section. For idea (1), we can use
the random matrix theory. Suppose we generated K -dimensional
random vectors denoted by {Nt(i)}iQ=1 for Q times. As above, we
are considering the case where Q is much larger than K . For
example, in our experiments, we consider the case with Q =
100 max{D, K }. If the matrix ⟨Nt(1), · · · Nt(Q )⟩ is full rank (i.e.,
rank = K ), some of the {Nt(i)}iQ=1 are linearly independent and
we can obtain any vector in RK (including Zt ) by the linear
combination of {Nt(i)}iQ=1 . The situation where the matrix is full
rank can be interpreted in that any singular value of the matrix
is non-zero. Many studies regarding the singular value of the
random matrix have been conducted [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ], and we used the result
of Tao and Vu [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ].
      </p>
      <p>
        Proposition 4.1 (Theorem 1.3 [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ]). Let ξ be a real random
variable, Mn (ξ ) be the random n × n matrix whose entries are i.i.d.
copies of ξ , and σi (M) be the i-th largest singular value of a matrix
M.
      </p>
      <p>Suppose that E[ξ ] = 0, E[ξ 2] = 1 and E[ξ C0 ] &lt; ∞ for some
suficiently large absolute constant C0. Subsequently, for all t &gt; 0,
we have</p>
      <p>p(nσn (Mn (ξ ))2 ≤ t ) = 1 − e−t /2−√t + O(n−c )
where c &gt; 0 is an absolute constant and implied constants in the
O(.) The notation depends on E[ξ C0 ] but are uniform on t</p>
      <p>
        This proposition can be used in any normalized square,
random matrix (i.e., each entry are i.i.d. random variables whose
mean and variance are 0 and 1, respectively). The stronger result
is obtained in the Gaussian case (Theorem 1.1 in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]). Moreover,
these results can be extended to rectangular matrices (Theorem
6.5 in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]).
      </p>
      <p>Idea (2) is mathematically formulated as follows.</p>
      <p>Proposition 4.2 (mathematical formulation of idea (2)).
Let
• zt 1, · · · , zt n ∈ R be i.i.d random variables.
• a1, · · · an ∈ R be constants.
• yt := a1zt 1 + · · · + anzt n for all t = 0, 1, · · · , T − 1
• wt 1, · · · , wtm ∈ R be i.i.d random variables
• where each pairs zt i , wt j are independent.
and we estimate the coeficients by Lasso, such that</p>
      <p>yt = b1zt 1 + · · · + bnzt n + bn+1wt 1 + · · · + bn+mwtm
holds. Then,
as T → ∞.</p>
      <p>• bi → ai for i = 1, · · · , n
• bi → 0 for i = n + 1, · · · n + m</p>
      <p>
        To prove this proposition, we use a seminal work by
Candes and Tao [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. They proved this proposition within this
assumption (restricted isometry constants)
      </p>
      <p>
        Definition 4.3 (Definition 1.1 [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] Restricted Isometry Constants).
Let F be the matrix with the finite collection of vectors (vj )j ∈J ∈
Rp as columns. For every integer 1 ≤ S ≤ | J |, we define the
S-restricted isometry constants δS to be the smallest quantity
such that FT obeys
      </p>
      <p>(1 − δS )∥c ∥2 ≤ ∥FT c ∥2 ≤ (1 + δS )∥c ∥2
for all subset T ⊂ J of cardinality |T | ≤ S, and all real coeficients
(cj )j ∈T . Similarly, we define the S, S ′-restricted orthogonality
constants θS,S′ for S + S ′ ≤ | J | to be the smallest quantity such
that</p>
      <p>|⟨FT c, FT c ′⟩| ≤ θS,S′ · ∥c ∥ ∥c ′∥
holds for all disjoint sets T , T ′ ⊂ J of cardinality |T | ≤ S and
|T ′| ≤ S ′.</p>
      <p>The numbers δS and θS,S′ measure how close the vectors are
to behaving as an orthonormal system. Subsequently, Proposition
4.2 can be proven as follows.</p>
      <p>
        Proposition 4.4 (Theorem 1.1 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]). Consider the linear model
y = X β + z where X ∈ Rn×p , y ∈ Rn , z ∼ N (0, σ 2In )
Suppose β ∈ Rp is a vector of parameters that satisfies
• ||β ∥L0 = |{i | βi , 0}| = S
• δ2S + θS,2S &lt; 1
where δ2S , θS,2S are the constants defined in Definition Def4.3.
Here, we will estimate β ′ by setting λp = p2 log p
      </p>
      <p>β ′ = argminβ˜ ∈Rp ∥β˜∥L1 subject to ∥X ∗(y − X β˜)∥L∞ ≤ λp σ
Then β ′ obeys</p>
      <p>∥β ′ − β ∥L22 ≤ C12 · (2 log p) · S · σ 2
where C1 = 4/(1 − δS − θS,2S ) with large probability. 3</p>
    </sec>
    <sec id="sec-8">
      <title>4.2 YMN Matrix Decomposition and</title>
    </sec>
    <sec id="sec-9">
      <title>Structural Change Point Detection</title>
      <p>Idea (1) can be interpreted as, "we can generate the true latent
time series zt ;" idea (2) can be interpreted as "we can obtain the
true latent time series zt ". Therefore, by combining ideas (1) and
(2), we will obtain idea (3), "we can generate and obtain the true
latent time-series and the true latent mixing matrix."</p>
      <p>Conjecture 4.5 (mathematical formulation of idea (3)).
Let
• yt = At zt
• yt ∈ RD×T be an observed matrix,
• zt ∈ RK ×T be a latent matrix,
• At ∈ RD×K be a mixing matrix,
• Nt ∈ RQ ×T be a large random matrix.
• Mt ∈ RD×Q be the sparsest matrix such that
∥yt − Mt Nt ∥ ≤ (1 − θ )∥yt ∥ holds for a given
hyperparameter θ .</p>
      <p>Then, distances d1 and d2 exist in RD×K and RD×Q such that
d1(At1 , At2 ) ≃ d2(Mt1 , Mt2 )
holds for all 0 &lt; t1 &lt; t2 &lt; T</p>
      <p>This conjecture suggests that we can approximate the change
of mixing matrix d1(At1 , At2 ) by the change of the sparse matrix
d2(Mt1 , Mt2 ) and that we don’t have to directly identify A in
change point detection problems. In the remainder of this section,
we show the matrix decomposition algorithm and the change
point detection algorithm based on this conjecture. We do not
know the properties of the distances d1andd2. However, we used
the function d2(Mt1 , Mt2 ) := ∥Mt1 ∥L2 /∥Mt2 ∥L2 , which is not a
distance but is robust under the permutations of the row vectors
of M .</p>
      <p>
        Algorithm 14 shows how our matrix decomposition using a
large random matrix is conducted. Our ultimate goal is to
decompose the observed data yt into a random matrix Nt and
sparse matrix Mt such that the change in the Mt is related to
the change in At . To obtain a sparse matrix Mt , we first
generate a large random matrix Nt ; subsequently, our method
minimizes the sparseness of the coeficient matrix within a constraint
∥Y −Mt Nt ∥ ≤ (1 −θ )∥Y ∥ where Y is the D ×T matrix whose row
vectors are yt . Algorithm 2 shows our change point detection
3 The term "with large probability" is used in the original paper [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. This term
means that the probability that the equation holds is above 1 − √π 1log p
4We name this algorithm "YMN" after the equation Y = M N .M is the first initial
of the mixing matrix and mechanism, and N is the first initial of the normalized
random matrix.
      </p>
      <p>Algorithm 1 Matrix decomposition using a large random matrix
(YMN)</p>
      <sec id="sec-9-1">
        <title>Definition</title>
        <p>Y : D × T matrix that we want to decompose
M: D × K sparse matrix
N : K × T large random matrix
Ni : 1 × T i-th row vector of N
E[x ], V [x ]: the mean and variance of vector x ’s entities
θ : threshold value of Lasso fitting score
α : hyperparameter of Lasso, i.e., the coeficient of the L1 term
δ : step value of α , used during hyperparameter optimization
Lasso(X , T ; α ): matrix W that minimizes ||T − W X ||2 +α ||W ||1</p>
      </sec>
      <sec id="sec-9-2">
        <title>End Definition</title>
        <p>N = ⟨N0, · · · , NK −1⟩ ∼ N (0, I )
for i = 0, · · · , K − 1 do</p>
        <p>Ni ⇐ (Ni − E[Ni ])/pV [Ni ]
end for
M ⇐ Lasso(Y , N ; α )
while ∥Y − M N ∥ ≤ (1 − θ )∥Y ∥ do
α ⇐ α + δ</p>
        <p>M ⇐ Lasso(Y , N ; α )
end while
algorithm. First, we use the series of window frames Yt for each
time and perform the matrix decomposition of Yt . At this time,
we have a series of matrix Mt . We calculate the change point
score using the distance between Mt and Mt +1.</p>
        <p>Algorithm 2 Change point detection algorithm using YMN
decomposition</p>
      </sec>
      <sec id="sec-9-3">
        <title>Definition</title>
        <p>w: window size, t : current time
yt : series vector, st : change point score
⟨, ⟩: concatenation of vectors into a matrix.</p>
        <p>MatrixDecomp(Y ): Decompose Y into two matrices by YMN
decomposition.</p>
      </sec>
      <sec id="sec-9-4">
        <title>End Definition</title>
        <p>for t = 0, · · · , T − w − 1 do
Yt ⇐ ⟨yt , yt +1, . . . , yt +w −1⟩
Nt Mt ⇐ MatrixDecomp(Yt )
st ⇐ d2(Mt , Mt +1)
end for
5</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>EXPERIMENT</title>
      <p>To demonstrate that our methods can detect structural change
points, we conducted two experiments. First, we conducted
experiments in artificial settings where their structures and changes
are clear. We demonstrate which type of structural change points
can be detected and how these changes are detected clearly by
these experiments.</p>
      <p>
        Subsequently, we conducted experiments in real-world
settings where the structures and changes in structures are not clear.
We used economics [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and financial [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] data that are popular
in the literature of structural change point detection. We
demonstrate that our method can be used in the real-world through
these experiments.
To verify that our method can detect the structural changes
described above, we conducted experiments with artificial data and
compared the AUC scores to those of the existing methods (PCA,
ICA, and NMF). All of the experiments were conducted in an
"online" setting. In other words, we fetched the time-series data yt
for w times and decomposed the matrix ⟨yt , yt +1, . . . , yt +w −1⟩
into two matrices Mt , Nt and detected the change in Mt . Change
point detection algorithms using PCA, ICA, and NMF are
executed in the same manner as Algorithm 2, that is, by defining
MatrixDecomp in Algorithm 2 as PCA, ICA, and NMF. For
example, if we perform a dimensional reduction from Yt to Zt such
that Zt = Mt Yt holds by PCA, then we calculate d2(Mt , Mt +1).
In addition, if we perform a matrix decomposition by ICA or NMF
such that Yt = Mt Nt holds, then we calculate d2(Mt , Mt +1). Note
that if we perform N MF , we consider Yt + C instead of Yt where
C is a constant such that Yt + C &gt; 0 holds for all t . Table 3
summarizes the settings of these experiments.
      </p>
      <p>In this experiment, we detected the following types of changes.
(1) Mean of Factor Loadings’ Change</p>
      <p>( A (t &lt; t ∗),
At =</p>
      <p>
        A + W
yt +1 = αt yt + βt 1 + γt ϵt
To verify that our method can be used in the real world, we
conducted experiments with two famous datasets in economics [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]
and finance [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. First, we calculated d2(Mt +1, Mt ) at each t . We
then listed A points whose d2(Mt +1, Mt ) are in the top A. Finally,
we clustered these A points into B points using k-means
clustering. Note that the listing and clustering are performed by of-line
settings, whereas our change-point detection method is an online
method. Because no ground truth of change points exists in
economic or financial data, we only consider true- and false-positive
cases. In other words, we listed change points detected by our
methods and checked whether a major event occurred in finance
or economics.
      </p>
      <p>5.2.1 US macroeconomic dataset. We analyzed the US
representative macroeconomic dataset of 101 time series, collected
monthly between 1959 and 2006 (T = 576), for the change points.</p>
      <p>
        We listed 10 remarkable changes resulting from the
aforementioned of-line listing method with P = 30, Q = 10 in Figure 5.2.
Our change point detection methods clearly show the following
major change points in the US economy.
• the Great Moderation period that started in 1983
• major economic recession in early 1970s
• major economic recession in early 1980s
• Internet Bubble between 1999 and 2001
• the oil crisis at 1973
These change points are consistent with the existing research
on business cycles [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. Moreover, the Internet Bubble and the
oil crisis are not detected in previous research in econometrics
[
        <xref ref-type="bibr" rid="ref27 ref5 ref7 ref9">5, 7, 9, 27</xref>
        ].
      </p>
      <p>
        5.2.2 US 48 Industry Portfolios. We analyzed the US 48
industry portfolio collected daily between 1926 and 2018 (T = 24350)
for the change points [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. We listed five most significant changes
by the same clustering methods as Figure 5.2 in Figure 5.3. Our
change point detection methods clearly show the following four
major change points in the US stock market history: Wall Street
Crash of 1929, Kennedy Slide of 1962, Black Monday of 1987, and
ifnancial crisis of 2007. Moreover, we observed that the
magnitudes of these change by the scores.
6
      </p>
    </sec>
    <sec id="sec-11">
      <title>DISCUSSION</title>
      <p>We demonstrated that our method can detect all three types of
structural changes defined in Section 3. This is because
Conjecture 4.5 is true to some degree. That is, by Proposition 4.1, we can
generate the same latent time-series by combining the random
time-series; further, by Proposition 4.2, we can detect true latent
time series by Lasso. Table 4 shows that this method is much
more accurate than the existing matrix decomposition methods.
This is because our method was initially created to detect the
structural changes that we defined, but the existing methods
cannot detect these changes.</p>
      <p>Figure 5.1 shows the L2 norm of the Mt and we can see that
the change of the original mixing matrix At and the change of
the sparse matrix Mt are highly correlated. Hence this figure
supports our Conjecture 4.5.</p>
      <p>However, we could not detect the change (6). This is because
the distance we used does not consider the permutation of the
row. In other words, we consider only ∥Mt ∥L2 at each time t .
Therefore, we cannot detect changes in swaps of elements of At .
We expect that our methods can detect these changes by choosing
some of the row vectors and by performing our decomposition
repeatedly. From the facts above, it is clear that our method can be
applied to the detection of structural changes in economics and
ifnance. As mentioned in the Introduction, econometrics is an
advanced research field in handling the change points; therefore,
we think that this method can be applied to the detection of
structural changes in many fields. For example, neuroscience and
network analysis have similar objectives to ours because they
cannot detect the exact model of the time-series, and structural
changes are important to them.
4.0
Mt
fo3.0
m
r
o
n
2L2.0</p>
    </sec>
    <sec id="sec-12">
      <title>7 CONCLUSION</title>
      <p>We herein proposed a new nonparametric change point detection
method of the structures of multidimensional time series. We
0 20 40 60 80</p>
      <p>time
(a) Experiment (1) Mean of At ’s Change
0 20 40 60 80</p>
      <p>time
(b) Experiment (2) Variance of At ’s Change
0 20 40 60 80</p>
      <p>time
(c) Experiment (3) Skew and Kurtosis of At ’s Change
used a large random matrix to generate the latent time series,
and Lasso to obtain good row vectors and to select the related
row vectors. We also demonstrated that our method could detect
not only typical structural changes used in econometrics and
ifnance but also new types of structural changes such as the
changes in the higher moment of the mixing matrix. With the
random matrix theory and statistical theory of Lasso, we partly
unveiled the mechanism and theory of our methods. However, a
conjecture that fully supports our methods remains unproven.
We demonstrated the efectiveness of our change point detection
techniques by artificial data and real-world data in economics and
ifnance. Similar structural changes may occur in many fields, but
we have allocated comparison with the state-of-the-art methods
and specialization in other fields as future work.</p>
    </sec>
    <sec id="sec-13">
      <title>ACKNOWLEDGMENT</title>
      <p>We would like to thank Tomoki Komatsu, Takeru Miyato,
Tomohiro Hayase, Kosuke Nakago, and Hirono Okamoto for insightful
comments and discussion.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Mohiuddin</given-names>
            <surname>Ahmed</surname>
          </string-name>
          , Abdun Naser Mahmood, and
          <string-name>
            <given-names>Jiankun</given-names>
            <surname>Hu</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>A survey of network anomaly detection techniques</article-title>
          .
          <source>Journal of Network and Computer Applications</source>
          <volume>60</volume>
          (jan
          <year>2016</year>
          ),
          <fpage>19</fpage>
          -
          <lpage>31</lpage>
          . https://doi.org/10.1016/j.jnca.
          <year>2015</year>
          .
          <volume>11</volume>
          .016
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Donald</surname>
            <given-names>W. K.</given-names>
          </string-name>
          <string-name>
            <surname>Andrews</surname>
          </string-name>
          .
          <year>1993</year>
          .
          <article-title>Tests for Parameter Instability and Structural Change With Unknown Change Point</article-title>
          .
          <source>Econometrica</source>
          <volume>61</volume>
          ,
          <issue>4</issue>
          (jul
          <year>1993</year>
          ),
          <volume>821</volume>
          . https://doi.org/10.2307/2951764
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Yasunori</given-names>
            <surname>Aoki</surname>
          </string-name>
          , Ryouhei Ishii,
          <string-name>
            <given-names>Roberto D.</given-names>
            <surname>Pascual-Marqui</surname>
          </string-name>
          , Leonides Canuet, Shunichiro Ikeda, Masahiro Hata, Kaoru Imajo, Haruyasu Matsuzaki, Toshimitsu Musha,
          <string-name>
            <given-names>Takashi</given-names>
            <surname>Asada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Masao</given-names>
            <surname>Iwase</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Masatoshi</given-names>
            <surname>Takeda</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Detection of EEG-resting state independent networks by eLORETA-ICA method</article-title>
          .
          <source>Frontiers in Human Neuroscience</source>
          <volume>9</volume>
          (feb
          <year>2015</year>
          ),
          <volume>31</volume>
          . https://doi.org/10.3389/ fnhum.
          <year>2015</year>
          .00031
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Reinaldo</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Arellano-Valle</surname>
            ,
            <given-names>Luis M.</given-names>
          </string-name>
          <string-name>
            <surname>Castro</surname>
          </string-name>
          , and
          <string-name>
            <surname>Rosangela</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Loschi</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Change Point Detection in The Skew-Normal Model Parameters</article-title>
          .
          <source>Communications in Statistics - Theory and Methods</source>
          <volume>42</volume>
          , 4 (feb
          <year>2013</year>
          ),
          <fpage>603</fpage>
          -
          <lpage>618</lpage>
          . https://doi.org/10.1080/03610926.
          <year>2011</year>
          .611321
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Jushan</given-names>
            <surname>Bai</surname>
          </string-name>
          and
          <string-name>
            <given-names>Serena</given-names>
            <surname>Ng</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Determining the Number of Factors in Approximate Factor Models</article-title>
          .
          <source>Econometrica</source>
          <volume>70</volume>
          ,
          <issue>1</issue>
          (jan
          <year>2002</year>
          ),
          <fpage>191</fpage>
          -
          <lpage>221</lpage>
          . https: //doi.org/10.1111/
          <fpage>1468</fpage>
          -
          <lpage>0262</lpage>
          .
          <fpage>00273</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Jushan</given-names>
            <surname>Bai</surname>
          </string-name>
          and
          <string-name>
            <given-names>Serena</given-names>
            <surname>Ng</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Large Dimensional Factor Analysis</article-title>
          .
          <source>Foundations and Trends® in Econometrics 3</source>
          ,
          <issue>2</issue>
          (
          <year>2008</year>
          ),
          <fpage>89</fpage>
          -
          <lpage>163</lpage>
          . https://doi.org/10.1561/ 0800000002
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Matteo</given-names>
            <surname>Barigozzi</surname>
          </string-name>
          , Haeran Cho, and
          <string-name>
            <given-names>Piotr</given-names>
            <surname>Fryzlewicz</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Simultaneous multiple change-point and factor analysis for high-dimensional time series</article-title>
          .
          <source>Journal of Econometrics 206</source>
          ,
          <issue>1</issue>
          (sep
          <year>2018</year>
          ),
          <fpage>187</fpage>
          -
          <lpage>225</lpage>
          . https://doi.org/10.1016/j. jeconom.
          <year>2018</year>
          .
          <volume>05</volume>
          .003
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Vincent</given-names>
            <surname>Brault</surname>
          </string-name>
          , Sarah Ouadah, Laure Sansonnet, and
          <string-name>
            <surname>Céline</surname>
          </string-name>
          Lévy-Leduc.
          <year>2018</year>
          .
          <article-title>Nonparametric multiple change-point estimation for analyzing large Hi-C data matrices</article-title>
          .
          <source>Journal of Multivariate Analysis</source>
          <volume>165</volume>
          (may
          <year>2018</year>
          ),
          <fpage>143</fpage>
          -
          <lpage>165</lpage>
          . https://doi.org/10.1016/j.jmva.
          <year>2017</year>
          .
          <volume>12</volume>
          .005
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Jörg</given-names>
            <surname>Breitung</surname>
          </string-name>
          and
          <string-name>
            <given-names>Sandra</given-names>
            <surname>Eickmeier</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Testing for structural breaks in dynamic factor models</article-title>
          .
          <source>Journal of Econometrics 163</source>
          ,
          <issue>1</issue>
          (jul
          <year>2011</year>
          ),
          <fpage>71</fpage>
          -
          <lpage>84</lpage>
          . https://doi.org/10.1016/j.jeconom.
          <year>2010</year>
          .
          <volume>11</volume>
          .008
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.J.</given-names>
            <surname>Candes</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Tao</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Decoding by Linear Programming</article-title>
          .
          <source>IEEE Transactions on Information Theory</source>
          <volume>51</volume>
          ,
          <issue>12</issue>
          (dec
          <year>2005</year>
          ),
          <fpage>4203</fpage>
          -
          <lpage>4215</lpage>
          . https: //doi.org/10.1109/tit.
          <year>2005</year>
          .858979
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Emmanuel</given-names>
            <surname>Candes</surname>
          </string-name>
          and
          <string-name>
            <given-names>Terence</given-names>
            <surname>Tao</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>The Dantzig selector: Statistical estimation when p is much larger than n</article-title>
          .
          <source>The Annals of Statistics 35</source>
          ,
          <issue>6</issue>
          (dec
          <year>2007</year>
          ),
          <fpage>2313</fpage>
          -
          <lpage>2351</lpage>
          . https://doi.org/10.1214/009053606000001523
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Liang</surname>
            <given-names>Chen</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Juan J.</given-names>
            <surname>Dolado</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Jesús</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Detecting big structural breaks in large factor models</article-title>
          .
          <source>Journal of Econometrics 180</source>
          , 1 (may
          <year>2014</year>
          ),
          <fpage>30</fpage>
          -
          <lpage>48</lpage>
          . https://doi.org/10.1016/j.jeconom.
          <year>2014</year>
          .
          <volume>01</volume>
          .006
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Ray-Bing</surname>
            <given-names>Chen</given-names>
          </string-name>
          , Ying Chen, and
          <string-name>
            <surname>Wolfgang</surname>
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Härdle</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>TVICA-Time varying independent component analysis and its application to financial data</article-title>
          .
          <source>Computational Statistics &amp; Data Analysis</source>
          <volume>74</volume>
          (jun
          <year>2014</year>
          ),
          <fpage>95</fpage>
          -
          <lpage>109</lpage>
          . https: //doi.org/10.1016/j.csda.
          <year>2014</year>
          .
          <volume>01</volume>
          .002
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Peter F Christofersen</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Forecasting Non-Stationary Economic Time Series</article-title>
          .
          <source>J. Amer. Statist. Assoc</source>
          .
          <volume>96</volume>
          ,
          <issue>453</issue>
          (mar
          <year>2001</year>
          ),
          <fpage>339</fpage>
          -
          <lpage>355</lpage>
          . https://doi.org/10. 1198/jasa.
          <year>2001</year>
          .s378
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cont</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Empirical properties of asset returns: stylized facts and statistical issues</article-title>
          .
          <source>Quantitative Finance</source>
          <volume>1</volume>
          ,
          <issue>2</issue>
          (feb
          <year>2001</year>
          ),
          <fpage>223</fpage>
          -
          <lpage>236</lpage>
          . https://doi.org/10.1080/ 713665670
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Graham</given-names>
            <surname>Elliott</surname>
          </string-name>
          and
          <string-name>
            <given-names>Allan</given-names>
            <surname>Timmermann</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Economic Forecasting</article-title>
          .
          <source>Journal of Economic Literature</source>
          <volume>46</volume>
          ,
          <issue>1</issue>
          (feb
          <year>2008</year>
          ),
          <fpage>3</fpage>
          -
          <lpage>56</lpage>
          . https://doi.org/10.1257/jel.46.
          <issue>1</issue>
          .
          <fpage>3</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Kenneth</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>French</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Detail for 48 Industry Portfolios</article-title>
          . http://mba.tuck.dartmouth.edu/pages/faculty/ken.french/Data_Library/det_ 48_ind_port.html
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Giampiero</given-names>
            <surname>Gallo</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>The methodology and practice of econometrics by J</article-title>
          . L. Castle;
          <string-name>
            <given-names>N.</given-names>
            <surname>Shephard</surname>
          </string-name>
          .
          <source>Journal of Economics</source>
          <volume>101</volume>
          (01
          <year>2010</year>
          ),
          <fpage>99</fpage>
          -
          <lpage>101</lpage>
          . https: //doi.org/10.2307/41795703
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Allan</given-names>
            <surname>Gibbard and Hal R. Varian</surname>
          </string-name>
          and.
          <year>1978</year>
          .
          <article-title>Economic Models</article-title>
          .
          <source>Journal of Philosophy</source>
          <volume>75</volume>
          ,
          <issue>11</issue>
          (
          <year>1978</year>
          ),
          <fpage>664</fpage>
          -
          <lpage>677</lpage>
          . https://doi.org/10.5840/jphil1978751111
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Xu</given-names>
            <surname>Han</surname>
          </string-name>
          and
          <string-name>
            <given-names>Atsushi</given-names>
            <surname>Inoue</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>TESTS FOR PARAMETER INSTABILITY IN DYNAMIC FACTOR MODELS</article-title>
          .
          <source>Econometric Theory</source>
          <volume>31</volume>
          (sep
          <year>2014</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>36</lpage>
          . https://doi.org/10.1017/s0266466614000486
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Zaïd</surname>
            <given-names>Harchaoui</given-names>
          </string-name>
          , Francis Bach, and
          <string-name>
            <given-names>Éric</given-names>
            <surname>Moulines</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Kernel Changepoint Analysis</article-title>
          .
          <source>In Proceedings of the 21st International Conference on Neural Information Processing Systems (NIPS'08)</source>
          . Curran Associates Inc., USA,
          <fpage>609</fpage>
          -
          <lpage>616</lpage>
          . http://dl.acm.org/citation.cfm?id=
          <volume>2981780</volume>
          .
          <fpage>2981856</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Guang-Bin</surname>
            <given-names>Huang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qin-Yu Zhu</surname>
          </string-name>
          , and
          <string-name>
            <surname>Chee-Kheong Siew</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Extreme learning machine: a new learning scheme of feedforward neural networks</article-title>
          ,
          <source>In 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541)</source>
          .
          <source>IEEE International Conference on Neural Networks - Conference Proceedings 2</source>
          ,
          <fpage>985</fpage>
          -
          <lpage>990</lpage>
          vol.
          <volume>2</volume>
          . https://doi.org/10.1109/ijcnn.
          <year>2004</year>
          .1380068
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hyvärinen</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Oja</surname>
          </string-name>
          .
          <year>2000</year>
          .
          <article-title>Independent component analysis: algorithms and applications</article-title>
          .
          <source>Neural Networks</source>
          <volume>13</volume>
          ,
          <fpage>4</fpage>
          -
          <lpage>5</lpage>
          (jun
          <year>2000</year>
          ),
          <fpage>411</fpage>
          -
          <lpage>430</lpage>
          . https://doi. org/10.1016/s0893-
          <volume>6080</volume>
          (
          <issue>00</issue>
          )
          <fpage>00026</fpage>
          -
          <lpage>5</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Daichi</surname>
            <given-names>Kitamura</given-names>
          </string-name>
          , Shinichi Mogami, Yoshiki Mitsui, Norihiro Takamune, Hiroshi Saruwatari, Nobutaka Ono,
          <string-name>
            <given-names>Yu</given-names>
            <surname>Takahashi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Kazunobu</given-names>
            <surname>Kondo</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Generalized independent low-rank matrix analysis using heavy-tailed distributions for blind source separation</article-title>
          .
          <source>EURASIP Journal on Advances in Signal Processing</source>
          <year>2018</year>
          ,
          <volume>1</volume>
          (may
          <year>2018</year>
          ),
          <volume>28</volume>
          . https://doi.org/10.1186/s13634-018-0549-5
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>Jeremias</given-names>
            <surname>Knoblauch</surname>
          </string-name>
          and
          <string-name>
            <given-names>Theodoros</given-names>
            <surname>Damoulas</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Spatio-temporal Bayesian On-line Changepoint Detection with Model Selection</article-title>
          .
          <source>In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research)</source>
          ,
          <source>Jennifer Dy and Andreas Krause (Eds.)</source>
          , Vol.
          <volume>80</volume>
          . PMLR, Stockholmsmassan, Stockholm Sweden,
          <fpage>2718</fpage>
          -
          <lpage>2727</lpage>
          . http://proceedings.mlr. press/v80/knoblauch18a.html
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Shuang</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Yao</given-names>
            <surname>Xie</surname>
          </string-name>
          , Hanjun Dai, and
          <string-name>
            <given-names>Le</given-names>
            <surname>Song</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>M-statistic for Kernel Change-point Detection</article-title>
          .
          <source>In Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 2 (NIPS'15)</source>
          . MIT Press, Cambridge, MA, USA,
          <fpage>3366</fpage>
          -
          <lpage>3374</lpage>
          . http://dl.acm.org/citation.cfm?id=
          <volume>2969442</volume>
          .
          <fpage>2969615</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Shujie</given-names>
            <surname>Ma</surname>
          </string-name>
          and
          <string-name>
            <given-names>Liangjun</given-names>
            <surname>Su</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Estimation of large dimensional factor models with an unknown number of breaks</article-title>
          .
          <source>Journal of Econometrics 207</source>
          ,
          <issue>1</issue>
          (nov
          <year>2018</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>29</lpage>
          . https://doi.org/10.1016/j.jeconom.
          <year>2018</year>
          .
          <volume>06</volume>
          .019
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>R.</given-names>
            <surname>David McLean</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jefrey E.</given-names>
            <surname>Pontif</surname>
          </string-name>
          .
          <year>2016</year>
          . Does Academic Research Destroy Stock Return Predictability?
          <source>The Journal of Finance 71</source>
          ,
          <issue>1</issue>
          (jan
          <year>2016</year>
          ),
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          . https://doi.org/10.1111/jofi.12365
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Inc</surname>
          </string-name>
          .
          <source>National Bureau of Economic Research</source>
          .
          <year>2012</year>
          .
          <article-title>US Business Cycle Expansions and Contractions</article-title>
          . https://www.nber.org/cycles.html
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>Grace</given-names>
            <surname>Ngunkeng</surname>
          </string-name>
          and
          <string-name>
            <given-names>Wei</given-names>
            <surname>Ning</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Information Approach for the ChangePoint Detection in the Skew Normal Distribution</article-title>
          and
          <string-name>
            <given-names>Its</given-names>
            <surname>Applications</surname>
          </string-name>
          .
          <source>Sequential Analysis</source>
          <volume>33</volume>
          ,
          <issue>4</issue>
          (oct
          <year>2014</year>
          ),
          <fpage>475</fpage>
          -
          <lpage>490</lpage>
          . https://doi.org/10.1080/07474946.
          <year>2014</year>
          .961845
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>Wei</given-names>
            <surname>Ning</surname>
          </string-name>
          and
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Gupta</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Change Point Analysis for Generalized Lambda Distribution</article-title>
          .
          <source>Communications in Statistics - Simulation and Computation</source>
          <volume>38</volume>
          , 9 (oct
          <year>2009</year>
          ),
          <fpage>1789</fpage>
          -
          <lpage>1802</lpage>
          . https://doi.org/10.1080/03610910903125314
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Bradley</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Paye</surname>
            and
            <given-names>Allan</given-names>
          </string-name>
          <string-name>
            <surname>Timmermann</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Instability of return prediction models</article-title>
          .
          <source>Journal of Empirical Finance</source>
          <volume>13</volume>
          ,
          <issue>3</issue>
          (jun
          <year>2006</year>
          ),
          <fpage>274</fpage>
          -
          <lpage>315</lpage>
          . https://doi. org/10.1016/j.jempfin.
          <year>2005</year>
          .
          <volume>11</volume>
          .001
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>Mario</given-names>
            <surname>Pitsillis</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>1 - Review of literature on multifactor asset pricing models</article-title>
          .
          <source>In Linear Factor Models in Finance, John Knight and Stephen Satchell (Eds.)</source>
          .
          <source>Butterworth-Heinemann</source>
          , Oxford,
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          . https://doi.org/10.1016/ B978-075066006-8.
          <fpage>50002</fpage>
          -
          <lpage>1</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>James</given-names>
            <surname>Stock</surname>
          </string-name>
          and
          <string-name>
            <given-names>Mark</given-names>
            <surname>Watson</surname>
          </string-name>
          .
          <year>1994</year>
          .
          <article-title>Evidence on Structural Instability in Macroeconomic Time Series Relations</article-title>
          .
          <source>Technical Report. National Bureau of Economic Research</source>
          . https://doi.org/10.3386/t0164
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>Terence</given-names>
            <surname>Tao</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Random matrices: A general approach for the least singular value problem</article-title>
          . https://terrytao.wordpress.com/
          <year>2008</year>
          /05/22/ random-matrices
          <article-title>-a-general-approach-for-the-least-singular-value-problem/</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>Terence</given-names>
            <surname>Tao and Van Vu</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Random Matrices: the Distribution of the Smallest Singular Values</article-title>
          .
          <source>Geometric and Functional Analysis</source>
          <volume>20</volume>
          ,
          <issue>1</issue>
          (mar
          <year>2010</year>
          ),
          <fpage>260</fpage>
          -
          <lpage>297</lpage>
          . https://doi.org/10.1007/s00039-010-0057-8
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>Jennifer</given-names>
            <surname>Thompson</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Smart beta funds pass $ 1tn in assets</article-title>
          . https: //www.ft.com/content/bb0d1830-e56b
          <string-name>
            <surname>-</surname>
          </string-name>
          11e7
          <string-name>
            <surname>-</surname>
          </string-name>
          8b99
          <string-name>
            <surname>-</surname>
          </string-name>
          0191e45377ec
          <fpage>2017</fpage>
          -
          <volume>12</volume>
          -27.
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <surname>Charles</surname>
            <given-names>Truong</given-names>
          </string-name>
          , Laurent Oudre, and
          <string-name>
            <given-names>Nicolas</given-names>
            <surname>Vayatis</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Selective review of ofline change point detection methods</article-title>
          . arXiv:arXiv:
          <year>1801</year>
          .00718
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <surname>Stefan De Wachter and Elias Tzavalis</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Detection of structural breaks in linear dynamic panel data models</article-title>
          .
          <source>Computational Statistics &amp; Data Analysis</source>
          <volume>56</volume>
          ,
          <issue>11</issue>
          (nov
          <year>2012</year>
          ),
          <fpage>3020</fpage>
          -
          <lpage>3034</lpage>
          . https://doi.org/10.1016/j.csda.
          <year>2012</year>
          .
          <volume>02</volume>
          .025
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <surname>Beilun</surname>
            <given-names>Wang</given-names>
          </string-name>
          , arshdeep Sekhon, and
          <string-name>
            <given-names>Yanjun</given-names>
            <surname>Qi</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Fast and Scalable Learning of Sparse Changes in High-Dimensional Gaussian Graphical Model Structure</article-title>
          .
          <source>In Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics (Proceedings of Machine Learning Research)</source>
          ,
          <source>Amos Storkey and Fernando Perez-Cruz (Eds.)</source>
          , Vol.
          <volume>84</volume>
          . PMLR,
          <string-name>
            <surname>Playa</surname>
            <given-names>Blanca</given-names>
          </string-name>
          , Lanzarote, Canary Islands,
          <fpage>1691</fpage>
          -
          <lpage>1700</lpage>
          . http://proceedings.mlr.press/v84/wang18f. html
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <surname>Wei</surname>
            <given-names>Xiao</given-names>
          </string-name>
          , Xiaolin Huang, Jorge Silva, Saba Emrani, and
          <string-name>
            <given-names>Arin</given-names>
            <surname>Chaudhuri</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Online Robust Principal Component Analysis with Change Point Detection</article-title>
          .
          <source>arXiv:arXiv:1702.05698</source>
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>Ian</given-names>
            <surname>En-Hsu</surname>
          </string-name>
          <string-name>
            <given-names>Yen</given-names>
            ,
            <surname>Wei-Cheng</surname>
          </string-name>
          <string-name>
            <given-names>Lee</given-names>
            ,
            <surname>Sung-En</surname>
          </string-name>
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , Arun Sai Suggala,
          <string-name>
            <given-names>ShouDe</given-names>
            <surname>Lin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Pradeep</given-names>
            <surname>Ravikumar</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Latent Feature Lasso</article-title>
          .
          <source>In Proceedings of the 34th International Conference on Machine Learning (Proceedings of Machine Learning Research)</source>
          ,
          <source>Doina Precup and Yee Whye Teh (Eds.)</source>
          , Vol.
          <volume>70</volume>
          . PMLR, International Convention Centre, Sydney, Australia,
          <fpage>3949</fpage>
          -
          <lpage>3957</lpage>
          . http://proceedings.mlr.press/v70/yen17a.html
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>