<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Analysis of Cryptocurrency Commodities with Motifs and LSTM</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Benjamin Barry</string-name>
          <email>benjamin.barry5@mail.dcu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martin Crane</string-name>
          <email>martin.crane@dcu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computing, Dublin City University</institution>
          ,
          <addr-line>Dublin 9</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Time Series Analysis has long been used as a method to help predict future events based on previous data. This area is wellde ned and studied but there are still areas of Time Series Analysis that are problematic. One such problem is isolating patterns in highly volatile datasets over long time intervals. In an e ort to help with such issues e.g. those seen in uctuating nancial datasets, a new approach is needed. Borrowing its name from Music and more recently Computational Biology, Motifs are small repeating patterns in datasets. Using these subsequences, further forecasting can be improved as Motifs can be shown to aid algorithms which require certain initial conditions. This method will potentially lead to new insights in the study of cryptocurrencies. This paper will present a case where an optimized Motif length will be used to aid the prediction of Bitcoin through the use of an LSTM Neural Network, yielding an 8% decrease in RSME for one test case.</p>
      </abstract>
      <kwd-group>
        <kwd>Time Series Analysis</kwd>
        <kwd>Cryptocurrency</kwd>
        <kwd>Motifs</kwd>
        <kwd>LSTM Neural Networks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Forecasting has derived many methods for calculating the future behaviour of a
system. Statistical methods, for instance, look to predict the relationship between
variables when they are reduced to a simplistic expression. The area of statistical
forecasting looking at the e ect time has on a system is known as Time Series
Analysis and this area will form the core of this work. Most aspects of forecasting
Time Series data revolve around nding patterns within said data, as an attempt
to see if these patterns can be projected into the future. As put forward by [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]
Time Series problems can be reduced down to data = pattern + error. Therefore,
if we can isolate the patterns and reduce the error (or noise) in a dataset we can
decompose a series into something we can extend forward in time. However,
what happens if the data has little to no \obvious" patterns? Here traditional
methods can begin to fail and the picture they project becomes less accurate and
hence less reliable. A new method is needed to try and better model these volatile
systems, a change in point of view from the big to the little. Using improvements
in computational power, we are no longer limited to looking at coarse granularity
to try to model patterns. We can now begin to look at microscopic intervals and
a new world of pattern repetition becomes apparent. In this paper, we look at
bringing together the power of time series Motifs to optimize Neural Network's
(NNs) ability to recognise patterns.
      </p>
      <p>In Section 2 we examine previous research in the areas of Motifs with a special
focus on the area of matrix pro les. We also examine prior work in Bitcoin
forecasting as well as looking at some inherent challenges presented. In Section
4 we detail the results of the work in terms of Similarity Scores and Occurrences
before looking at the overall accuracy and a comparison between LSTMs with
and without Motifs. Section 4.4 discusses these results, while Section 5 concludes
with some possible future directions.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <sec id="sec-2-1">
        <title>Motifs</title>
        <p>
          Certain bodies of work have introduced Motifs (unknowingly) as opposed to in
a theoretical framework for their study and evolution. For this, we need to look
to three particular papers that represent a linear progression of the work.
Although there have been several co-writers across the three papers [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]
and [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], Eamonn Keogh features heavily in all and he can be considered the
instigator of the eld. The fundamental concept of Motifs is encapsulated by
Keogh in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] where he de nes a Motif as the subsequence within a time series
which is most similar to the most other subsequences in that series, D(Ci;Cj) &gt;
2R, for all 1 &lt; i K. Here R is a numeric value picked using a Distance Measure
such as the Euclidean norm. A consequence of the de nition is that all Ci are
mutually exclusive. In Figure 1, A, B and C are a single repeating structure found
at di erent points in the time series. This can be seen to be nearly identical in
structure by eye and when compared mathematically they are seen to fall within
our distance range to be called the \same" i.e. our R from above.
The goal of most research within the eld of Matrix Pro les, is to reduce the
run time of the Brute Force approach (O(n2m) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]). The method known as
STAMP (Scalable Time series Any-time Matrix Pro le ) uses the Fast Fourier
Transform to reduce the computation time to O(n2logn) [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The method used
here is STOMP (Scalable Time series Ordered-search Matrix Pro le ) which runs
in O(n2) [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In this algorithm, through algebraic changes that better use the
component order and invariance of the matrix, run time is improved.
The algorithm uses Piecewise Aggregate Approximation to reduce
dimensionality and then Discretization is performed to transform the time series into
\Words" which can be easily compared. This allows for a more e cient
calculation of distances exactly [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] (unlike other methods).
2.3
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Bitcoin and its Modelling Challenges</title>
        <p>
          Operating under the pseudonym \Satoshi Nakamoto", a developer published a
paper [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] where they propose the idea of a Cryptocurrency. From that point
onwards the rise of cryptocurrencies has been exponential and highly volatile 1.
The rise and fall of Bitcoin has led many to try to predict what it will do next [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
The price of Bitcoin has been modelled by, among others, Krystoufek [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] who
found by studying Bitcoin Price according to fundamental Economic laws, the
price is not ever-increasing. Since 2018 Bitcoin is close to its model-implied price.
This nding would seem to back up the notion that Bitcoin had entered a Mature
Phase [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. Due to the (relatively) low amount of trading done on Bitcoin pre-April
2017, the price remained less volatile. With Bitcoins increased popularity, its
value changed wildly from day to day. Hence Bitcoin is best considered in 2 ages,
its Immature state and its Mature state. As the price of Bitcoin in the Immature
state has been found to be almost perfectly correlated with other Cryptocurrency
prices [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], this has given rise to arbitrage opportunities which have been pursued
as a means to nd the "next" Bitcoin. Further investigation of the variability of
Bitcoin price has taken the form of research into existence of regime changes in
the price, [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] found that Recurrent Neural Networks outperform Hidden Markov
Models at predicting volatility spikes.
        </p>
        <p>
          Other authors have also found challenges due to the extreme uctuations seen
in the Bitcoin market over small increments of time. For example in July 2017,
its value increased 25% in a single day. Previous research has attempted to solve
the problem using the correlation between the number of transactions and price.
This yielded positive results for Bitcoin returns without, however, being able
to predict volatility [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. The issue with this approach is the need to know the
number of transactions in order to create a prediction. A method is needed that
can model forward the market with less restrictive input parameters.
A statistical approach is taken by Shah and Zhang [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] where they use Bayesian
Regression to model future values of Bitcoin. They achieve good results with
this approach and are able to nearly double their initial investment in less than
a 60-day period.
        </p>
        <p>LSTM RNNs have been found to be an e cient way to predict Bitcoin price
from historical prices and moving average technical indicators. Other Machine</p>
        <sec id="sec-2-2-1">
          <title>1 Its max value over $18,000 to current</title>
          <p>
            $8,000 (coinbase.com/price/bitcoin).
Learning methods, e.g. [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ] found LSTM RNNs produced the highest accuracy on
price prediction. [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ] used LSTM model to predict Bitcoin price with, as model
inputs, macroeconomics factors such as global currency ratios and blockchain
information. [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] have used them to mine news articles and tweets to infer the
relationship between their information, certain commodity prices and Bitcoin
price direction yielding an accurate prediction technique.
          </p>
          <p>(a) Immature Bitcoin
(b) Mature Bitcoin</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Methods</title>
      <sec id="sec-3-1">
        <title>Matrix Pro les</title>
        <p>
          What underpins the Keogh algorithms [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] and [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] is the concept of a
Matrix Pro le. In essence, this reduces a time series T into a matrix where each
cell represents the distance between a subsequence and all other subsequences
of equal length in the time series. This produces a matrix of size N by N (N
is the length of the time series T less the length on the subsequence). For a
time series T with 100 data points the Matrix Pro le for all subsequences of
length 10, would consist of a 90 by 90 matrix. A common metric used is the
z-normalized Euclidean distance due to its ability to account for variance and
mean. Consider Figure 3, where the blue and red lines represent 2 di erent time
series T and S. Here the 2 series are mapped on top of each other over a period
of time with arbitrary values on the y-Axis. The vertical lines show the distance
between corresponding points on each series (i.e at the same point in time). The
modi ed sum of these distances represents the Euclidean distance at a simpli ed
level. For Motifs, we are only interested in the lowest distance value i.e. the most
similar subsequence.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Data Used</title>
        <p>Three separate time granularities were chosen for both the Mature and Immature
data within Bitcoin. Since the Poloniex2 data source was accurate to 5-minutes,
this was the lowest granularity picked. From here, 15-minute and 30-minute
intervals were also chosen.This data counts are shown in Table 1. Graphs of the
Mature and Immature Bitcoin data are shown in Figure 2.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <sec id="sec-4-1">
        <title>Matrix Pro le Outputs</title>
        <p>A sample of the standard STUMPY algorithm outputs are seen in Table 2.
This table shows the Starting position of a subsequence (Index ), the Euclidean
Distance (Similarity ) to its closest copy, the starting index of where that copy
is (Start ) and how often the pattern occurs (Occurrence).</p>
        <sec id="sec-4-1-1">
          <title>2 https://poloniex.com/</title>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>Matrix Pro le Graphs</title>
        <p>The data in Table 2 can be used to create an overall result for all granularities
within the data set graphically. Figure 4 shows the Similarity (Euclidean
Distance) of each subsequence of length 3, 6, 12, 24 and 48. With Figure 4(e) as
one of the clearest examples, we can see peaks within the graph or \Discords".
These can be considered the opposite of Motifs i.e. they represent seldom
repeating occurrences or anomalies. The lower the value on the y-Axis the closer
the Motifs are to each other, i.e. over larger windows we can see the Similarity
Score increasing so the Motifs are becoming less similar.
4.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Long Short Term Memory (LSTM)</title>
        <p>Due to the large number of points in the input data (Immature 15-minutes and
Mature 15-minutes) and limited hardware resources3 the LSTM NN takes a lot
of time to run. The computation times were noted as 23.1 hours, 7.5 hours and
26.6 hours, 8.4 hours on the Mature and Immature dataset running the Motif
and the standard algorithm respectively. A comparison can be made between
our version (with Motifs) and a standard 1 point LSTM NN in Figures 7 (a) &amp; 8
(a) for the Mature and Immature cases respectively. Table 3 show Mean Square
Error for the Trained and Tested case for Mature and Immature for Motifs with
LSTM NN and the 1 point LSTM NN.
Optimum Motif Length The goal from analysis of the outputs is to attempt
to nd an \Optimized" Motif that we can then pass into the LSTM to try to
improve its accuracy. To do this, it was best to compare the Motif lengths across
3 The code was run on an Intel i7-5600 CPU @ 2.60GHz with 8GB of RAM
(a) Matrix Pro le Immature 5 Minutes
(b) Matrix Pro le Mature 5 Minutes
(c) Matrix Pro le Immature 15 Minutes
(d) Matrix Pro le Mature 15 Minutes
(e) Matrix Pro le Immature 30 Minutes (f) Matrix Pro le Mature 30 Minutes</p>
        <p>Fig. 4. Matrix Pro le for Di erent Granularities
two metrics: Similarity Scores and Motif Occurrence. Hence, to establish which
Motif length to input, each will have to be taken into account and an overall best
value will need to be selected. The answer from this section mixes the qualitative
and quantitative. We need to visually interpret graphical outputs, in conjunction
with the underlying numeric scores. As such, the optimum Motif length is open
to interpretation.</p>
        <p>
          Similarity Score For both the Mature and Immature dataset for the ve
subsequence lengths tested, an output as seen in Table 2 was created. The longer
Motifs tend to get less similar (this makes sense intuitively with more points
to calculate the Euclidean distance between). The pitfall at this point would
be to just consider the length 3 Motif as the best option due to it having the
smallest mean Euclidean distance. However, we must consider the reason behind
this. Figure 5 contains the graphed Similarity Score (in intervals of 0.5 where
the nal value on the x-Axis contains all values above it). All these graphs show
the dominance of the zero score Motifs. These occur often at low granularity
due to minute di erences in Bitcoin's value. These e ectively are straight lines
in the graph. This behaviour is dominant in the 5-minute 3 and 6 length Motifs
(it is also dominant in most of the Immature data due to the lack of trading at
the beginning of Bitcoins life). We need to nd the best compromise between
high Similarity Scores and the dominance of the straight-line \0" score. In all
six graphs in Figure 5, the blue (3 Motif) and the orange (6 Motif) plots are
dominated by the 0 score Motifs, while the purple (48 Motif) plot occurs more
often at the higher values. The green (12 Motif) plot and the red (24 Motif)
plot are our best compromise. Of note in these graphs is the apparent Gaussian
distribution sitting on the spread of some scores (this is most obvious for the red
(24 Motif) plots). This may be explained using the Law of Large Numbers [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
Due to the lack of dominance of the 0 score and as the there are fewer larger
scores, the values are tending towards their Expected Value. This e ect is more
pronounced in the Mature dataset due to the movement of Bitcoin towards a
traditional commodity [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. To split our choice between a 12 Motif and 24 Motif
we consider the Occurrence counts of our Motifs.
        </p>
        <p>Motif Occurrence When we speak of the \Occurrence" of a Motif we are
referring to how often it appears in a dataset. If we look back to Table 2 we
can explain the eld as how many times the Start, i.e. the starting index of the
repeating subsequence relative to the Index, occurs. This is to say that the
subsequence at Index 0, along with 26 other subsequences are most similar to the
subsequence at Index 14292. This means the Motif repeats throughout the time
series multiple times. Although all 27 subsequences may not have the same
Similarity Score as 0, they will form a Motif. Highly repeating structures are again
usually akin to straight lines and often, after a further look, are uninformative.
Due to the high number of possible values for the Occurrence in the graphs, the
values need to be bucketed: Anything with an Occurrence value of 1 or 2 is kept
as is, whilst 3-5, 5-10 and 10-50 are grouped and nally anything above 50 are
(a) Similarity Scores Immature 5 Minutes
(b) Similarity Scores Mature 5 Minutes
(c) Similarity Scores Immature 15 Minutes
(d) Similarity Scores Mature 15 Minutes
(e) Similarity Scores Immature 30 Minutes
(f) Similarity Scores Mature 30 Minutes
included in the 50+ bucket. Subsequences with an Occurrence value in these
categories are counted and used as the y-Axis value. In Figure 6, we are
interested in the green (12 Motif) and red (24 Motif) plots from analysis in Section
4.4. From the graphs there is little to choose between the two Motif lengths. In
general, the green plot seems to have more of the undesired Occurrence values
of 1 and Occurrence values in the 50+ range. Hence, we can select the 24 Motif
as the best option between the two and proceed to using it in the LSTM NN.
4.5</p>
      </sec>
      <sec id="sec-4-4">
        <title>LSTM Accuracy</title>
        <p>The LSTM NN was not optimized as that was not the goal of this paper, as such,
some of the hyperparameters chosen are not useful in a real world scenario. This
is especially true of the batch size of 1. An increase here would have reduced
the computation time. In this sense there is little to no value in using the
computation times as a metric to measure if the Motif using LSTM outperforms a
traditional LSTM.</p>
        <p>The simple goal of the LSTM NN is to see does using Motifs as an
intermediary step between raw data and model help our outcome. Thus Figures 7 (a) and
(b) let us compare our Motif version to the standard version of the algorithm for
the Mature dataset. It is clear that the former tracks its training data much more
e ectively. What is very obvious is that after only 25 epochs the test outputs
(a) Motif Repetition Immature 5 Min- (b) Motif Repetition Mature 5 Minutes
utes
(c) Motif Repetition Immature 15 Min- (d) Motif Repetition Mature 15
Minutes utes
(e) Motif Repetition Immature 30 Min- (f) Motif Repetition Mature 30
Minutes utes
trend very closely to the true values, just shifted down in value by roughly 750.
Table 3 shows a reduction in the test RMSE by 8% and training of 24%. There
are signi cant computation time increases (66%), but with script optimization
and better hardware this could be reduced. Figures 8 (a) and (b) show the e ect
introducing Motifs has to the Immature dataset. Table 3 shows us that there is
a reduction in the test RSME by 33% and training of 50%.
(a) LSTM NN for Mature Bitcoin using Motifs
(b) LSTM NN for Mature Bitcoin not using Motifs
The goal of this work was to study the application of Motifs in pattern detection
in volatile data. The computation speed and exactness of the Matrix Pro le
make it a powerful tool and results back this. Future work could be to further
integrate Motifs into LSTMs by putting Motifs in the individual NN nodes. If,
training a LSTM, a recognised Motif occurred, forcing the node to abandon its
standard approach of randomizing the weights matrix and simply complete the
Motif could help our learning rate. This would require a more pro cient NN to
be built to better handle the computations and reduce run-times.
There is the potential that capturing enough Motifs in a sequence could replace
a series with a Motif piecewise reduction, reducing the total information stored
by potentially multiples of bits. All these are possible future directions.
The work presented here is a simple use of Motifs in a practical way. The results
indicate that there is strong potential to embed Motifs in di erent areas of
Computer Science.
6</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>The authors are grateful to Prof. H. J. Ruskin for her helpful comments. This
research was (partially) conducted with the nancial support of SFI under Grant
Agreement No. 17/SP/5447 at the ADAPT SFI Research Centre at DCU. The
ADAPT SFI Centre for Digital Media Technology is funded by Science
Foundation Ireland through the SFI Research Centres Programme and is co-funded
under the European Regional Development Fund (ERDF) through Grant no.
13/RC/2106.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Balcilar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bouri</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roubaud</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Can volume predict Bitcoin returns and volatility?</article-title>
          <source>Economic Modelling</source>
          <volume>64</volume>
          ,
          <issue>74</issue>
          {
          <fpage>81</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chiu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keogh</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lonardi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Probabilistic discovery of time series motifs</article-title>
          .
          <source>In: Proc. of 9th Int. Conf. on KDD</source>
          , Washington, DC, Aug
          <volume>24</volume>
          -
          <fpage>27</fpage>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cretarola</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Figa-Talamanca</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>Detecting bubbles in Bitcoin price dynamics via market exuberance</article-title>
          .
          <source>Annals of Operations Research (Jul</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Drozdz</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gebarowski</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Minati</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oswiecimka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Watorek</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Bitcoin market route to maturity? CoRR abs/
          <year>1804</year>
          .05916 (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Feller</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>The strong law of large numbers</article-title>
          . In:
          <article-title>An introduction to probability theory and its applications</article-title>
          , vol.
          <volume>1</volume>
          (
          <issue>3</issue>
          ), pp.
          <volume>243</volume>
          {
          <fpage>245</fpage>
          .
          <string-name>
            <surname>Wiley</surname>
          </string-name>
          (
          <year>1968</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Jang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ko</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Predicting bitcoin prices by using rolling window lstm model</article-title>
          .
          <source>In: Proc of KDD Data Science In Fintech Workshop (DSF)</source>
          (
          <year>July 2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kinderis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bezbradica</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crane</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Bitcoin currency uctuation</article-title>
          .
          <source>In: 3rd Int Conf on Complexity, Future Information Systems and Risk</source>
          ,
          <volume>20</volume>
          -21
          <string-name>
            <surname>Mar</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Kodama</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pichl</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaizoji</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Regime change and trend prediction for Bitcoin time series data</article-title>
          .
          <source>In: CBU Int. Conf. Proc. vol. 5</source>
          , pp.
          <volume>384</volume>
          {
          <issue>388</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Kristoufek</surname>
          </string-name>
          , L.:
          <article-title>Is the Bitcoin price dynamics economically reasonable? Physica</article-title>
          A p.
          <volume>120873</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Law</surname>
            ,
            <given-names>S.M.:</given-names>
          </string-name>
          <article-title>STUMPY: A powerful and scalable python library for time series data mining</article-title>
          .
          <source>The Journal</source>
          <volume>4</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keogh</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lonardi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Finding motifs in time series</article-title>
          .
          <source>Proceedings of the Second Workshop on Temporal Data</source>
          Mining pp.
          <volume>53</volume>
          {
          <issue>68</issue>
          (10
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Makridakis</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wheelwright</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hyndman</surname>
          </string-name>
          , R.:
          <article-title>Forecasting methods and applications</article-title>
          . John Wiley &amp; Sons (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>McNally</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roche</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Predicting the price of bitcoin using machine learning</article-title>
          .
          <source>In: 26th Euromicro Int Conf on Parallel, Distributed and Network-based Processing (March</source>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Mueen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keogh</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cash</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Westover</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Exact discovery of time series motifs</article-title>
          .
          <source>In: Proc 2009 SIAM Int conf on data mining</source>
          . pp.
          <volume>473</volume>
          {
          <issue>484</issue>
          (04
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Nakamoto</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Bitcoin: A peer-to-peer electronic cash system (</article-title>
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <surname>K.</surname>
          </string-name>
          :
          <article-title>Bayesian regression and Bitcoin</article-title>
          .
          <source>In: 52nd Allerton Conf on Communication, Control, and Computing</source>
          , Monticello, IL, Sept 30 - Oct 3 (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>