<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Physics-Informed Spatiotemporal Deep Learning for Emulating Coupled Dynamical Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anishi Mehta</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cory Scott</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Diane Oyen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nishant Panda</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gowri Srinivasan</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Georgia Institute of Technology</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Los Alamos National Laboratory</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of California-Irvine</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1989</year>
      </pub-date>
      <abstract>
        <p>Accurately predicting the propagation of fractures, or cracks, in brittle materials is an important problem in evaluating the reliability of objects such as airplane wings and concrete structures. Efficient crack propagation emulators that can run in a fraction of the time of high-fidelity physics simulations are needed. A primary challenge of modeling fracture networks and the stress propagation in materials is that the cracks themselves introduce discontinuities, making existing partial differential equation (PDE) discovery models unusable. Furthermore, existing physics-informed neural networks are limited to learning PDEs with either constant initial conditions or changes that do not depend on the PDE outputs at the previous time. In fracture propagation, at each timestep, there is a damage field and a stress field; where the stress causes further damage in the material. The stress field at the next time step is affected by the discontinuities introduced by the propagated damage. Thus, both stress and damage fields are heavily dependent on each other; which makes modeling the system difficult. Spatiotemporal LSTMs have shown promise in the area of real-world video prediction. Building on this success, we approach this physics emulation problem as a video generation problem: training the model on simulation data to learn the underlying dynamic behavior. Our novel deep learning model is a Physics-Informed Spatiotemporal LSTM, that uses modified loss functions and partial derivatives from the stress field to build a data-driven coupled dynamics emulator. Our approach outperforms other neural net architectures at predicting subsequent frames of a simulation, enabling fast and accurate emulation of fracture propagation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction and Motivation</title>
      <p>
        Brittle materials fail suddenly with little warning due to the
growth of micro-fractures that quickly propagate and
coalesce. Prediction of fracture propagation in brittle
materials is a multi-scale modeling problem whose time
dynamics are well understood at the micro-scale but do not scale
well to the macro-scale necessary for practical evaluation
of materials under strain
        <xref ref-type="bibr" rid="ref10 ref11 ref15">(White 2006; Hyman et al. 2016;
Kim et al. 2014)</xref>
        . Fracture formation in brittle materials
      </p>
      <p>Micro cracks
and loading
in one cell</p>
      <p>Continuum model</p>
      <p>Damage evolution
accounts for crack</p>
      <p>interactions
Constitutive</p>
      <p>model
for summary</p>
      <p>statistics
Emulate
dynamics</p>
      <p>
        t
is typically simulated using parallel implementations of
finite discrete element methods (FDEM). Industrial software
packages applying these methods have been developed,
many of which are capable of representing the high-fidelity
dynamics and are extremely paralelized
        <xref ref-type="bibr" rid="ref1 ref9">(Hyman et al. 2015;
Rougier et al. 2014)</xref>
        . Yet, these codes are unable to simulate
samples large enough to have real-world scientific
applications, due to the large computational requirements of
simulating the behavior at the spatial and temporal resolutions
necessary. Upscaled continuum representations are used as
an approximation because they discard topological features
of the simulated material and are therefore faster; however,
precisely because they omit these features, they fail to match
experimental observations (Vaughn et al. 2019). Thus, we
develop a spatio-temporal machine learning model to
emulate the micro-scale physics model and estimate the
necessary quantities of interest needed to ensure accuracy of the
continuum-scale model, as in Figure 1.
      </p>
      <p>The goal is to predict summary statistics, or quantities of
interest, for both the damage field and the stress field in a
simulated 2-dimensional material from initial conditions
until the point of failure (when a single fracture spans the width
of the material). The dynamics of the stress field cannot be
modeled without the damage and vice versa. When
damage is static, the evolution of stress over the material
mimics properties of fluid flow. However, the damage caused in
the material changes the behavior of stress to no longer be
governed by a single PDE, e.g. stress accumulates at crack
tips and causes cracks (each of which is a discontinuity in
the stress field) to spread further. Thus, instead of using
solid state dynamics equations to predict this stress field, we
must extend approaches successfully demonstrated in
machine learning to couple the dynamics of the damage field
and stress field.</p>
      <p>It is tempting to treat damage and the stress tensor at each
location simply as different channels in the same time
series and apply methods from the extensive prior work on
video prediction (Wang et al. 2018). However this approach
is ineffective because although the damage and stress fields
are highly coupled, they have dramatically different
dynamics in time. Therefore, one model cannot easily predict both
quantities simultaneously. The damage data is binary-valued
and sparse: most of the finite elements remain undamaged
for the entire simulation, as shown in Figure 2a. The stress
data is real-valued, where values as small as 10 6 are
significant yet magnitudes also range up to 108 (see Figure 2); and
the stress field has spatial discontinuities wherever damage
has occurred. Furthermore, unlike video prediction which is
concerned with precise pixel-by-pixel accuracy, we need to
emulate the most important features of the simulation over
a long time horizon (hundreds of frames in the future) with
high enough accuracy to predict several quantities of interest
needed by the continuum model.</p>
      <p>
        In order to capture the long-term frame dependencies,
recurrent neural networks (RNNs)
        <xref ref-type="bibr" rid="ref17">(Williams and Zipser 1995)</xref>
        have been recently applied to video predictive learning.
Former state-of-the-art models applied complex nonlinear
transition functions from one frame to the next, constructing
a dual memory structure (Wang et al. 2018) upon Long
Short-Term Memory (LSTM)
        <xref ref-type="bibr" rid="ref7 ref8">(Hochreiter and Schmidhuber
1997b)</xref>
        . To emulate the spatio-temporal model, we propose
a Physics-Informed Spatiotemporal LSTM model. First,
linear interpolation is used to coarsen the damage data to
retain the important fracture features while discarding the
uninformative undamaged regions. Next, a modified recurrent
neural network learns temporal evolution in the latent space
representation (Wang et al. 2018). Finally, the predictions
from the recurrent neural network are passed to the decoder
sub-network of the convolutional autoencoder, and decoded
into time-advanced simulation states. As input to the
convolutional autoencoder network, we include point estimates
of partial derivatives of stress values. This allows us to
predict Coupled Dynamical PDEs unlike existing PDE
discovery models.
      </p>
      <p>Results show that this approach makes accurate
predictions of fracture propagation. Our method outperforms other
neural net architectures at predicting subsequent frames of
a simulation, and reproduces physical quantities of interest
with higher fidelity.</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Machine learning-based prediction of behavior in physical
systems in general, and partial differential equations
specifically, is an area of active research. Broadly, machine
learning approaches to PDE emulation fall into one of two
categories. In the first category are approaches that
accelerate methods to solve a PDE whose form is known using
data; for example,
        <xref ref-type="bibr" rid="ref4">(Han, Jentzen, and E 2018)</xref>
        . (Long, She,
and Mukhopadhyay 2018) uses convolutions in the LSTM
cells in a fully convolutional network to train a PDE solver
with varying input perturbations. Our work falls under the
second category of approaches; which emulate the
behavior of a system governed by PDEs; such as fluid
dynamics
        <xref ref-type="bibr" rid="ref12 ref14 ref16 ref3">(Kim et al. 2019; Wiewel, Becher, and Thuerey 2019;
White, Ushizima, and Farhat 2019; Guo, Li, and Iorio 2016)</xref>
        .
Unlike these fluid dynamics emulators where the boundary
conditions and topology are constant, in our case the
evolution of the damage field changes both the boundary
conditions and topology. Furthermore, our problem has a
bidirectional relationship between stress (which is governed
by a PDE, when damage is constant) and damage (which is
not) and hence we cannot fit a simple PDE and simply unroll
forward in time.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Physics-Informed Spatiotemporal Model</title>
      <p>Formally, the problem to solve is: given initial conditions,
predict a time series of damage field and stress field
evolution. The initial conditions given to this generative model
are some number of simulated frames, from which the rest
of the time-series is predicted.</p>
      <sec id="sec-3-1">
        <title>Architecture of the deep learning model</title>
        <p>
          Our data-driven approach to predict physical behavior in a
complex system leverages advances in deep neural networks
          <xref ref-type="bibr" rid="ref6">(Hinton et al. 2012)</xref>
          . We use a Convolutional Neural
Network (CNN)
          <xref ref-type="bibr" rid="ref13 ref6">(Krizhevsky, Sutskever, and Hinton 2012)</xref>
          to
learn a nonlinear mapping from the stress and damage
values in local neighborhoods at time t to the stress and damage
fields at the next time-step. CNNs are designed for problems
with high spatial correlation and translation invariance,
making them an ideal choice for physical problems.
        </p>
        <p>
          In prior work, we found that using a CNN alone to make
predictions at the next time step, tends to make biased
predictions of lower stress and damage values than the truth. As
we unroll the predictions over time, these errors compound
resulting in highly inaccurate predictions of stress values
after 10 or so frames in time; and virtually no predictions of
damage occurring. We incorporate an explicit modeling of
the time component using a recurrent neural network (RNN)
which shares weights over subsequent time-steps of the
input (Pearlmutter 1989). The hidden state of the RNN after
consuming an entire time series thus is a fixed-length
encoding of that (varying-length) time series. Specifically, we
use a Long Short-Term Memory network (LSTM) that
allows the network to separately “remember” both long-term
global context, as well as short-term recent context
          <xref ref-type="bibr" rid="ref7 ref8">(Hochreiter and Schmidhuber 1997a)</xref>
          .
        </p>
        <p>The spatial and temporal elements can be combined with
a Convolutional LSTM or ConvLSTM which maintains the
spatial structure of the input as it processes time series. We
find that the best model is a Spatiotemporal LSTM
(STLSTM) (Lu, Hirsch, and Scholkopf 2017). The main
reason for the improved predictive power is the inclusion of the
Spatiotemporal Memory in each LSTM block in addition to
(a) Damage
(b) Stress t = 10
(c) Coarsened damage field
the Temporal Memory. While temporal states are only shared
horizontally between time-steps, the spatiotemporal state is
shared between the stacked ST-LSTM blocks. This enables
efficient flow of spatial information. We make the memory
representations of the ST-LSTM cells common between all
input fields. This feature allows us to model the highly
codependent nature of the stress and damage fields. Figure 3
shows our novel physics-informed architecture. We
introduce various aspects inspired by the physical properties of
the damage propagation problem allowing for a closer
fitting PDE.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Coarsening of input damage images</title>
        <p>The damage field is very sparse with damaged pixels
forming less than 2% of the entire spatial domain. The increase
in damaged pixels from the initial seed damage at t = 0,
to the damage at the final step when the sample has failed,
is less than 0.2% of the total pixels, as shown in Figure 2a.
This makes it difficult for an ML model to capture and
predict this information, since (formulated as a binary
classification task) the two prediction classes are extremely
imbalanced. Furthermore the distance between cracks is quite
large relative to the size of a crack which complicates the use
of convolutional filters. Hence, we coarsen the damage data
with a linear Lanczos method (Lanczos 1950) with a filter
of 3x3 (see Figure 2c). We then convert the damage field to
a binary 0-1 field by applying a threshold of 0.11 which is
a standard threshold in this domain beyond which damage
cannot be repaired, i.e. any pixels with values higher than
this will be considered as a damaged pixel and all others
are non-damaged. In this manner, we effectively coarsen the
fields by a factor of 8. Empirically, we find that this
coarsening method preserves the important features to accurately
predict physical quantities of interest.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Informing the model with partial derivatives</title>
        <p>The FDEM model that we are emulating is a Markovian
process: the damage and stress fields of the next time-step
are completely determined by the current state. Unlike the
FDEM model, the machine learning model does not have the
actual PDE to define the dynamics, and so we predict each
time-step from up to k previous time-steps. We use k = 3 so
that dynamic information can be observed.</p>
        <p>
          The stress field, without damage, follows a 2nd-order
PDE. To build a deep learning model that fits to such PDEs,
we include the 1st and 2nd order partial spatial and
temporal derivatives as input. Using k = 3 time-steps as input
allows us to capture temporal derivative information
accurately. We use the gradient and Hessian calculating
functions of Tensorflow to calculate the gradients
          <xref ref-type="bibr" rid="ref1">(Abadi and
others 2015)</xref>
          . At each step, we append the derivatives to the
input fields and predict them as part of the next-step
prediction. This enables the spatiotemporal memory blocks of
the network to carry information about spatial and temporal
derivatives of stress. Through this, we overcome the issue of
vanishing gradients discussed in (Wang et al. 2018), as well
as capture the monotonically increasing nature of the
damage field. We ensure the mean squared error of the predicted
derivatives and derivatives calculated from the stress fields
lies within a pre-decided threshold as a self-check. These
modeling choices were imbibed from the physical
principles governing the damage propagation process creating a
novel ”physics-informed” deep learning model. Empirical
results show that this physics-informed approach of training
the model significantly improves accuracy.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiment</title>
      <p>We focus on three variants of LSTM models: Stacked
LSTM, convLSTM, and ST-LSTM. All three models take in
the first k = 3 time-steps of coarsened data as input frames.
Next, to encourage the model to fit to a PDE, we calculate
the 1st and 2nd derivatives of the stress field w.r.t. time and
append that information to the inputs. The LSTM block
calculates predicted values for the next frame corresponding to
each pixel in the input frames. Each model with derivatives
as inputs is called Physics-Informed. The whole model is
unrolled to predict the entire simulation.</p>
      <p>The dataset consists of 61 simulations each of which has
260 time-steps. We split our dataset into 41 simulations used
for training, 10 for validation, and 10 as test cases. We train
our models until saturation which we reach between
350400 epochs. To prevent the model from overfitting, we
perform one round of cross-validation at the end of each epoch.</p>
      <p>
        Our models are designed to allow us to plugin different
ML architectures, choose whether to include partial
derivative information, and test various loss functions. This
modular approach makes it easy to train the necessary
components as needed. Our experiments show that we achieve the
best performance by using 6 ST-LSTM blocks stacked on
top of each other, each of size 128. We use tanh
        <xref ref-type="bibr" rid="ref7 ref8">(Hochreiter and Schmidhuber 1997a)</xref>
        as the activation function for
the LSTM and Leaky-ReLu
        <xref ref-type="bibr" rid="ref5">(Maas, Hannun, and Ng 2013)</xref>
        for the CNN. For our experiments, we test several
combinations of loss functions such as L1 loss, L2 loss, L2 loss
weighted by pixel values, cross-entropy loss, etc. We use the
same loss functions for all our models to directly compare
performance. For the best results, we treat the damage fields
as a binary classification problem i.e. deducing whether a
given pixel i; j is damaged or not, and use a cross-entropy
loss
        <xref ref-type="bibr" rid="ref2">(Goodfellow, Bengio, and Courville 2016)</xref>
        . The
crossentropy loss LD for the damage field is:
      </p>
      <p>LD =
0:5 X</p>
      <p>X
yDij;c log(pDij;c);
(1)
i;j c=f1;2g
where y is a binary indicator (0 or 1) if class label c is the
correct prediction for damage field observation Dij and p is
the probability of the model predicting class c for damage
field observation Dij .</p>
      <p>For the stress fields, we use L1 and L2 losses and a
gradient difference loss (GDL) which sharpens the image
prediction (Mathieu, Couprie, and LeCun 2016). The stress field
loss function LS is given by:
LS =</p>
      <p>3LGDL +
LGDL =</p>
      <p>X
i;j</p>
      <p>X
i;j
jSi;j
jSi;j 1
1(Sij</p>
      <p>S^ij )2 +
2jSij
^</p>
      <p>Sij j ;
Si 1;j j jS^i;j
^</p>
      <p>Si 1;j j +
Si;j j jS^i;j 1</p>
      <p>S^i;j j ;
where i; j ranges over the pixels, S^ij are the predicted stress
values, Sij are the true values and 1, 2, and 3 are
hyperparameters that weight the relative importance of each
term of the loss function. We use 1 = 0:3, 2 = 0:1, and
3 = 0:1.</p>
      <sec id="sec-4-1">
        <title>Prediction of quantities of interest</title>
        <p>The continuum-scale model requires as input several
quantities that describe a material behavior under given conditions.
These quantities of interest (QoI) are: (a) number of cracks
as a function of time; (b) distribution of crack lengths as a
function of time; and (c) maximum stress over the field as
a function of time. To predict these quantities of interest,
we collect stress and damage predictions from our
physicsinformed spatiotemporal generative model; then, we
calculate the QoI.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Evaluation Metrics</title>
        <p>We evaluate model performance with two standard video
similarity metrics and by quantifying the prediction of
quantities of interest. MSE: Mean Squared Error compares the
squared difference between prediction and truth, averaged
(2)
over all pixels. SSIM: The Structural Similarity Index
Metric considers perception-based similarity between two
images (Wang et al. 2004). Note that higher is better for SSIM.
QoI: We weigh the quantities of interest (QoI) defined above
equally and measure the mean absolute error which indicates
how well the continuum model will perform with this model
as an emulator.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <p>Our Physics-Informed ST-LSTM outperforms other models
particularly on MSE and on predicting the QoI, as shown
in Table 1. ST-LSTM does perform slightly better according
to the SSIM metric, but the difference is small and SSIM
measures visual similarity which is not our main goal.
Qualitatively, we see in Figures 5 and 6 our Physics-Informed
ST-LSTM model can faithfully emulate both the stress and
damage field propagation. The Stacked LSTM model in
particular, tends to predict overly smooth stress fields and no
change in damage, even with the Physics Informed model.</p>
      <p>Our model learns an approximation to the physical
equations governing the evolution of stress and damage fields
allowing it to make predictions on previously unseen
conditions. The quantities of interest are then extracted from these
predictions. As an example, Figures 4 and 7 show the results
for these quantities of interest for a held-out test simulation.
From this example, we can see generally that our model
predicts cracks coalescing with neighbor cracks slightly earlier
than when it actually occurs; causing (a) the total damage
to be overestimated, (b) the number of cracks are
underestimated, and (c) the length of individual cracks are over
estimated during the most dynamic parts of the simulation.
This is likely due to the coarsening of the damage field and
is not a major concern. We predict the entire stress field for
all three directions (or channels) of stress and then extract
the maximum value from our prediction to compare against
the maximum value in the ground truth stress field. Figure 7
shows that our model routinely under-estimates the
maximum stress value, yet generally gets the trend and peaks of
the time series. This is a typical result from machine
learning prediction, which tends not to predict extreme values.
We could improve our prediction of this quantity by
optimizing specifically for the prediction of the maximum stress
rather than predicting the entire stress field, but leave this for
future work.</p>
      <sec id="sec-5-1">
        <title>Run-time and speed-up</title>
        <p>High-fidelity simulators for material failure are
computationally expensive, taking on the order of 1500 CPU-hours
to run one simulation of a 2-dimensional material for 260
time-steps, such as in the dataset we use (Rougier et al.
2014). Physics-Informed ST-LSTM accelerates the entire
workflow by generating approximate QoI in a fraction of the
time. We train each model to saturation in 10-12 hours on
four GeForceGTX1080Ti2.20GHz GPUs, after which
emulation of the physical behavior is on the order of
milliseconds, rather than minutes, per timestep. This is a speedup on
the order of 50,000 times faster. Furthermore, once trained,
the model can generate QoI for any number of simulations
drawn from the same initial conditions.</p>
      </sec>
      <sec id="sec-5-2">
        <title>Discussion</title>
        <p>
          The complexity of our model architecture and loss functions
are necessary for accurately emulating a complex
spatiotemporal process over a long time horizon. The LSTM learns the
monotonically increasing nature of the damage field
without any constraints being imposed. This physically-plausible
learned model is an important result that favored the use
of LSTMs that can capture time-dependent evolution
better than conventional neural network architectures.
Explicitly calculating the partial derivatives and including them as
input improves prediction. This is because the model now
fits to a PDE which is a closer approximation to the original
physical problem. The dual memory representation of
spatial and temporal information in our ST-LSTM cell improves
performance of our model on this problem significantly. The
failure of Stacked-LSTM
          <xref ref-type="bibr" rid="ref5">(Hermans and Schrauwen 2013)</xref>
          is also evidence of this. Furthermore, both local and global
spatio-temporal information is important to reduce
compounding errors to make predictions at any given time.
        </p>
        <p>We see that the maximum stress is consistently
underpredicted, even after weighting the losses by actual stress
values. We believe this is due to the inherent nature of ML
finding an average representation from training data and the
inherently difficult inference problem of estimating a
maximum statistic. However, an important point to take note of
is that our model is able to follow the peaks and trends of
the maximum stress quite accurately. Future work in
uncertainty quantification could learn the correction in our
maximum stress estimate.</p>
        <p>The damage model tends to predict crack coalescence
early. We coarsen the simulation data before giving it as
input to our model, which proportionately reduces the
nondamaged regions between cracks. Due to this, our model
tends to predict crack coalescence a few steps earlier than
ground truth. However, the model is able to converge to the
correct number of cracks towards the end of the simulations
(see Figure 4). In future work, learning a coarse
representation, such as with a convolutional autoencoder (Masci et al.
2011), could learn to correct this bias.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>Emulation of complex physical systems has long been a goal
of artificial intelligence because although we can write down
(a) Total proportion of damaged elements
(b) Number of cracks
(c) Crack length distribution prediction
the micro-scale physics equations of such a system, it is
computationally intractable to simulate the physics model
to obtain meaningful predictions on a large scale; yet the
macro-scale patterns of these dynamic systems can be quite
intuitive to humans (Lerer, Gross, and Fergus 2016). We
present Physics-Informed ST-LSTM, an extension and
application of Spatiotemporal LSTM (ST-LSTM) neural
network models to emulate the time dynamics of a physical
simulation of stress and damage in a material. Unlike PDE
emulators that assume a PDE form, our entirely data driven
framework, can be used equally well on high dimensional
experimental studies where binary variables can arise. We
demonstrate that ST-LSTMs outperform two other machine
learning models at predicting these time dynamics and
physical quantities of interest, and furthermore that all three
models increase in performance when they are physics-informed,
that is they have access to the underlying physics of the
simulation. Physics information comes both in the form of
spatiotemporal derivatives, and in a loss function which takes
into account the QoI. We furthermore demonstrate that a
reduced-order model can gainfully capture the time
dynamics of these physical QoI without needing pixel-perfect
accuracy, an important step towards using machine learning to
massively accelerate prediction of complex physics.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>Research supported by the Laboratory Directed Research
and Development program of Los Alamos National
Laboratory (LANL) under project number 20170103DR. AM
supported by the LANL Applied Machine Learning Summer
Research Fellowship. CS supported by the LANL Center for
Non-Linear Studies.</p>
      <p>Lanczos, C. 1950. An Iteration method for the solution of the
eigenvalue problem of linear differential and integral operators.
United States Governm. Press Office.</p>
      <p>Lerer, A.; Gross, S.; and Fergus, R. 2016. Learning physical
intuition of block towers by example. In International Conference on
Machine Learning.</p>
      <p>Long, Y.; She, X.; and Mukhopadhyay, S. 2018. Hybridnet:
Integrating model-based and data-driven learning to predict evolution
of dynamical systems. In Conference on Robot Learning.
Lu, C.; Hirsch, M.; and Scholkopf, B. 2017. Flexible
spatiotemporal networks for video prediction. In The IEEE Conference
on Computer Vision and Pattern Recognition (CVPR).
Maas, A. L.; Hannun, A. Y.; and Ng, A. Y. 2013. Rectifier
nonlinearities improve neural network acoustic models. In Proceedings
of International Conference on Machine Learning.</p>
      <p>Masci, J.; Meier, U.; Cires¸an, D.; and Schmidhuber, J. 2011.
Stacked convolutional auto-encoders for hierarchical feature
extraction. In Artificial Neural Networks and Machine Learning.
Mathieu, M.; Couprie, C.; and LeCun, Y. 2016. Deep multi-scale
video prediction beyond mean square error. In International
Conference on Learning Representations.</p>
      <p>Rougier, E.; Knight, E. E.; Broome, S. T.; Sussman, A. J.; and
Munjiza, A. 2014. Validation of a three-dimensional finite-discrete
element method using experimental results of the split hopkinson
pressure bar test. International Journal of Rock Mechanics and
Mining Sciences.</p>
      <p>Vaughn, N.; Kononov, A.; Moore, B.; Rougier, E.; Viswanathan,
H.; and Hunter, A. 2019. Statistically informed upscaling of
damage evolution in brittle materials. Theoretical and Applied Fracture
Mechanics.</p>
      <p>Wang, Z.; Bovik, A. C.; Sheikh, H. R.; and Simoncelli, E. P. 2004.
Image quality assessment: from error visibility to structural
similarity. In IEEE Transactions on Image Processing.</p>
      <p>Wang, Y.; Gao, Z.; Long, M.; Wang, J.; and Yu, P. S. 2018.
Predrnn++: Towards a resolution of the deep-in-time dilemma in
spatiotemporal predictive learning. In International Conference on
Machine Learning.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , et al.
          <year>2015</year>
          .
          <article-title>TensorFlow: Large-scale machine learning on heterogeneous systems</article-title>
          .
          <source>Software available from tensorflow.org.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ; Bengio,
          <string-name>
            <given-names>Y.</given-names>
            ; and
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <year>2016</year>
          . Deep Learning.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Iorio</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Convolutional neural networks for steady flow approximation</article-title>
          .
          <source>In ACM SIGKDD Int. Conference on Knowledge Discovery and Data Mining.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>J</given-names>
          </string-name>
          .;
          <string-name>
            <surname>Jentzen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and E,
          <string-name>
            <surname>W.</surname>
          </string-name>
          <year>2018</year>
          .
          <article-title>Solving high-dimensional partial differential equations using deep learning</article-title>
          .
          <source>Proceedings of the National Academy of Sciences.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Hermans</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Schrauwen</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Training and analysing deep recurrent neural networks</article-title>
          .
          <source>In Neural Information Processing Systems</source>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Hinton</surname>
          </string-name>
          , G.;
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; Dahl,
          <string-name>
            <given-names>G.</given-names>
            ;
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          -r.; Jaitly,
          <string-name>
            <given-names>N.</given-names>
            ;
            <surname>Senior</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ;
            <surname>Vanhoucke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ;
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ;
            <surname>Kingsbury</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          ; et al.
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1997a</year>
          .
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural Computation.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1997b</year>
          .
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural Computation</source>
          <volume>9</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Hyman</surname>
            ,
            <given-names>J. D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Karra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Makedonska,
          <string-name>
            <given-names>N.</given-names>
            ;
            <surname>Gable</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. W.</surname>
          </string-name>
          ; Painter,
          <string-name>
            <surname>S. L.</surname>
          </string-name>
          ; and Viswanathan,
          <string-name>
            <surname>H. S.</surname>
          </string-name>
          <year>2015</year>
          .
          <article-title>dfnworks: A discrete fracture network framework for modeling subsurface flow and transport</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Hyman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <article-title>Jime´nez-Mart´ınez</article-title>
          , J.; Viswanathan,
          <string-name>
            <given-names>H. S.</given-names>
            ;
            <surname>Carey</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. W.</surname>
          </string-name>
          ; Porter,
          <string-name>
            <given-names>M. L.</given-names>
            ;
            <surname>Rougier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ;
            <surname>Karra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            ;
            <surname>Frash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ;
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          ; et al.
          <year>2016</year>
          .
          <article-title>Understanding hydraulic fracturing: A multi-scale problem</article-title>
          .
          <source>Philosophical Transactions of the Royal Society A: Mathematical</source>
          , Physical and
          <string-name>
            <given-names>Engineering</given-names>
            <surname>Sciences</surname>
          </string-name>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Um</surname>
            ,
            <given-names>E. S.</given-names>
          </string-name>
          ; Moridis,
          <string-name>
            <surname>G. J.</surname>
          </string-name>
          ; et al.
          <year>2014</year>
          .
          <article-title>Fracture propagation, fluid flow, and geomechanics of water-based hydraulic fracturing in shale gas systems and electromagnetic geophysical monitoring of fluid migration</article-title>
          .
          <source>In SPE Hydraulic Fracturing Technology Conference.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Azevedo</surname>
            ,
            <given-names>V. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Thuerey</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gross</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Solenthaler</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Deep fluids: A generative network for parameterized fluid simulations</article-title>
          . In Computer Graphics Forum.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Krizhevsky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and Hinton,
          <string-name>
            <surname>G. E.</surname>
          </string-name>
          <year>2012</year>
          .
          <article-title>Imagenet classification with deep convolutional neural networks</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>White</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ushizima</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Farhat</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Neural networks predict fluid dynamics solutions from tiny datasets</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .00091.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>White</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Review of methods and approaches for the structural risk assessment of aircraft</article-title>
          .
          <source>Technical report, Australian Government Department of Defence, Defence Science and Technology Organisation</source>
          , DSTO-TR-
          <year>1916</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Wiewel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Becher,
          <string-name>
            <given-names>M.</given-names>
            ; and
            <surname>Thuerey</surname>
          </string-name>
          ,
          <string-name>
            <surname>N.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Latent space physics: Towards learning the temporal evolution of fluid flow</article-title>
          . In Computer Graphics Forum.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>R. J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Zipser</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>Gradient-based learning algorithms for recurrent networks and their computational complexity</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>