<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Continuous Representation of Molecules Using Graph Variational Autoencoder</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>GDCN Decoder</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Mohammadamin Tavakoli, Pierre Baldi Department of Computer Science, University of California</institution>
          ,
          <addr-line>Irvine</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In order to continuously represent molecules, we propose a generative model in the form of a VAE which is operating on the 2D-graph structure of molecules. A side predictor is employed to prune the latent space and help the decoder in generating meaningful adjacency tensor of molecules. Other than the potential applicability in drug design and property prediction, we show the superior performance of this technique in comparison to other similar methods based on the SMILES representation of the molecules with RNN based encoder and decoder.</p>
      </abstract>
      <kwd-group>
        <kwd>Adjacency tensor</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Using machine learning to predict molecular structure
properties is a challenging problem [
        <xref ref-type="bibr" rid="ref3">7, 3</xref>
        ]. While the governing
equations (e.g. Schrodinger equation) are difficult and
computationally expensive to solve, the fact that an underlying
model exists is appealing for machine learning techniques.
However, this problem is difficult from a technical point of
view. The space of molecules is discrete and non-numerical.
Thus, “how to best represent molecules and atoms for
machine learning problems?” is still a question.
      </p>
      <p>
        Despite having numerous ways to represent molecules
such as methods introduced in [
        <xref ref-type="bibr" rid="ref14">18, 1</xref>
        ], all the representations
are suffering from a few shortcomings, such as 1) discrete
representation, 2) lengthy representation, 3) non-injective
mapping, and 4) non-machine readable representation.
      </p>
      <p>
        Here, we proposed a new method that borrows the main
idea from [
        <xref ref-type="bibr" rid="ref1">5</xref>
        ] and [
        <xref ref-type="bibr" rid="ref8">12</xref>
        ] and overcomes all the aforementioned
shortcomings. Our method which takes the graphical
structure of the molecule as the inputs consists of a variational
framework with a side predictor to better prune the structure
of the latent space. Then an inner product decoder transfers
the samples of latent space into meaningful adjacency
tensors. To compare with the main benchmark which is a
textbased encoding of molecules [
        <xref ref-type="bibr" rid="ref5">9</xref>
        ] we performed two
experiments on the QM9 dataset [
        <xref ref-type="bibr" rid="ref11 ref12">16, 15</xref>
        ] and ZINC [
        <xref ref-type="bibr" rid="ref7">11</xref>
        ]. Both
experiments show the success of this method. Although this
Copyright c 2020, for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0 International
(CCBY 4.0).
      </p>
      <p>Node-feature matrix
Adjacency tensor</p>
      <p>Side target

GCN
Encoder

GCN
Encoder</p>
      <p>Side predictor</p>
      <p>Pooling
work is presenting preliminary results of Graph VAE, further
experiments and comparisons are left to future work.</p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <p>Molecules and Graphs A molecule can be represented by
an undirected graph G = (V; E; R), with nodes (atoms) vi
2 V and labeled edges (bonds) (vi; e; vj ) 2 E where r 2 R
is an edge type. Since we focus on small molecules with four
bond types, R is equal to 4. An n by d node-feature matrix
H is also carrying more information about each node. These
two tensors, together, represent a molecular structure.
Variational Autoencodes To help ensure that points in the
latent space correspond to valid realistic molecules, and to
minimize the dead areas of the latent space, we chose to use
a variational autoencoder (VAE). To further ensure that the
outputs of the decoder are corresponding valid molecules we
employed the open-source cheminformatics suite RDKit30
to validate the chemical structures of output molecules in
terms of atomic valence. All invalid outputs are discarded.
It is necessary to mention that the ordering of the nodes
assumed to be unchanged.</p>
      <p>
        VAE and Side Prediction To better learn the graph
structure of the molecules, the encoder part of the VAE consists
of GCN layers. The same method as [
        <xref ref-type="bibr" rid="ref13">17</xref>
        ] has been employed
to perform relational update which can be formulated as:
hl+1 =
i
(X
      </p>
      <p>X
where Nri denotes the set of nodes connected to node i
through the edge type r 2 R. Since we are focusing on small
molecules, we applied three layers of GCN in our encoder
model to gather information from 3-hop neighbors of each
atom. The structure of encoder consists of two, three-layer
GCNs for both mean and the covariance. GCNs of the
encoder share the filters of the first two layers. Here we can
formulate the encoding and sampling scheme as follows:
q(ZjH; A) =</p>
      <p>N
Y qi(zijH; A);
1
qi(zijH; A) = N (zijGCN ; GCN )</p>
      <p>
        The GCN and similarly GCN are: GCN (H; A) =
A^ (A^ (A^H W0)W1)W2, where the A^ is the normalized
adjacency tensor, Wi is the filter parameter of each layer, and
is the activation function [2]. Finally, as suggested in [
        <xref ref-type="bibr" rid="ref8">12</xref>
        ]
we use the simplest form of the decoder which can be seen
as graph deconvolution network. The output of the encoder
is simply the inner product between latent variable:
      </p>
      <p>N N
p(AjZ) = Y Y p(Aij jzi; zj );</p>
      <p>1 1
p(Aij = 1jzi; zj ) =
(ziT zj )</p>
      <p>For the side prediction part, we employ a simple
regression model in the form of a multilayer perceptron (MLP) to
the network that predicts the properties from the latent space
representation. The input of the side predictor is a vector
obtained through a pooling mechanism of the latent
representation as follows:</p>
      <p>N</p>
      <p>G(H (L)) = X sof tmax(hiL:Wp)
Where WP is the pooling weight matrix and H (L) is the
output of the GCN .</p>
      <p>Finally, the autoencoder is trained jointly on the
reconstruction task and a property prediction task; The joint loss
function is the summation of the two losses, as follows:
i=1
L = ELBO + negative log likelihood
= Eq(ZjH;A)</p>
      <p>KL(q(ZjH; A)jjp(Z))
+ M SE(sidenetwork)</p>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>We performed two experiments to show the usefulness of
continuous representation. In the first experiment, we focus
on the prediction of property and the generation of the valid
molecules. In the second experiment, we use this continuous
representation to propose a new metric for measuring the
molecular similarity.</p>
      <p>Nicotine</p>
      <p>Ecstasy
Amphetamine</p>
      <p>Caffeine</p>
      <p>
        Aspirin
Using a subset of QM9 dataset [
        <xref ref-type="bibr" rid="ref11">15</xref>
        ] as the training set,
we extract 48,000 molecules covering a broad range of
molecules. Each molecule in the training set is chosen to
have up to 20 atoms. The training objective on the side
predictor was set to be one of the Solubility, Druglikeliness,
and Synthesizability. We employ the continuous
representation of molecules using each network to predict the other
two unseen properties. The performance of each model plus
the percentages of validly generated molecules are
summarized in Table 1. In order to check the validity of the
outcome, we only check for the validity of the atomic valence.
As it is shown in Table 1 the accuracy of each property is
comparable to the state of the art property predictions
mentioned in [
        <xref ref-type="bibr" rid="ref4">8</xref>
        ]. Although Graph VAE is not outperforming the
predictions based on [
        <xref ref-type="bibr" rid="ref4">8</xref>
        ], it shows that using a property as a
heuristic to prune the latent space, can help with predicting
other molecule properties.
      </p>
      <p>
        Molecular Similarity Measure
Numerous similarity or distance measures have been used
widely to calculate the similarity or dissimilarity between
two samples. Since metrics are focusing more on
2dimensional representation rather than 3-dimensional
structure, our model as a “2D structure-aware representation”
is an accurate metric for the similarity measure.
Normalized Euclidean distance between the latent representation of
two molecules after pooling operation is the metric we
define to capture the similarity. Here we compare three
wellknown similarity measures with our technique and also to
the methods introduced in [
        <xref ref-type="bibr" rid="ref5">9</xref>
        ]. This method which is
using the SMILES representation of the molecules as the
input employs a VAE with a side predictor. Both encoder
and decoder parts of the VAE are based on RRN and
sequence to sequence model. Although all the graphical
information of the molecule is encoded within the SMILES
representation, inferring the graphical structure (e.g.,
adjacency tensor) from the SMILES string is an exhausting
process that is based on several rules. Despite the numerous
techniques built upon using the SMILES representation of
the molecules [
        <xref ref-type="bibr" rid="ref10 ref2 ref6 ref9">6, 10, 4, 14, 13</xref>
        ], it has been shown that it
is more efficient to take advantage of the graph structures
and employ GCNs to process molecular structures. Here,
we chose Aspirin as a sample drug and compare its
similarity with four different drugs with four different similarity
measures. We compare the performances of our technique
with [
        <xref ref-type="bibr" rid="ref5">9</xref>
        ], which is using a similar approach but operating on
text representation of molecules. Our experiment shows that
graph-based hidden representation is carrying more
information than only text. Table 2 is summarizing the result of
the similarity measure experiment.
      </p>
      <p>As it is shown in table 2, our metric is very well aligned
with all other well-known metrics which is another proof for
the applicability of our model.</p>
    </sec>
    <sec id="sec-4">
      <title>Experiment Details</title>
      <p>GVAE consists of two GCNs for the encoder, a pooling
mechanism, and a multi-layer perceptron for the side
prediction. Both GCNs are three-layer networks with filter
matrices W0; W1, and W2 of 32*32, 32*32m and 32*16
respectively. The pooling weight matrix Wp is of size 1*64
which outputs a vector of length 64 to represent the whole
molecule. A two-layer MLP with 32 and 1 hidden units is
employed to perform the regression task.</p>
      <p>In Table 2, we use our own implementation of the SMILES
VAE. Both GVA and SMILES VAE are trained using a
dataset of 70,000 molecules which are randomly selected
from ZINC.</p>
      <p>In Table 2, all measures except the continuous
representations are calculated with the same fingerprinting algorithm.
It identifies and hashes topological paths (e.g. along with
bonds) in the molecule and then uses them to set bits in a
fingerprint of length 2048. The set of parameters used by
the algorithm is - minimum path size: 1 bond - maximum
path size: 7 bonds - number of bits set per hash: 2 - target
on-bit density 0.3.</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>We proposed a generative model through which we can find
continuous representation for molecules. As shown in the
experiments section, this technique can be used in different
chemoinformatics tasks such as drug design, drug discovery
and property prediction. As future work, one can think of
attention based graph convolutions and more complicated
decoders. These two extensions can be studied in future works.</p>
      <p>References
[1] E. J. Bjerrum. Smiles enumeration as data augmentation
for neural network modeling of molecules. arXiv preprint
arXiv:1703.07076, 2017.
[2] D.-A. Clevert, T. Unterthiner, and S. Hochreiter. Fast and
accurate deep network learning by exponential linear units
(elus), 2015.
[3] C. W. Coley, W. Jin, L. Rogers, T. F. Jamison, T. S. Jaakkola,
W. H. Green, R. Barzilay, and K. F. Jensen. A
graphconvolutional neural network model for the prediction of
chemical reactivity. Chemical Science, 10(2):370–377, 2019.
[4] A. Dalke. Deepsmiles: An adaptation of smiles for use in.</p>
      <p>2018.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Duvenaud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Maclaurin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Iparraguirre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bombarell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hirzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aspuru-Guzik</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>R. P.</given-names>
            <surname>Adams</surname>
          </string-name>
          .
          <article-title>Convolutional networks on graphs for learning molecular fingerprints</article-title>
          . In C. Cortes,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Lawrence</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sugiyama</surname>
          </string-name>
          , and R. Garnett, editors,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>28</volume>
          , pages
          <fpage>2224</fpage>
          -
          <lpage>2232</lpage>
          . Curran Associates, Inc.,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Elton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Boukouvalas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Fuge</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Chung</surname>
          </string-name>
          .
          <article-title>Deep learning for molecular design-a review of the state of the art</article-title>
          .
          <source>Molecular Systems Design &amp; Engineering</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Fooshee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mood</surname>
          </string-name>
          , E. Gutman,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tavakoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Huynh</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. Van Vranken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and P.</given-names>
            <surname>Baldi</surname>
          </string-name>
          .
          <article-title>Deep learning for chemical reaction prediction</article-title>
          .
          <source>Molecular Systems Design &amp; Engineering</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gilmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Schoenholz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Riley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Dahl</surname>
          </string-name>
          .
          <article-title>Neural message passing for quantum chemistry</article-title>
          .
          <source>In Proceedings of the 34th International Conference on Machine Learning-</source>
          Volume
          <volume>70</volume>
          , pages
          <fpage>1263</fpage>
          -
          <lpage>1272</lpage>
          . JMLR. org,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Go</surname>
          </string-name>
          <article-title>´ mez-</article-title>
          <string-name>
            <surname>Bombarelli</surname>
            ,
            <given-names>J. N.</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Duvenaud</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Herna</surname>
          </string-name>
          <article-title>´ndez-</article-title>
          <string-name>
            <surname>Lobato</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <article-title>Sa´nchez-</article-title>
          <string-name>
            <surname>Lengeling</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Sheberla</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Aguilera-Iparraguirre</surname>
            ,
            <given-names>T. D.</given-names>
          </string-name>
          <string-name>
            <surname>Hirzel</surname>
            ,
            <given-names>R. P.</given-names>
          </string-name>
          <string-name>
            <surname>Adams</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Aspuru-Guzik</surname>
          </string-name>
          .
          <article-title>Automatic chemical design using a datadriven continuous representation of molecules</article-title>
          .
          <source>ACS central science</source>
          ,
          <volume>4</volume>
          (
          <issue>2</issue>
          ):
          <fpage>268</fpage>
          -
          <lpage>276</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hirohara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Saito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sato</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sakakibara</surname>
          </string-name>
          .
          <article-title>Convolutional neural network based on smiles representation of compounds for detecting chemical motif</article-title>
          .
          <source>BMC bioinformatics</source>
          ,
          <volume>19</volume>
          (
          <issue>19</issue>
          ):
          <fpage>526</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Irwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sterling</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Mysinger</surname>
            ,
            <given-names>E. S.</given-names>
          </string-name>
          <string-name>
            <surname>Bolstad</surname>
            , and
            <given-names>R. G.</given-names>
          </string-name>
          <string-name>
            <surname>Coleman.</surname>
          </string-name>
          <article-title>Zinc: a free tool to discover chemistry for biology</article-title>
          .
          <source>Journal of chemical information and modeling</source>
          ,
          <volume>52</volume>
          (
          <issue>7</issue>
          ):
          <fpage>1757</fpage>
          -
          <lpage>1768</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Kipf</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          .
          <article-title>Semi-supervised classification with graph convolutional networks</article-title>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Krenn</surname>
          </string-name>
          ,
          <string-name>
            <surname>F.</surname>
          </string-name>
          <article-title>Ha¨se, A</article-title>
          . Nigam,
          <string-name>
            <given-names>P.</given-names>
            <surname>Friederich</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>AspuruGuzik. Selfies</surname>
          </string-name>
          <article-title>: a robust representation of semantically constrained graphs with an example application in chemistry</article-title>
          . arXiv preprint arXiv:
          <year>1905</year>
          .13741,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kwon</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Yoon</surname>
          </string-name>
          . Deepcci:
          <article-title>End-to-end deep learning for chemical-chemical interaction prediction</article-title>
          .
          <source>In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics</source>
          , pages
          <fpage>203</fpage>
          -
          <lpage>212</lpage>
          . ACM,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ramakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. O.</given-names>
            <surname>Dral</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rupp</surname>
          </string-name>
          , and
          <string-name>
            <surname>O. A.</surname>
          </string-name>
          von Lilienfeld.
          <article-title>Quantum chemistry structures and properties of 134 kilo molecules</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ruddigkeit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Van</given-names>
            <surname>Deursen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. C.</given-names>
            <surname>Blum</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.-L.</given-names>
            <surname>Reymond</surname>
          </string-name>
          .
          <article-title>Enumeration of 166 billion organic small molecules in the chemical universe database gdb-17</article-title>
          .
          <source>Journal of chemical information and modeling</source>
          ,
          <volume>52</volume>
          (
          <issue>11</issue>
          ):
          <fpage>2864</fpage>
          -
          <lpage>2875</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schlichtkrull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Kipf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bloem</surname>
          </string-name>
          , R. Van Den Berg, I. Titov, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          .
          <article-title>Modeling relational data with graph convolutional networks</article-title>
          .
          <source>In European Semantic Web Conference</source>
          , pages
          <fpage>593</fpage>
          -
          <lpage>607</lpage>
          . Springer,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ramsundar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. N.</given-names>
            <surname>Feinberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gomes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Geniesse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Pappu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Leswing</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Pande</surname>
          </string-name>
          .
          <article-title>Moleculenet: a benchmark for molecular machine learning</article-title>
          .
          <source>Chemical science</source>
          ,
          <volume>9</volume>
          (
          <issue>2</issue>
          ):
          <fpage>513</fpage>
          -
          <lpage>530</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>