<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Advancing Materials Property Prediction: A Comparative Study of Graph Neural Network Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ioannis Papadimitriou</string-name>
          <email>i.papadimitriou@iti.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ilias Gialampoukidis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefanos Vrochidis</string-name>
          <email>stefanos@iti.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yiannis Kompatsiaris</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Information Technologies Institute, Center for Research and Technology Hellas</institution>
          ,
          <addr-line>6</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This study investigates the efficacy of Graph Neural Networks (GNNs) in predicting material properties by comparing a baseline GraphSage model with a hybrid model incorporating Convolutional Graph Neural Network (CGCNN), Graph Attention Net work (GAT), and GraphSage layers. Both models secure positions on leaderboards, but the proposed hybrid model significantly outperforms the baseline across diverse tasks. The baseline struggles in 4 out of 9 tasks, emphasizing limitations in capturing intricate dependencies. Conversely, the hybrid model consistently excels, ranking in the top 10 for 5 tasks and top 5 in the critical dielectric task. Insights highlight the importance of holistic approaches, considering structural and edge-related features. Future research aims to refine models, addressing materials science intricacies and fostering advancements in predictive accuracy. Overall, our findings contribute to the evolving landscape of materials property prediction, emphasizing the need for sophisticated models at the intersection of machine learning and materials science.</p>
      </abstract>
      <kwd-group>
        <kwd>GNNs</kwd>
        <kwd>GraphSage</kwd>
        <kwd>GAT</kwd>
        <kwd>CGCNN</kwd>
        <kwd>Matbench</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>against the existing benchmark leaderboard. Through this work, we aim to contribute nuanced insights to the
realm of generalization of material properties prediction.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>
        Materials property prediction has seen a paradigm shift with the advent of Graph Neural Networks (GNNs),
offering a unique approach to understanding complex relationships in materials datasets [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. GNNs, as a subset
of neural networks designed to handle graph structured data, have shown promise in capturing intricate
structural patterns within materials, making them particularly apt for applications in materials science [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Our methodology revolves around the careful construction and comparison of two distinct Graph Neural
Network (GNN) models, each tailored for materials property prediction on the Materials Project benchmark
datasets.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Model Architectures</title>
        <p>
          Baseline model (GraphSage): This baseline model exclusively incorporates GraphSage layers [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] (Fig. 1).
GraphSage is renowned for its ability to capture local structural information, making it a suitable candidate for
materials datasets where intricate crystallographic patterns are prevalent. GraphSage is considered a sampling
method and the main idea behind it is to uniformly sample a set of nodes from its neighborhood, aggregate the
feature information and perform graph/node classification. The model’s architecture consisted of ten
GraphSage layers and a model head of four dense layers.
        </p>
        <p>
          Proposed model (CGCNN+GAT+GraphSage): In contrast, our proposed model rep resents a more intricate
architecture, combining Convolutional Graph Neural Network (CGCNN) [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] (Fig. 2), Graph Attention Network
(GAT) [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] (Fig. 3), and GraphSage layers [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
        </p>
        <p>GAT and CGCNN layers are specifically chosen for their ability to leverage edge features (in contrast to the
GraphSage layers), enhancing the model’s capability to capture non-local dependencies and relationships within
the materials. GATs stand out for their integration of the attention mechanism, a concept extensively employed
in domains like natural language processing, into graph neural networks. The distinctive feature of GATs lies
in utilizing the attention mechanism to assign weights to nodes within a graph. This approach enables the
model to prioritize certain nodes over others during information processing, a critical aspect for effectively
capturing the intricate nuances present in graph-structured data. Moreover, regarding CGCNNs, the key idea
behind them is to adapt the convolutional operation to the irregular and non-grid nature of graphs, enabling
the model to learn and understand the inherent connectivity and relationships present in the data, both in the
node as well as the edge level. The model’s architecture included three CGCNN layers, followed by one
twoheaded and three one-headed GAT layers, leading to four GraphSage layers. The model head consisted of four
dense layers.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Featurization, Data and Setup</title>
        <p>
          The featurization of materials is a critical step in ensuring that our models encapsulate both structural and
edgerelated features. We employ the CGCNN method for featurization, utilizing the deepchem library [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], drawing
inspiration from convolutional neural networks to capture spatial relationships within crystal structures.
Starting from a typical pymatgen crystal structure form [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], the deepchem library enables the user to easily
featurize and produce graphs with a standard number of node and edge features, namely 92 and 41, respectively.
We utilise standard training techniques, such as stochastic gradient descent, and evalu ate model performance
using appropriate metrics for materials property prediction. Hyper parameters were kept the same throughout
the training and testing of all datasets, as the focus of this study was the comparative performance of the models
and not the achievement of good performance in a single task. The training set was split into training and
validation sets at 80/20 ratio as no validation set was included in the data, the training lasted for 200 epochs,
using Adam optimizer [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] with a 0.001 learning rate and a constant random seed for reproducibility. The pytorch
framework [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] was used for training and the pytorch geometric library was used for the graph dataloader
construction [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results and Discussion</title>
      <p>
        Our results present a comprehensive analysis of the performance of the baseline (GraphSage) and proposed
(CGCNN+GAT+GraphSage) models on the Materials Project benchmark datasets. The comparative evaluation
highlights the impact of incorporating edge features through CGCNN and GAT layers in the proposed model.
(a) Task: dielectric
(b) Task: jdft2d
(c) Task: gvrh
(d) Task: kvrh
(e) Task: mp-gap
(f) Task: mpe-form
(g) Task: is-metal
(h) Task: perovskites
(i) Task: phonons
Both models successfully secured positions on the leaderboards, signifying the effectiveness of Graph Neural
Networks (GNNs) in predicting material properties. However, as we scrutinize their performance, notable
distinctions emerge. The baseline GraphSage model, while making a commendable entry, consistently lags in 4
out of 9 tasks, as it is next to last in the gvrh, kvrh, mp-gap and phonons. This pattern suggests limitations in
capturing intricate dependencies within certain materials when relying solely on local node information.
In stark contrast, the proposed hybrid model significantly outperforms the baseline across all evaluated tasks.
This performance disparity underscores the advantages of integrating diverse GNN layers, especially those
capable of leveraging edge features, in materials property prediction. The consistent excellence of the proposed
model is evident as it se cures a position within the top 10 models in 5 out of 9 tasks. Notably, in the dielectric
task, the proposed model emerges among the top 5 performers, emphasizing its exceptional predictive
capabilities. Lastly, it should be noted that the proposed model outperforms the CGCNN [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] model in 6 out of 8
benchmark tasks.
      </p>
      <p>These results carry implications for materials science, affirming the relevance of machine learning, particularly
GNNs, in advancing our understanding of material behaviors. The challenges faced by the baseline model
underscore the complexities inherent in materials science tasks and highlight the need for more sophisticated
models.</p>
      <p>The successes, challenges, and observed disparities provide valuable insights for re searchers seeking to harness
the power of machine learning in the intricate realm of materials science. Future considerations may involve
fine-tuning hyperparameters, exploring interpretability, and potentially incorporating additional GNN layers
to further enhance predictive accuracy. This ongoing research aims to refine and advance the capabilities of
models in materials property prediction.</p>
      <p>In conclusion, our comprehensive examination of the baseline and proposed models sheds light on the intricate
landscape of materials science. The successes and challenges observed underscore the evolving role of machine
learning, encouraging continued exploration and refinement to unlock new possibilities in materials discovery
and understanding.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In the culmination of our study, we reflect on the insights gained from comparing the baseline GraphSage model
and the proposed hybrid model (CGCNN+GAT+GraphSage) across a spectrum of tasks within the Materials
Project benchmark datasets. Both models demonstrated their competence by securing positions on the
leaderboards, affirming the applicability of Graph Neural Networks (GNNs) in predicting material properties.
However, a nuanced examination revealed distinct performance characteristics that unveil valuable
considerations for the field of materials science.</p>
      <p>The baseline GraphSage model, although making an entry, consistently faced challenges in capturing intricate
dependencies within materials, particularly evident in its lower standings in 4 out of 9 tasks. This emphasizes
the limitations associated with relying solely on local structural information for materials property prediction.
In contrast, the proposed hybrid model emerged as a formidable solution, showcasing significant performance
advantages across all evaluated tasks. The incorporation of Convolutional Graph Neural Network (CGCNN)
and Graph Attention Network (GAT) layers, along with GraphSage layers, demonstrated a remarkable ability
to leverage edge features and non-local dependencies, thereby vastly outperforming the baseline model. The
proposed model’s consistent excellence, securing positions in the top 10 for 5 out of 9 tasks and achieving a
top5 status in the critical dielectric task, highlights its robust predictive capabilities. This success underscores the
importance of a holistic approach, considering both structural and edge-related features, in materials property
prediction.</p>
      <p>Moving forward, research will concentrate on refining hyperparameters, enhancing interpretability, and
integrating additional GNN layers to improve predictive accuracy. These efforts aim to advance machine
learning models in tackling the complexities of materials science tasks.</p>
      <p>In conclusion, our study underscores the importance of sophisticated models capable of understanding both
local and non-local relationships within materials. As we navigate the intersection of machine learning and
materials science, our findings propel ongoing exploration, fostering advancements in materials discovery and
understanding.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This work was supported by the Horizon Europe Framework Programme and the EC-funded project DiMAT
under grant agreement No 101091496.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cheng</surname>
          </string-name>
          , C. Zhang, and L. Dong, “
          <article-title>A geometric-information-enhanced crystal graph network for predicting properties of materials,” Communications Materials</article-title>
          , vol.
          <volume>2</volume>
          , no.
          <issue>1</issue>
          , p.
          <fpage>92</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dunn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ganose</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dopp</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Jain</surname>
          </string-name>
          , “
          <article-title>Benchmarking materials property prediction methods: the matbench test set and automatminer reference al gorithm</article-title>
          ,
          <source>” npj Computational Materials</source>
          , vol.
          <volume>6</volume>
          , no.
          <issue>1</issue>
          , p.
          <fpage>138</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>W.</given-names>
            <surname>Hamilton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ying</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          , “
          <article-title>Inductive representation learning on large graphs</article-title>
          ,
          <source>” Advances in neural information processing systems</source>
          , vol.
          <volume>30</volume>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Xie</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Grossman</surname>
          </string-name>
          , “
          <article-title>Crystal graph convolutional neural networks for an ac curate and interpretable prediction of material properties,” Physical review letters</article-title>
          , vol.
          <volume>120</volume>
          , no.
          <issue>14</issue>
          , p.
          <fpage>145301</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5] P. Veliˇckovi´c, G. Cucurull,
          <string-name>
            <given-names>A.</given-names>
            <surname>Casanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lio</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          , “
          <article-title>Graph attention networks</article-title>
          ,
          <source>” arXiv preprint arXiv:1710.10903</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. Juarez, and
          <string-name>
            <given-names>B. G.</given-names>
            <surname>Sumpter</surname>
          </string-name>
          , “
          <article-title>Benchmarking graph neural net works for materials chemistry</article-title>
          ,
          <source>” npj Computational Materials</source>
          , vol.
          <volume>7</volume>
          , no.
          <issue>1</issue>
          , p.
          <fpage>84</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>B. Ramsundar,</surname>
          </string-name>
          <article-title>Molecular machine learning with DeepChem</article-title>
          .
          <source>PhD thesis</source>
          , Stanford University,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S. P.</given-names>
            <surname>Ong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. D.</given-names>
            <surname>Richards</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jain</surname>
          </string-name>
          , G. Hautier,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kocher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cholia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gunter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. L.</given-names>
            <surname>Chevrier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Persson</surname>
          </string-name>
          , and G. Ceder, “
          <article-title>Python materials genomics (pymatgen): A robust, open-source python library for materials analysis</article-title>
          ,
          <source>” Computational Materials Science</source>
          , vol.
          <volume>68</volume>
          , pp.
          <fpage>314</fpage>
          -
          <lpage>319</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Ba</surname>
          </string-name>
          , “
          <article-title>Adam: A method for stochastic optimization</article-title>
          ,
          <source>” arXiv preprint arXiv:1412.6980</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paszke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bradbury</surname>
          </string-name>
          , G. Chanan,
          <string-name>
            <given-names>T.</given-names>
            <surname>Killeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gimelshein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antiga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Desmaison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kopf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>DeVito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Raison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tejani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chilamkurthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Steiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bai</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Chintala</surname>
          </string-name>
          , “
          <article-title>Pytorch: An imperative style, high-performance deep learning library</article-title>
          ,”
          <source>in Advances in Neural Information Processing Systems</source>
          <volume>32</volume>
          , pp.
          <fpage>8024</fpage>
          -
          <lpage>8035</lpage>
          , Curran Associates, Inc.,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fey</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Lenssen</surname>
          </string-name>
          , “
          <article-title>Fast graph representation learning with pytorch geometric</article-title>
          .
          <source>arxiv</source>
          <year>2019</year>
          ,” arXiv preprint arXiv:
          <year>1903</year>
          .02428,
          <year>1903</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>