<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>HexaClus: Interpretable Hexagonal Supervised Spatial Clustering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yameng Guo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Seppe vanden Broucke</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Business Informatics and Operations Management, Ghent University</institution>
          ,
          <addr-line>Tweekerkenstraat 2, 9000 Gent</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Research Centre for Information Systems Engineering, KU Leuven</institution>
          ,
          <addr-line>Naamsestraat 69, 3000 Leuven</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Geospatial inference is a critical component in many domains. Conventional methodologies such as Kriging and Geographically Weighted Regression (GWR) exhibit inherent limitations in inference capabilities, while contemporary machine learning approaches demonstrate superior predictive accuracy but significantly lack interpretability within spatial contexts. To bridge this gap, this work introduces HexaClus, a novel interpretable supervised spatial clustering framework based on mergers over a hexagonal grid that efectively resolves the tension between predictive performance and interpretability in geospatial analysis. HexaClus partitions a study area into hexagonal cells, assigns geospatial features to each cell, and iteratively merges adjacent regions based on predictive consistency to form spatially coherent clusters. A key innovation of HexaClus lies in its architecture-agnostic design, enabling seamless integration with any machine learning model. Through an empirical evaluation, we demonstrate that HexaClus substantially outperforms traditional geospatial models and black-box machine learning techniques across dual dimensions of inference capability and interpretability. The complete implementation of HexaClus is publicly accessible for reproducibility and further development1.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;geospatial clustering</kwd>
        <kwd>supervised learning</kwd>
        <kwd>geospatial interpretation</kwd>
        <kwd>estates evaluation</kwd>
        <kwd>hexagonal lattice</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Geospatial inference plays a crucial role in addressing both environmental and socio-economic issues [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1,
2, 3</xref>
        ], increasing the demand for accurate yet interpretable geospatial models. Such models enable
decision makers to analyze spatial trends, allocate resources eficiently, and mitigate potential risks.
However, balancing predictive accuracy and interpretability remains a significant challenge, particularly
in the presence of spatial dependencies and heterogeneous environments.
      </p>
      <p>
        Traditional geospatial inference methodologies, such as kriging [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ] and Geographically Weighted
Regression (GWR)[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], have been widely used to model spatial relationships by assigning diferent
weights to nearby observations. Although these methods provide solid mathematical and statistical
foundations, they face several limitations in inference capability, interpretability, and generalization.
For example, both Kriging and GWR assume stationarity, making them less efective in complex terrains
and urban environments where spatial relationships vary significantly [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. Moreover, they are
highly sensitive to parameter choices, such as the selection of the bandwidth in GWR [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and the
variogram parameters in Kriging [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], which can significantly afect the performance of constructed
models. Another major drawback is their limited ability to incorporate external geospatial features,
such as Points of Interest (POIs), which are essential for modeling human activities and land use
patterns. Furthermore, their predictive capabilities remain constrained, as they are primarily designed
for interpolation rather than robust inference in the presence of varying spatial relationships.
      </p>
      <p>
        Stronger machine learning (ML) techniques have emerged as a powerful alternative to geospatial
prediction due to their robustness and high predictive accuracy. These models can handle large-scale
datasets and complex feature interactions, making them well-suited for geospatial tasks. However,
this increase in predictive power typically comes at a cost of hampering interpretability, limiting their
utility in many practical geospatial inference settings. Ensemble models are often considered to act as
“black boxes”, making it dificult to understand why certain regions exhibit higher or lower predicted
values. Although interpretation techniques such as SHAP (Shapley Additive Explanations) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] can be
used in this context, they generally lack a geospatial-specific perspective, failing to fully capture spatial
dependencies and relationships between adjacent areas in an appealing manner.
      </p>
      <p>To address these challenges, we propose HexaClus, a novel hexagonal supervised spatial clustering
method that balances interpretability and predictive precision. Our approach starts with dividing the
study area into hexagonal cells, and then aggregates geospatial features within each cell. Next, local
supervised models are constructed per cell and are used to iteratively merge adjacent hexes until a
validation set-based convergence criterion is met. By developing this algorithm, we aim to:
• Improve interpretability of geospatial models by structuring spatial clusters, providing an intuitive
visualization of geospatial distributions and predictions.
• Develop a flexible framework that dynamically integrates local geospatial information. Unlike
ifxed-radius aggregation approaches around given target instances, our method enables the
localized incorporation of spatial embeddings on a smaller scale over the hexagonal grid, ensuring
a more accurate representation of local geospatial features.
• Remain compatible with various advanced machine learning models whilst maintaining an
interpretable clustering structure.
• Provide an interpretable and adaptive approach to geospatial inference by leveraging hexagonal
spatial partitioning and supervised clustering.</p>
      <p>An implementation of our work is made publicly available1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <sec id="sec-2-1">
        <title>2.1. Geospatial Inference Techniques</title>
        <p>Geospatial modeling relies on capturing spatial dependencies to improve inference accuracy. Existing
approaches can be broadly categorized into three main groups: statistical models, machine
learningbased methods, and deep learning-based methods.</p>
        <p>
          Traditional statistical approaches such as Kriging [
          <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
          ] and Geographically Weighted Regression
(GWR) [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] are widely used in geospatial modeling. These methods leverage spatial dependencies by
assigning weights to nearby observations. Kriging, a geostatistical interpolation technique, predicts
unknown values based on the spatial correlation of known data points. However, its efectiveness is
generally limited in heterogeneous landscapes due to the assumption of stationarity [
          <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
          ].
        </p>
        <p>
          Similarly, GWR extends traditional regression modeling by allowing coeficients to vary spatially,
ofering insights into how relationships change across diferent regions. Though GWR improves spatial
interpretability, it is sensitive to bandwidth selection [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] and struggles with high-dimensional feature
spaces. Furthermore, both Kriging and GWR have limited predictive capabilities, as they are primarily
designed for interpolation rather than robust inference.
        </p>
        <p>In practical applications, both methods also impose strict mathematical conditions, leading to complex
nonlinear computations that are often computationally expensive [12], causing these models to also be
less suitable for large-scale datasets.</p>
        <p>Next, machine learning (ML) techniques, particularly tree-based ensemble methods like Random
Forest [13] and Gradient Boosting [14, 15, 16], have gained popularity in geospatial inference due to
their robustness and predictive accuracy. These models eficiently handle non-linear relationships
and large datasets while also ofering flexibility in terms of being able to incorporate local Points of</p>
        <sec id="sec-2-1-1">
          <title>1See: https://github.com/ArmonGo/HexaClus</title>
          <p>Divide the area into hexes Merge the neighbor hex
Figure 1: Hexagonal cell clustering and merging process
Reach the convergence
The area under study is initially divided into hexagonal cells based on a specified resolution.
Then, pairs of touching polygons that yield the greatest reduction in training MSE are
identified and merged into a single polygon. This process is repeated iteratively until a
convergence condition is met.</p>
          <p>Interest (POIs). However, integrating POIs often requires extensive feature engineering [17] to query
the surrounding area around targets of interest and then appropriately aggregate POI counts. Typically,
a predefined fixed-radius approach is used to define a restricted area (in the shape of a box or circle) to
incorporate spatial context, but this may fail to capture dynamic spatial interactions efectively.</p>
          <p>A promising direction is the combination of machine learning and spatial correlation modeling.
Some hybrid approaches [18, 19] integrate both spatial statistic models and machine learning models to
improve geospatial generalization. Yet, the lack of explicit spatial structure often results in reduced
interpretability compared to traditional geostatistical models.</p>
          <p>Deep learning, in particular Graph Neural Networks (GNNs), has emerged as a promising approach
for geospatial modeling [20, 21]. GNNs represent spatial structures as graphs, enabling models to learn
complex spatial relationships. However, despite their theoretical advantages, GNN-based geospatial
inference faces significant practical challenges related to graph construction and training. First, there
is currently no standard well-accepted method for defining an appropriate spatial graph, as most
approaches rely on arbitrary distance thresholds. Second, empirical studies have shown that GNNs
can struggle to outperform classical machine learning methods in geospatial inference while requiring
significantly higher computational resources and tuning eforts [22].</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Interpretability and Visualization</title>
        <p>Compared to other inference tasks, geospatial inference places greater emphasis on interpretability
due to its involvement with complex spatial relationships, heterogeneous data sources, and real-world
decision-making processes, all of which require a high level of trust and transparency.</p>
        <p>
          A common approach to interpreting model mechanisms is assessing feature importance using methods
such as SHAP (Shapley Additive Explanations)[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and LIME (Local Interpretable Model-agnostic
Explanations)[23]. While these techniques provide an initial understanding of feature attributions—
including e.g. coordinate features—they fail to show local spatial dependencies and variations in an
appealing manner as the spatial coordinates are regarded as normal features.
        </p>
        <p>Beyond feature importance, prediction results also play a crucial role in interpretability. Since output
values are often tied to specific locations, researchers commonly aggregate raw predictions and visualize
them spatially using heatmaps or region-based summaries [24, 25] within administrative boundaries.
Additionally, methods such as GWR, Kriging, and Gaussian Processes [26] enable spatially varying
coeficients or spatial uncertainty estimation, ofering a deeper insight into local spatial variations and
uncertainties.</p>
        <p>Despite the availability of various methods for interpreting geospatial predictions, they share common
challenges. For example, most spatial visualization rely on predefined administrative boundaries, which
may not align with actual spatial variations that transcend these divisions. Furthermore, existing
map-based techniques, such as hotspot maps and density-based visualizations, are often sensitive to
data point density variations, making them susceptible to data collection bias rather than reflecting
true spatial patterns. For instance, urban areas may appear as high-risk zones simply due to there being
more instances in those areas rather than there being genuine geospatial variation.</p>
        <p>To address the previously mentioned challenges, we propose a straightforward hexagonal supervised
spatial clustering framework that provides an adaptive and interpretable solution to balance predictive
accuracy and spatial interpretability. Unlike fixed-radius approaches, our framework dynamically
integrates spatial embeddings, enabling a more context-aware representation of geospatial features.
Additionally, HexaClus supports a variety of machine learning models while preserving an interpretable
spatial structure, ensuring strong predictive performance. Most importantly, it facilitates clear geospatial
visualization and enables local variation analysis beyond administrative boundaries by structuring
spatial clusters in a hexagonal format.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The HexaClus algorithm is built on hexagonal partitions and operates by iteratively clustering geospatial
regions using a supervised hierarchical merging strategy. The cell shape of a grid system is an important
consideration. We opt for hexagonal cell shapes over e.g. square or other covering partitioning strategies
due to their appealing properties. For simplicity, it should be constant-area polygon that tiles regularly,
i.e. the triangle, the square, or the hexagon. Of these, triangles and squares have neighbors with
diferent distances. Triangles have three diferent distances, and squares have two diferent distances.
For hexagons, all neighbors are equidistant and additionally have the property of expanding rings of
neighbors approximating circles. Hexagons are also optimally space-filling. On average, a polygon
may be filled with hexagon tiles with a smaller margin of error than would be present with square tiles.
Particularly, in the context of our clustering task, the uniform structure helps reduce directional bias
and prevents potential distortions in spatial patterns when aggregating features among all neighbors.
Note that other polygonal covering strategies, such as e.g. based on Voronoi maps, could be considered
without significant efort, but are left out of scope for this work.</p>
      <p>Algorithm 1 and Figure 1 outline the steps to construct a clustering solution solution.</p>
      <p>Given a training and validation set (the latter used for stopping when convergence is reached), as
well as a set of initial (hexagonal) cells, the algorithm begins by removing all cells which do not contain
any train instances. To construct the set of initial hexagonal cells, we utilize Uber’s h3 library 2, which
can construct hexagonal cells for each location on the globe with diferent sizes, using a resolution
parameter .</p>
      <p>Next, for each initial cell, one can optionally construct a spatial representation vector containing
point of interest (POI) information, land use data information, or terrain attributes describing the initial
cell. Whilst this is a flexible component which can be either skipped or easily extended by end users,
we by default utilize the Python library SRAI to obtain a vector containing spatial information given an
initial hexagonal cell [27], similarly to what is done in e.g. Hex2Vec [28].</p>
      <p>Next, a prediction model base learner should be specified which is used to train over a set of spatially
located instanced to predict their target value. We assume target values to be continues so that regression
models are constructed, but our approach can be easily extended to a binary classification setting as
well. Before the merging procedure of cells can start, a local regression model is hence trained for each
of the initial cells. To avoid issues of overfitting and make the resulting local models more interpretable,
we use Ridge Regression as our default choice to construct the models, as it applies the 2 penalty on
large coeficients to prevent overfitting. The objective is defined as</p>
      <p>min ‖y −</p>
      <p>X ‖22 +  ‖ ‖22
(1)
where  controls regularization strength. A single regularization strength value for all constructed
models is used (which can be globally tuned if desired).</p>
      <sec id="sec-3-1">
        <title>2See: https://h3geo.org/.</title>
        <p>Next, the algorithm enters an iterative merging phase, where spatially neighboring pairs of polygons
(hexagonal cells initially) are evaluated for a potential merge. For this, a neighborhood function should
be specified which can indicate whether a pair of polygons are valid neighbors. By default, we simply
consider touching polygons to be valid. For each candidate pair, with their associated local models,
a new model is trained using the combined set of train instances, and the reduction of train loss is
utilized to select the best pair to merge. The merge is finalized, and the process continues until no
further improvement is observed on the validation loss of the solution so far as a whole for a predefined
number of iterations (early stopping with patience), ensuring that the final clusters are both spatially
coherent and predictive.</p>
        <p>Note that to obtain predictions for instances, the polygon-local model is retrieved, corresponding
with the polygon the instance falls into. Note here that by design, HexaClus does not support inference
for geospatial extrapolated points; instead, it restricts predictions to regional instances to prevent spatial
variation from afecting extrapolation performance (alternatively, a default prediction corresponding to
the mean train target value can be returned). Also remark that in the case where spatial representation
vectors where constructed for the initial hexagonal cells, the features of the given instance need to be
concatenated with the representational vector corresponding with the initial hexagonal cell the instance
falls into.</p>
        <p>Therefore, for a new unseen data point, we identify both its original hexagonal cell (before merging)
and its corresponding merged polygon, and then concatenate the representational vector derived from
the initial cell, but use the local model fitted on the merged polygon to make the prediction.</p>
        <p>Note that instead of using a strong internal leaner, we deliberately use Ridge Regression for a couple
of reasons. First, individual cells often contain a limited number of observations, making complex
models prone to overfitting, while Ridge Regression, with its L2 regularization, can mitigate this efect.
Additionally, Ridge is computationally eficient, allowing the merging process to scale efectively. Finally,
Ridge provides an interpretable framework, where feature coeficients can be analyzed per polygon
to explain spatial variations, addressing a major limitation of black-box machine learning methods in
geospatial inference. While Ridge Regression serves as the default model, HexaClus is flexible to this
regard and can integrate more complex supervised learning algorithms when desired.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <p>In this section, we compare our proposed method in terms of predictive performance across multiple
real-world datasets with the baseline default strategy of considering a single global model.</p>
      <p>We utilize three publicly available property valuation datasets covering London, New York, and Paris.
All datasets are sourced from the Kaggle data hub, with links being available on our GitHub repository
de quo. Each dataset is partitioned into a training set (70%), validation set (10%), and test set (20%).
Comparisons between our method and the global Ridge model use the same data partitions.</p>
      <p>To evaluate the performance in the setup, we compare a global Ridge Regression model with HexaClus.
For the global model, we tune the regularization strength hyperparameter  over a range of 0.1 to 0.9
in increments of 0.1. For HexaClus, we tune the hexagonal resolution parameter , varying from 5 to 9,
and do not tune the regularization strength hyperparameter  , instead simply keeping it constant to 0.1.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>Table 1 presents a comparison of RMSE (Root Mean Squared Error) between the global model and our
proposed approach across the three datasets. The results clearly demonstrate that HexaClus outperforms
the global approach on all three datasets, highlighting the efectiveness and predictive capability of our
clustering approach across diferent urban environments.</p>
      <p>In addition, we provide three visualized clustering solution maps (Figure 2, 3a, and 3b) for London,
New York, and Paris, colored based on the local predictions generated by HexaClus. These maps
Algorithm 1 HexaClus: Supervised Hexagonal Spatial Clustering
Input: ⟩ = {(, , f, )}=1: train dataset with spatial coordinates (, ), feature vectors f ∈ R, and target values 
Input: |⊑ = {( ,  , f ,  )}=1: validation dataset
Input: ℱ : regression model learner
Input: ℋ: initial set of cells (e.g. hexagonal cells with  the hexagonal resolution)
Input: max: maximum patience threshold
Input: : maximum number of epochs
Input:  (1 ∈ ℋ^ , 2 ∈ ℋ^ ) boolean neighborhood function indicating whether two polygons are valid neighbors
Input: ( ∈ ℋ): function returning optional non-empty feature vector for an initial cell
Output: ℋ^ * : optimized hexagonal partitions
Output: ℳ^ * : polygon-local models</p>
      <p>′ ← 
end for
end for
return ′
end function
function ConcatFeats(, ℋ)
′ = ∅
for  ∈ ℋ do
for (, , f, . . . ) ∈ (, ) do</p>
      <p>′ ∪ {(f | (), . . . )}
function MseLoss(, ℋ^ , ℱ^ )
for ,  ∈ ℋ^ , ℱ^ do
for (, , f, ) ∈ (, ) do</p>
      <p>←  + ((, , f) − )2
end for
end for
return /||
end function
Initialization
ℋ^ ← { | ∈ ℋ : ∃ ∈ ⟩, (, )}
^
ℳ ← {  ← ℱ</p>
      <p>( ((, ), ℋ))| ∈ ℋ^ }
Iterative Merging
 ← 0
for  = 1 to  do
ℒvparelv ←  (⊑, ℋ^ , ℱ^ )
 = ∅
for  ∈ ℋ^ , &gt; ∈ ℋ^ :  (,  ) do
* ←  ∪ 
* ← (, ) ∪ ( , )
 * ← ℱ ( (* , ℋ))
Δℒ =  (* , * ,  * )
−  ((, ), {}, {})
−  (( , ), { }, {})</p>
      <p>(,  , * ,  * , Δℒ}
 ←  ∪ {
end for
,  , * ,  * , Δℒ = arg</p>
      <p>max
(, ,* , * ,Δℒ)∈</p>
      <p>Δℒ
Update partitions: ℋ^ ← ℋ ^ ∖ {,  } ∪ {* }
Update models: ℳ^ ← ℳ ^ ∖ {,  } ∪ { * }
ℒvnaelw ←  (⊑, ℋ^ , ℱ^ )
if ℒvnaelw &lt; ℒvparelv then
 ← 0
ℋ^ * ← ℋ ^
ℳ ← ℳ ^
^ *
else</p>
      <p>←  + 1
end if
if  ≥ max then</p>
      <p>break for
end if
end for
◁ Concatenate spatial features to instance features</p>
      <p>◁ Initialize patience
◁ Initial validation loss
◁ Initialize candidate merge set
◁ Merged polygons
◁ Merged instances
◁ Construct model over merged instances
◁ Calculate train loss reduction</p>
      <p>◁ New validation loss
◁ Save the best solution so far
(a) New York property prices clustering solution
visualized with landmarks
(b) Paris property prices clustering solution visualized</p>
      <p>with landmarks
represent property prices on top of the clustered cells, where blue indicates comparatively lower prices
and red represents higher prices.</p>
      <p>Notably, the figures reveal irregular boundaries of high-price areas that do not strictly conform to
administrative regions. By adopting a hexagonal cell-based interpretation paradigm, local variations in
target values become more apparent. For instance, in London, distinct high-price areas emerge, which
might otherwise be obscured by over-smoothed summary maps.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>In this work, we proposed HexaClus, a novel interpretable hexagonal supervised spatial clustering
method that efectively balances predictive performance and interpretability in geospatial inference. By
leveraging a hexagonal grid structure, our approach dynamically merges adjacent regions based on
prediction consistency, ensuring that the final clusters reflect meaningful spatial patterns, but meanwhile
provide reliable predictions. Unlike traditional geospatial methods which rely on strong stationarity
assumptions, HexaClus provides a flexible framework to incorporate local geospatial features while
maintaining spatial interpretability. Furthermore, in contrast to black-box machine learning models,
HexaClus ofers a transparent and explainable clustering process, making it particularly useful for
geospatial decision making applications.</p>
      <p>Several areas for future improvement exist. First, optimizing the computational eficiency of the
merging process remains an important direction, particularly for large-scale datasets. Second, while we
employ Ridge Regression as a simple and interpretable local model, future iterations can incorporate
more complex models, such as tree-based methods, to improve predictive performance. Third,
considering diferent initialization strategies other than the hexagonal approach considered in this work can be
an interesting avenue for future work as well. Finally, extensive evaluation across diverse geographic
areas will be necessary to assess the generalizability of HexaClus across diferent spatial contexts.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT in order to check the grammar and
spelling. The authors reviewed the content as needed and take full responsibility for the publication’s
content.
[12] P. Harris, C. Brunsdon, A. S. Fotheringham, Links, comparisons and extensions of the
geographically weighted regression model when used as a spatial predictor, Stochastic environmental
Research and Risk assessment 25 (2011) 123–138.
[13] L. Breiman, Random forests, Machine Learning 45 (2001) 5–32.
[14] T. Chen, C. Guestrin, Xgboost: A scalable tree boosting system, in: Proceedings of the 22nd acm
sigkdd international conference on knowledge discovery and data mining, 2016, pp. 785–794.
[15] G. Ke, Q. Meng, T. Finley, T. Wang, W. Chen, W. Ma, Q. Ye, T.-Y. Liu, Lightgbm: A highly eficient
gradient boosting decision tree, in: I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus,
S. Vishwanathan, R. Garnett (Eds.), Advances in Neural Information Processing Systems, volume 30,
Curran Associates, Inc., 2017.
[16] L. Prokhorenkova, G. Gusev, A. Vorobev, A. V. Dorogush, A. Gulin, Catboost: unbiased boosting
with categorical features, Advances in neural information processing systems 31 (2018).
[17] K. Patroumpas, D. Skoutas, G. Mandilaras, G. Giannopoulos, S. Athanasiou, Exposing points of
interest as linked geospatial data, in: Proceedings of the 16th International Symposium on Spatial
and Temporal Databases, 2019, pp. 21–30.
[18] B. S. Murphy, Pykrige: development of a kriging toolkit for python, in: AGU fall meeting abstracts,
volume 2014, 2014, pp. H51K–0753.
[19] G. Erdogan Erten, M. Yavuz, C. V. Deutsch, Combination of machine learning and kriging for
spatial estimation of geological attributes, Natural Resources Research 31 (2022) 191–213.
[20] Y. Guo, S. vanden Broucke, Enhancing geospatial prediction models with feature engineering
from road networks: a graph-driven approach, International Journal of Geographical Information
Science 38 (2024) 1611–1632.
[21] W. Zhang, H. Liu, L. Zha, H. Zhu, J. Liu, D. Dou, H. Xiong, Mugrep: A multi-task hierarchical
graph representation learning framework for real estate appraisal, in: Proceedings of the 27th
ACM SIGKDD conference on knowledge discovery &amp; data mining, 2021, pp. 3937–3947.
[22] M. Geerts, J. De Weerdt, et al., Graph neural networks for house price prediction: do or don’t?,</p>
      <p>International Journal of Data Science and Analytics (2024) 1–31.
[23] M. T. Ribeiro, S. Singh, C. Guestrin, "why should I trust you?": Explaining the predictions of any
classifier, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, 2016, pp. 1135–1144.
[24] J. Yu, Z. Zhang, M. Sarwat, Geosparkviz: a scalable geospatial data visualization framework in the
apache spark ecosystem, in: Proceedings of the 30th international conference on scientific and
statistical database management, 2018, pp. 1–12.
[25] J. Hynek, J. Kachlík, V. Rusnák, Geovisto: A toolkit for generic geospatial data visualization., in:</p>
      <p>VISIGRAPP (3: IVAPP), 2021, pp. 101–111.
[26] C. K. Williams, C. E. Rasmussen, Gaussian processes for machine learning, volume 2, MIT press</p>
      <p>Cambridge, MA, 2006.
[27] P. Gramacki, K. Lesniara, K. Raczycki, S. Wozniak, M. Przymus, P. Szymanski, SRAI: towards
standardization of geospatial AI, in: Proceedings of the 6th ACM SIGSPATIAL International
Workshop on AI for Geographic Knowledge Discovery, GeoAI 2023, Hamburg, Germany, 13
November 2023, ACM, 2023, pp. 43–52.
[28] S. Woźniak, P. Szymański, Hex2vec: Context-aware embedding h3 hexagons with openstreetmap
tags, in: Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic
Knowledge Discovery, 2021, pp. 61–71.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Rai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Singh</surname>
          </string-name>
          , et al.,
          <source>Geospatial Technology for Landscape and Environmental Management</source>
          , Springer,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Thakur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramanathan</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. B. K. Prasad</surname>
          </string-name>
          , W. Gossel,
          <article-title>Geospatial techniques for managing environmental resources</article-title>
          , Springer Science &amp; Business
          <string-name>
            <surname>Media</surname>
          </string-name>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N. N.</given-names>
            <surname>Kussul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. V.</given-names>
            <surname>Sokolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. I.</given-names>
            <surname>Zyelyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. A.</given-names>
            <surname>Zelentsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. V.</given-names>
            <surname>Skakun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Shelestov</surname>
          </string-name>
          ,
          <article-title>Disaster risk assessment based on heterogeneous geospatial information</article-title>
          ,
          <source>Journal of Automation and Information Sciences</source>
          <volume>42</volume>
          (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. G.</given-names>
            <surname>Krige</surname>
          </string-name>
          ,
          <article-title>A statistical approach to some basic mine valuation problems on the witwatersrand</article-title>
          ,
          <source>Journal of the Southern African Institute of Mining and Metallurgy</source>
          <volume>52</volume>
          (
          <year>1951</year>
          )
          <fpage>119</fpage>
          -
          <lpage>139</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Matheron</surname>
          </string-name>
          , Principles of geostatistics,
          <source>Economic geology 58</source>
          (
          <year>1963</year>
          )
          <fpage>1246</fpage>
          -
          <lpage>1266</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Brunsdon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fotheringham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Charlton</surname>
          </string-name>
          ,
          <article-title>Geographically weighted regression</article-title>
          ,
          <source>Journal of the Royal Statistical Society: Series D (The Statistician)</source>
          <volume>47</volume>
          (
          <year>1998</year>
          )
          <fpage>431</fpage>
          -
          <lpage>443</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Kleijnen</surname>
          </string-name>
          ,
          <string-name>
            <surname>W. C. Van Beers</surname>
          </string-name>
          ,
          <article-title>Robustness of kriging when interpolating in random simulation with heterogeneous variances: some experiments</article-title>
          ,
          <source>European Journal of Operational Research</source>
          <volume>165</volume>
          (
          <year>2005</year>
          )
          <fpage>826</fpage>
          -
          <lpage>834</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>De Marsily</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Delay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonçalvès</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Renard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Teles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Violette</surname>
          </string-name>
          ,
          <article-title>Dealing with spatial heterogeneity</article-title>
          ,
          <source>Hydrogeology Journal</source>
          <volume>13</volume>
          (
          <year>2005</year>
          )
          <fpage>161</fpage>
          -
          <lpage>183</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ma</surname>
          </string-name>
          , L. Zhang,
          <article-title>Comparison of bandwidth selection in application of geographically weighted regression: a case study</article-title>
          ,
          <source>Canadian Journal of Forest Research</source>
          <volume>38</volume>
          (
          <year>2008</year>
          )
          <fpage>2526</fpage>
          -
          <lpage>2534</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Toal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. W.</given-names>
            <surname>Bresslof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Keane</surname>
          </string-name>
          ,
          <article-title>Kriging hyperparameter tuning strategies</article-title>
          ,
          <source>AIAA journal 46</source>
          (
          <year>2008</year>
          )
          <fpage>1240</fpage>
          -
          <lpage>1252</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A unified approach to interpreting model predictions</article-title>
          , in: I. Guyon,
          <string-name>
            <given-names>U. V.</given-names>
            <surname>Luxburg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wallach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vishwanathan</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2017</year>
          , pp.
          <fpage>4765</fpage>
          -
          <lpage>4774</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>