<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>IEEE Transactions on Image Processing</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1109/TIP</article-id>
      <title-group>
        <article-title>Approaches for Deep Learning-Based 2-m Temperature Downscaling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Antigoni Moira</string-name>
          <email>a.moira@ipta.demokritos.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stelios Karozis</string-name>
          <email>skarozis@ipta.demokritos.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Theodoros Giannakopoulos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Efrosyni</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karakitsou</string-name>
          <email>e.karakitsou@ipta.demokritos.gr</email>
          <email>ts@ipta.demokritos.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolaos Gounaris</string-name>
          <email>gounaris@ipta.demokritos.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Athanasios Sfetsos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Climate Downscaling</institution>
          ,
          <addr-line>T2m, Elevation, DEM, Super-resolution, EDSR, Deep Learning</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Computational Intelligence Laboratory, Institute of Informatics and Telecommunications, National Centre of Scientific Research</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Demokritos</institution>
          ,
          <addr-line>15341 Agia Paraskevi</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Environmental Research Laboratory, Institute of Nuclear &amp; Radiological Sciences &amp; Technology, Energy &amp; Safety, National Centre</institution>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>of Scientific Research Demokritos</institution>
          ,
          <addr-line>15341 Agia Paraskevi</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>13</volume>
      <issue>2004</issue>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Global Climate Models (GCMs) provide valuable climate projections, but operate at coarse spatial resolutions, limiting their usefulness for local-scale applications. Downscaling techniques are therefore essential to bridge this gap. This study investigates how the integration of elevation data can improve the performance of CNN-based architecture deep learning models to downscale the near-surface air temperature (T2m) from a 0.5°×0.5° grid to a 0.25°×0.25° resolution. We evaluate diferent elevation data integration strategies and demonstrate their impact on downscaling efectiveness, highlighting the role of terrain-related features in refining temperature estimates.</p>
      </abstract>
      <kwd-group>
        <kwd>Downscaling</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>CEUR
Workshop</p>
      <p>ISSN1613-0073
data can include latitude and longitude (especially for large regions covering diferent climate zones),
terrain characteristics (such as elevation and land use/cover), seasonal patterns, and more.</p>
      <p>
        In ”Deep learning downscaled high-resolution daily near surface meteorological datasets over East Asia”
Lin etal. (2023) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] developed a bias correction and downscaling approach, named BC-UNet, to construct
a Climate Change for East Asia with Bias Corrected UNet Dataset (CLIMEA-BCUD) based on CMIP6.
The BC-UNet downscaling approach firstly applied Quantile Delta Mapping (QDM) to correct CMIP6
models and then the UNet is trained for downscaling the biased corrected GCM dataset.
CLIMEABCUD provides nine meteorological variables including 2-m air temperature, 2-m daily maximum and
minimun air temperature, precipitation, 10-m wind speed, 2-m relative humidity, 2-m specific humidity,
downward shortwave radiation and downward longwave radiation with 0.1° horizontal resolution at
daily intervals over the historical period of 1950–2014 and three future scenarios (SSP1-2.6, SSP2-4.5
and SSP5-8.5) of 2015–2100. In their implementation, they used static elevation data from the Global 30
Arc-Second Elevation dataset as an additional input channel to the model, alongside the other input
data, representing the simplest and most direct way to integrate elevation into the model.
      </p>
      <p>
        In their study, Sha et al.(2020) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] modify UNet, a semantic-segmentation CNN, and apply it to the
downscaling of daily maximum/minimum 2-m temperature (TMAX/TMIN) over the western continental
United States from 0.258° to 4-km grid spacings. They selected HR elevation, LR elevation and LR
TMAX/TMIN as inputs and trained UNet using Parameter–Elevation Regressions on Independent
Slopes Model (PRISM) data over the south- and central-western United States from 2015 to 2018. They
found that the original UNet cannot generate enough fine-grained spatial details when transferred
to the new northwestern U.S. domain. In response, they modified the original UNet by assigning an
extra HR elevation output branch/loss function and training the modified UNet to reproduce both
the supervised HR TMAX/TMIN and the unsupervised HR elevation. They named this improvement
”UNet-Autoencoder(AE).” The UNet-AE showed better gridpoint-level performance with more than 10%
mean absolute error (MAE) reduction relative to the original UNet.
      </p>
      <p>
        In the work of Bhakare et al.(2025) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], three machine learning models examined and specifically,
Artificial Neural Network (ANN), Random Forest (RF), and Convolutional Neural Network (CNN), for
downscaling of seasonal forecasts of daily minimum temperature from 12 km to 250 m horizontal
resolution. The study area was part of the Trentino-Alto Adige region in northern Italy, characterized
by complex terrain (elevations ranging from 166 m to 2628 m) and varied local microclimatic
conditions. Data with spatial resolution 0.125° × 0.125° (approximately 12 km) from the fith-generation
Seasonal Forecast System SEAS5 operated by the European Centre for Medium-Range Weather Forecasts
(ECMWF) used as predictors, while the target was the daily minimum temperature on a 250 m grid
developed by interpolating more than 200 station observations in Trentino—South Tyrol. Elevation
data from NASA’s Shuttle Radar Topography Mission were also used as a static predictor, having been
spatially aggregated to achieve approximately 250 m resolution, matching the pixel size of the target
data. ANN and the RF model were provided with 9 dynamic features and 1 static feature (elevation),
while the CNN model is provided with 9 dynamic features without explicitly providing elevation as an
additional feature, relying on CNN’s ability to understand spatial features on its own. Results suggest
that CNN outperforms ANN and RF, achieving lower Mean Absolute Error (MAE) (1.59 °C to 2.03 °C).
In a previous work by Bhakare et al. (2024) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], where they attempted downscaling of daily mean
temperature for the same region but with a diferent resolution step (from 9 km ERA5-Land reanalysis
to 1 km), they also described how they derived a few more static features from the elevation, such as
slope, aspect, cross-sectional curvature (C-curv), and longitudinal curvature (L-curv), in order to obtain
more auxiliary predictors. The schematic of these features is shown in Figure 1.
      </p>
      <p>The aim of this study is to examine whether the integration of elevation data can enhance the
performance of a CNN-based DL model for downscaling T2m from a 0.5°×0.5° grid to a 0.25°×0.25°
resolution. Several elevation integration strategies, derived from the related literature, as well as their
combinations, were tested to identify the most efective approach for the given task.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>The Enhanced Deep Super-Resolution (EDSR) network, a widely used CNN architecture for improving
spatial resolution, was selected for this study. An initial baseline model, excluding elevation data, was
ifrst evaluated. Next, three elevation integration strategies were implemented and tested to assess
their impact on downscaling performance and to identify the most efective architecture. Finally, an
ablation study was conducted on the two selected elevation-derived features (aspect and slope) and
their combinations with elevation to examine whether they improved performance.</p>
      <sec id="sec-2-1">
        <title>2.1. Model Architecture</title>
        <p>
          The EDSR network is a high-performing CNN model commonly applied to image super-resolution tasks.
It builds upon residual learning by removing unnecessary components, such as batch normalization
layers, which were shown to hinder performance in image restoration tasks. EDSR utilizes a very deep
residual network with a simplified structure and greater model capacity, allowing it to learn more
powerful mappings from low-resolution to HR images[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>To conserve computational resources, a simplified EDSR model is used in this study, as the focus
is on comparing elevation integration strategies rather than model architecture. The model includes
four residual blocks, each consisting of two 3×3 convolutional layers with LeakyReLU activation and a
scaling factor (0.1) to stabilize training. It begins with an initial convolutional layer, followed by the four
residual blocks, and includes a skip connection that adds the initial features back to the processed ones.
After feature extraction, an upsampling layer is applied to double the resolution of the input image.
This architecture, shown in Figure 2, is used in combination with the elevation integration strategies
described below.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Elevation Integration Strategies</title>
        <p>
          Informed by the literature review presented in the Introduction section, which suggests that
incorporating elevation information from Digital Elevation Models (DEMs) can positively influence prediction
accuracy, the following approaches were explored to further enhance the performance of the EDSR
model:
1. Early Fusion: The first approach involved adding elevation data as a second input channel
alongside the primary input, allowing the model to process spatial and elevation features in
parallel. However, this method required the DEM data to be upscaled to match the resolution of
the input grid, which may result in a loss of fine-grained elevation details.
2. Late Fusion: In an attempt to address the low-resolution limitation of early fusion, a second
approach was tested. In this case, HR DEM data was introduced after the upsampling stage of
the model as a second channel to the intermediate output. This was followed by two fusion
convolutional layers that incorporated the elevation information into the final prediction. This
method aimed to preserve the spatial richness of the DEM and enable more meaningful integration
into the model’s output.
3. Early &amp; Late Fusion Combination: Since each of the above strategies has its own limitations,
a third model was developed that combines both approaches simultaneously. The architecture of
this hybrid model is illustrated in Figure 2. This integrated design yielded the most significant
improvement in performance, as demonstrated in the Results section.
4. Elevation &amp; Elevation-derived Features Fusion: Finally, inspired by the approach in the
literature review study by Bhakare et al. (2024) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], additional static features were extracted
from the available DEM data. Specifically, slope and aspect were derived and incorporated as
extra input channels alongside elevation, both in the early and late fusion stages.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Setup</title>
      <p>This section outlines the procedures and resources used to conduct the experiments. It describes the
dataset and study area, the preprocessing steps applied to the input and target data, the evaluation
metrics used to assess model performance and the implementation details of the deep learning models.</p>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>
          The source of ground truth data used to train the DL models is the ERA5 dataset [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], available through
the Copernicus Climate Data Store. ERA5 is the fith-generation reanalysis product, by the ECMWF,
providing global climate and weather data from 1940 to the present at hourly temporal resolution.
Specifically, the T2m variable is used, representing the air temperature at 2 meters above the surface of
land and sea. In addition, the Land-Sea Mask parameter is employed to compute certain evaluation
metrics over land areas, where temperature variability is more noticeable on both a daily and seasonal
basis. This parameter indicates the proportion of land within each grid cell (ranging from 0 to 1) and is
a dimensionless value that distinguishes land from ocean or inland water bodies such as lakes, rivers,
reservoirs and coastal waters.
        </p>
        <p>The present study focuses on a broad region of the Northern Hemisphere, selected to encompass all
four major climate zones (tropical, subtropical, temperate, and polar), as well as a range of topographic
complexity. The study domain extends from the vicinity of Greenland to the equator and the Indian
Ocean, covering mountainous regions and relatively flat terrain. GRIB files were obtained from the
Copernicus Climate Data Store for the period 2000 to 2020, with a temporal resolution of 6 hours (00:00,
06:00, 12:00 and 18:00). The spatial extent spans latitudes from 80.0°N to 0.0°N and longitudes from
60.0°W to 85.0°E, capturing substantial climatic and topographic variability across the region.</p>
        <p>
          The training input consists of ERA5 reanalysis data paired with altitude information from the U.S.
Geological Survey 3D Elevation Program DEM [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], integrated using various strategies to enhance the
performance of the deep learning models. The original spatial resolution of 0.25° × 0.25° was upscaled
to 0.5° × 0.5° using bicubic interpolation for the T2m and DEM grids, in order to generate the LR dataset.
This LR dataset serves as the input for the deep learning model, which is trained to reconstruct the
corresponding HR target data from the original ERA5 reanalysis.
        </p>
        <p>Finally, two additional static features were derived from the available elevation data: slope and aspect.
Aspect refers to the compass direction a slope faces and is measured in radians from north. These
features were included at both low and high resolutions alongside elevation, aiming to enhance the
model’s performance.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Data Preprocessing</title>
        <p>Before evaluating the elevation integration strategies described in the Methodology section, the dataset
was uniformly preprocessed to ensure a fair comparison of results. The data was first standardized and
then partitioned into three subsets: 70% for training, 15% for validation, and 15% for testing.</p>
        <p>For standardization, the z-score normalization technique was applied, transforming the data, so that
most values falling within the range of -1 to 1. This required calculating the global mean and standard
deviation across the entire dataset, yielding a mean of 286.30798 and a standard deviation of 64.25680.
The z-score was computed for each grid cell using the formula:</p>
        <p>( −  )
 =</p>
        <p />
        <p>Finally, to eliminate temporal dependencies and ensure randomness, the data was shufled prior to
splitting, disrupting its original chronological order. A fixed random seed was also used to guarantee
reproducibility, enabling consistent data partitioning across diferent experiments and ensuring a fair
comparison of integration strategies.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Evaluation Metrics and Loss Function</title>
        <p>To evaluate the performance of the investigated elevation integration methods for temperature grid
downscaling, four statistical metrics were employed: Mean Absolute Error (MAE), Mean Squared Error
(MSE), Peak Signal-to-Noise Ratio (PSNR), and Structural Similarity Index (SSIM). Since this task can
be framed as a single-image super-resolution (SISR) problem, these metrics efectively capture both
pixel-wise accuracy and structural fidelity.</p>
        <p>
          MSE, the most commonly used full-reference quality metric, quantifies the average squared diferences
between predicted and actual temperature values, emphasizing larger errors. PSNR measures the ratio
between the maximum possible signal and the level of reconstruction error, providing a logarithmic
assessment of prediction quality. To assess perceptual quality, SSIM is included, which evaluates
structural similarity based on luminance, contrast, and spatial texture [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Finally, MAE ofers an
intuitive and interpretable measure of the average absolute error, making it especially useful for
real-world temperature applications. Together, these metrics provide a comprehensive assessment of
numerical accuracy and spatial structure preservation.
        </p>
        <p>
          For model training, MAE was selected as the loss function instead of MSE. Also known as L1 loss,
MAE computes the average of the absolute diferences between predicted and actual values [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Unlike
MSE, which squares these diferences and disproportionately penalizes larger errors, MAE treats all
errors linearly, making it less sensitive to outliers [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. This characteristic is particularly beneficial in
temperature downscaling, where localized extremes or noise could otherwise distort the learning process.
Moreover, MAE aligns more closely with real-world temperature interpretation, as it retains the original
units (e.g., kelvin) and directly reflects the average deviation [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. This improves the interpretability of
the model’s performance and ensures that the objective remains focused on minimizing overall error
across all regions, rather than overemphasizing extreme discrepancies.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Implementation Details</title>
        <p>During this study, DL models were optimized using the Adam optimizer. This choice is based on the fact
that Adam is computationally eficient, has little memory requirement, is invariant to diagonal rescaling
of gradients, and is well suited for problems that are large in terms of data or parameters. Although
Adam dynamically adapts the learning rate during the training process, training can sometimes stall or
converge slowly. Therefore, a ReduceLROnPlateau scheduler was applied to decrease the learning rate
when the loss function stops improving for 70 epochs. A batch size of 32 was used and the models were
trained for 250 epochs. Finally, the early stopping technique was employed, triggered after 100 epochs
without improvement in model performance.</p>
        <p>
          All data preprocessing and the training of the DL models were performed using Google Colab
notebooks, utilizing the A100 GPU with the high-RAM runtime option enabled. The implementation
of the models was carried out using the TensorFlow [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] framework, while for data preprocessing
tasks and data splitting, the scikit-learn (sklearn) library [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] was employed. Visualization of training
progress, model performance metrics and output comparisons was performed using Matplotlib [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Post-Processing and Model Evaluation</title>
        <p>
          To efectively evaluate each method, Land MAE was calculated in both normalized units and degrees
Celsius to provide a meaningful interpretation of the temperature diferences. Since temperature is
strongly afected by terrain and can vary significantly over land throughout the day, MAE was calculated
specifically for land areas to provide a more meaningful evaluation. This approach helps avoid bias
introduced by large water bodies, such as oceans, which have high heat capacity and tend to moderate
temperature changes[
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. Including these regions in the evaluation could underestimate errors, as
temperature fluctuation over water is generally much lower than over land.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In this section, the results obtained from each approach in the study are presented and analyzed.
The performance of the diferent methods is evaluated using the selected metrics, enabling a direct
comparison. First, a simple EDSR model was trained without incorporating any DEM data, serving as
the baseline. Subsequently, each of the elevation integration methods described in the Methodology
section was tested to enhance the baseline model’s performance. Specifically, the Early Fusion model
integrates elevation as an additional input channel alongside the low-resolution T2m data. The Late
Fusion model tests the efect of incorporating high-resolution elevation after upscaling, by fusing it as a
second channel just before the final output. The Early &amp; Late Fusion approach combines both strategies
simultaneously.</p>
      <p>Figure 4 presents the training history and validation loss for each of the four models, providing a
visual comparison and facilitating interpretation of their efectiveness. The baseline shows relatively
stable convergence but higher loss compared to all elevation-integrated strategies. This indicates that
incorporating elevation data improves model performance. Among the tested strategies, Early &amp; Late
Fusion achieved the lowest overall loss curves. In this case, the learning rate decreased later (at epoch
212) compared to the other strategies, suggesting that the model continued to learn efectively for
more epochs before requiring a reduction. This delayed activation of the learning rate scheduler is
advantageous, as it indicates that the model maintained steady improvements over a longer period,
ultimately leading to better convergence.</p>
      <p>The results of the baseline and elevation integration approaches are summarized in Table 1. While
the diferences in the metrics are not very large, it is clear that the integration of elevation has a positive
efect on performance. The small variations in the metrics are likely due to the fact that the study area
is very large (all metrics are averaged over the 185,600 cells of the grid), and a significant portion of
it consists of sea, where, as shown in Figure 6, the MAE is close to zero. For this reason, it was also
deemed necessary to calculate the MAE over land only. In the land-only metrics, the improvement is
more noticeable, with the mean MAE decreasing by 0.03°C when using the Early &amp; Late Fusion method.
Of particular importance is the 0.3°C reduction observed in the maximum MAE. Qualitatively, this
indicates that even in the worst-case scenarios, this architecture deviates from the true value by only
about 1°C on average—significantly lower than the roughly 1.3°C deviation observed in the No DEM
variant.</p>
      <sec id="sec-4-1">
        <title>Trainable params ↓</title>
        <p>No DEM</p>
        <p>After establishing Early &amp; Late Fusion as the most efective architecture, additional static features
derived from the DEM (slope and aspect), were tested to further improve performance. In this analysis,
elevation was first substituted with aspect and slope individually, followed by an evaluation of their
combined use with elevation. Figure 5 presents the training history and validation loss for each of
the four models. All feature configurations display stable convergence with similar loss trajectories.
However, the Elevation &amp; Slope combination achieves slightly lower overall losses and triggers a
reduction in the learning rate at a later stage, indicating that the model sustained efective learning
over a longer training period.</p>
        <p>The results of this ablation study are summarized in Table 2. According to the evaluation metrics,
the diferences between the approaches are minimal, with the Elevation &amp; Slope combination showing
a slight improvement in most cases. All configurations achieve results comparable to those reported in
Table 1 for elevation alone, suggesting that the model extracts similar information from elevation and
its derived features. Nevertheless, the Elevation &amp; Slope combination yields marginal but consistent
improvements in most metrics and is therefore recommended as the most efective approach.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Trainable params ↓</title>
        <p>Aspect
0.00141
0.000008</p>
        <p>Figure 6 shows the average per-pixel MAE in °C on the test dataset for both the No Elevation Aware
baseline model and the best-performing Elevation and Slope Fusion model. In both cases, errors are
more prominent in temperate zones and regions with complex terrain, where temperature variations
tend to be greater. The baseline model, in particular, exhibits higher errors in mountainous areas such
as the Alps, Caucasus, Ethiopian Highlands, and Himalayas, highlighting its dificulty in capturing
temperature variability in such regions. While the proposed method does not eliminate these errors
entirely, the figure demonstrates a noticeable reduction in their value.</p>
        <p>Finally, Figure 7 illustrates an example output from best-performing Elevation and Slope Fusion
model, alongside the LR input data and the expected HR target for comparison. To highlight the
diferences more clearly, a zoomed-in view of the wider European region has been provided.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion &amp; Conclusions</title>
      <p>This study presented an efective method for downscaling climate data (both in terms of performance
and inference time) capable of doubling the spatial resolution (from a 0.5°×0.5° grid to 0.25°×0.25°),
while maintaining a low average error of approximately 0.14°C per cell. The main focus of the research
was the integration of geospatial information (elevation and elevation-devired static features) into the
DL model. This approach enhanced the accuracy of the downscaled outputs, particularly in complex
terrains such as mountainous regions, where traditional methods tend to perform poorly. The study
shows that this relatively simple yet targeted intervention shift the problem from a generic SISR task
toward a domain-informed climate application. By doing so, the model is better able to capture spatial
variability in temperature that is influenced by underlying geographic and topographic features. This
not only improves performance but also increases the physical interpretability of the results, which is
considered an important consideration in climate modeling.</p>
      <p>
        While the proposed strategy has demonstrated promising results in improving the spatial resolution
of temperature data using DL and geospatial augmentation, several directions remain open for further
development. The following key points are suggested for further investigation:
1. Finer-scale downscaling: The current elevation integration strategy may be even more
beneficial for higher spatial resolution downscaling, as HR DEM data (which are available at scales of a
few meters) can help the model capture finer local variability, particularly in areas with complex
topography.
2. Exploration and evaluation of diferent static feature integration across varying
resolutions: When downscaling is performed in multiple steps, diferent static features can be
incorporated at each stage, depending on the factors that influence climate at the corresponding
spatial scale. For instance, when downscaling from 250 km to 100 km, broad features such as
land–sea masks may be more relevant. At intermediate resolutions (e.g., 100 km to 25 km), factors
like latitude or season may play a larger role. Finally, for resolutions below 25 km, more detailed
and information-rich features, such as elevation, land cover or land use, become increasingly
important for capturing fine-scale temperature variations.
3. Expansion to more climate variables: The proposed elevation integration strategy could also
be applied to other climate variables, such as precipitation and wind. Extending the approach
in this way would help assess whether it can preserve physical consistency between variables,
while exploring the feasibility of multi-variable downscaling.
4. Model architecture optimizations: As previously mentioned, this study focuses on exploring
diferent strategies for integrating DEM data to enhance the performance of the deep learning
model, rather than modifying the model architecture itself. However, architectural improvements
(such as employing a deeper EDSR model) could increase the network’s capacity to learn complex
patterns. Additionally, testing more advanced architectures with attention mechanisms, such as
RCAN, could be a potential direction to scale up the approach.
5. Use of temporal models: In this study, the temporal dimension was entirely ignored and
training, validation and testing sets were shufled to disrupt their temporal continuity. In future
work, the integration of temporal information could be explored using models designed to capture
sequential dependencies, such as ConvLSTMs [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. Since temperature patterns exhibit temporal
correlations, incorporating such models may improve downscaling accuracy by capturing seasonal
or daily trends that static models do not represent.
      </p>
      <p>These directions highlight the potential for both methodological and practical improvements in this
climate data downscaling method.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT in order to: Grammar and spelling
check, reword, drafting content and improve writing style. After using this tool, the authors reviewed
and edited the content as needed and take full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-7">
      <title>Funding</title>
      <p>This research was funded by the New Enabling Visions and Tools for End-useRs and stakeholders
thanks to a common MOdeling approach towards a ClimatE neutral and resilient society (NEVERMORE)
project, Grant agreement ID: 101056858 and the Adaptation solutions to reduce climate change impact
on health in the Mountain area (MOUNTADAPT) project, Grant agreement ID: 101155958.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>N.</given-names>
            <surname>Politi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vlachogiannis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sfetsos</surname>
          </string-name>
          , P. T. Nastos,
          <article-title>High-resolution dynamical downscaling of era-interim temperature and precipitation using wrf model for greece</article-title>
          ,
          <source>Climate Dynamics</source>
          <volume>57</volume>
          (
          <year>2021</year>
          )
          <fpage>799</fpage>
          -
          <lpage>825</lpage>
          . URL: http://dx.doi.org/10.1007/s00382-021-05741-9. doi:
          <volume>10</volume>
          .1007/s00382- 021- 05741- 9.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Tomasi</surname>
          </string-name>
          , G. Franch,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cristoforetti</surname>
          </string-name>
          ,
          <article-title>Can ai be enabled to perform dynamical downscaling? a latent difusion model to mimic kilometer-scale cosmo5.0_clm9 simulations</article-title>
          ,
          <source>Geoscientific Model Development</source>
          <volume>18</volume>
          (
          <year>2025</year>
          )
          <fpage>2051</fpage>
          -
          <lpage>2078</lpage>
          . URL: http://dx.doi.org/10.5194/gmd-18-2051-2025. doi:
          <volume>10</volume>
          .5194/ gmd- 18- 2051-
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. P.</given-names>
            <surname>Sahu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Bazal</surname>
          </string-name>
          ,
          <article-title>Literature review on single image super resolution</article-title>
          ,
          <source>International Journal of Trend in Scientific Research and Development</source>
          Volume-
          <volume>2</volume>
          (
          <year>2018</year>
          )
          <fpage>2485</fpage>
          -
          <lpage>2490</lpage>
          . URL: http: //dx.doi.org/10.31142/ijtsrd18339. doi:
          <volume>10</volume>
          .31142/ijtsrd18339.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ren</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>C.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <article-title>Deep learning in statistical downscaling for deriving high spatial resolution gridded meteorological data: A systematic review</article-title>
          ,
          <source>ISPRS Journal of Photogrammetry and Remote Sensing</source>
          <volume>208</volume>
          (
          <year>2024</year>
          )
          <fpage>14</fpage>
          -
          <lpage>38</lpage>
          . URL: http://dx.doi.org/10.1016/j.isprsjprs.
          <year>2023</year>
          .
          <volume>12</volume>
          .011. doi:
          <volume>10</volume>
          .1016/j.isprsjprs.
          <year>2023</year>
          .
          <volume>12</volume>
          .011.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N.</given-names>
            <surname>Rampal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hobeichi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. B.</given-names>
            <surname>Gibson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baño-Medina</surname>
          </string-name>
          , G. Abramowitz,
          <string-name>
            <given-names>T.</given-names>
            <surname>Beucler</surname>
          </string-name>
          , J. GonzálezAbad, W. Chapman,
          <string-name>
            <given-names>P.</given-names>
            <surname>Harder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Gutiérrez</surname>
          </string-name>
          ,
          <article-title>Enhancing regional climate downscaling through advances in machine learning</article-title>
          ,
          <source>Artificial Intelligence for the Earth Systems</source>
          <volume>3</volume>
          (
          <year>2024</year>
          ). URL: http: //dx.doi.org/10.1175/aies-d
          <string-name>
            <surname>-</surname>
          </string-name>
          23
          <source>-0066</source>
          .1. doi:
          <volume>10</volume>
          .1175/aies- d
          <string-name>
            <surname>-</surname>
          </string-name>
          23
          <source>- 0066</source>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N. M.</given-names>
            <surname>Pawar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Soltanmohammadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Mahjour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Faroughi</surname>
          </string-name>
          ,
          <article-title>Esm data downscaling: a comparison of super-resolution deep learning models</article-title>
          ,
          <source>Earth Science Informatics</source>
          <volume>17</volume>
          (
          <year>2024</year>
          )
          <fpage>3511</fpage>
          -
          <lpage>3528</lpage>
          . URL: http://dx.doi.org/10.1007/s12145-024-01357-9. doi:
          <volume>10</volume>
          .1007/s12145- 024- 01357- 9.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>F.</given-names>
            <surname>Merizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Asperti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Colamonaco</surname>
          </string-name>
          ,
          <article-title>Wind speed super-resolution and validation: from era5 to cerra via difusion models</article-title>
          ,
          <source>Neural Computing and Applications</source>
          <volume>36</volume>
          (
          <year>2024</year>
          )
          <fpage>21899</fpage>
          -
          <lpage>21921</lpage>
          . URL: http://dx.doi.org/10.1007/s00521-024-10139-9. doi:
          <volume>10</volume>
          .1007/s00521- 024- 10139- 9.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          , G. Dong,
          <article-title>Deep learning downscaled high-resolution daily near surface meteorological datasets over east asia</article-title>
          ,
          <source>Scientific Data</source>
          <volume>10</volume>
          (
          <year>2023</year>
          ). URL: http://dx.doi.org/10. 1038/s41597-023-02805-9. doi:
          <volume>10</volume>
          .1038/s41597- 023- 02805- 9.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sha</surname>
          </string-name>
          , D. J.
          <string-name>
            <surname>Gagne</surname>
            <given-names>II</given-names>
          </string-name>
          , G. West,
          <string-name>
            <given-names>R.</given-names>
            <surname>Stull</surname>
          </string-name>
          ,
          <article-title>Deep-learning-based gridded downscaling of surface meteorological variables in complex terrain. part i: Daily maximum and minimum 2-m temperature</article-title>
          ,
          <source>Journal of Applied Meteorology and Climatology</source>
          <volume>59</volume>
          (
          <year>2020</year>
          )
          <fpage>2057</fpage>
          -
          <lpage>2073</lpage>
          . URL: http://dx.doi.org/10. 1175/
          <string-name>
            <surname>JAMC-D-</surname>
          </string-name>
          20-
          <issue>0057</issue>
          .1. doi:
          <volume>10</volume>
          .1175/jamc- d
          <string-name>
            <surname>-</surname>
          </string-name>
          20
          <source>- 0057</source>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhakare</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Crespi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zardi</surname>
          </string-name>
          ,
          <article-title>Spatial downscaling of daily temperature minima using machine learning methods and application to frost forecasting in two alpine valleys</article-title>
          ,
          <source>Atmosphere</source>
          <volume>16</volume>
          (
          <year>2025</year>
          )
          <article-title>38</article-title>
          . URL: http://dx.doi.org/10.3390/atmos16010038. doi:
          <volume>10</volume>
          .3390/atmos16010038.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhakare</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Dal</given-names>
            <surname>Gesso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Venturini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zardi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Trentini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Petitta</surname>
          </string-name>
          ,
          <article-title>Intercomparison of machine learning models for spatial downscaling of daily mean temperature in complex terrain</article-title>
          ,
          <source>Atmosphere</source>
          <volume>15</volume>
          (
          <year>2024</year>
          )
          <article-title>1085</article-title>
          . URL: http://dx.doi.org/10.3390/atmos15091085. doi:
          <volume>10</volume>
          .3390/atmos15091085.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Son</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. M.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Enhanced deep residual networks for single image superresolution</article-title>
          ,
          <year>2017</year>
          . URL: http://dx.doi.org/10.1109/cvprw.
          <year>2017</year>
          .
          <volume>151</volume>
          . doi:
          <volume>10</volume>
          .1109/cvprw.
          <year>2017</year>
          .
          <volume>151</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>C. C. C.</surname>
          </string-name>
          <article-title>Service, Era5 hourly data on pressure levels from 1940 to present</article-title>
          ,
          <year>2018</year>
          . URL: https: //cds.climate.copernicus.eu/doi/10.24381/cds.bd0915c6. doi:
          <volume>10</volume>
          .24381/CDS.BD0915C6.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>U.S.</given-names>
            <surname>Geological</surname>
          </string-name>
          <string-name>
            <surname>Survey</surname>
          </string-name>
          ,
          <source>The 3d elevation program (3dep)</source>
          , https://www.usgs.
          <source>gov/ core-science-systems/ngp/3dep</source>
          ,
          <year>2023</year>
          . https://www.usgs.
          <source>gov/core-science-systems/ngp/3dep.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bovik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sheikh</surname>
          </string-name>
          , E. Simoncelli,
          <article-title>Image quality assessment: from error visibility to</article-title>
          <year>2003</year>
          .
          <volume>819861</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Deep learning for weather forecasting: A cnn-lstm hybrid model for predicting historical temperature data (</article-title>
          <year>2024</year>
          ). URL: https://arxiv.org/abs/2410.14963. doi:
          <volume>10</volume>
          .48550/ARXIV.2410.14963.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>C.</given-names>
            <surname>Willmott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Matsuura</surname>
          </string-name>
          ,
          <article-title>Advantages of the mean absolute error (mae) over the root mean square error (rmse) in assessing average model performance</article-title>
          ,
          <source>Climate Research</source>
          <volume>30</volume>
          (
          <year>2005</year>
          )
          <fpage>79</fpage>
          -
          <lpage>82</lpage>
          . URL: http://dx.doi.org/10.3354/cr030079. doi:
          <volume>10</volume>
          .3354/cr030079.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Willmott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Matsuura</surname>
          </string-name>
          ,
          <article-title>On the use of dimensioned measures of error to evaluate the performance of spatial interpolators</article-title>
          ,
          <source>International journal of geographical information science 20</source>
          (
          <year>2006</year>
          )
          <fpage>89</fpage>
          -
          <lpage>102</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>A</article-title>
          . et al,
          <source>TensorFlow: Large-scale machine learning on heterogeneous systems</source>
          ,
          <year>2015</year>
          . URL: https://www.tensorflow.org/, software available from tensorflow.
          <source>org.</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>P.</surname>
          </string-name>
          et al,
          <article-title>Scikit-learn: Machine learning in python</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>Matplotlib: A 2d graphics environment</article-title>
          ,
          <source>Computing in Science &amp; Engineering</source>
          <volume>9</volume>
          (
          <year>2007</year>
          )
          <fpage>90</fpage>
          -
          <lpage>95</lpage>
          . doi:
          <volume>10</volume>
          .1109/
          <string-name>
            <surname>MCSE</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <volume>55</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Tong</surname>
          </string-name>
          , Y. Liu,
          <article-title>Impact of high temperature heat waves on ocean carbon sinks: Based on literature analysis perspective</article-title>
          ,
          <source>Journal of Sea Research</source>
          <volume>198</volume>
          (
          <year>2024</year>
          )
          <article-title>102487</article-title>
          . URL: https:// www.sciencedirect.com/science/article/pii/S1385110124000200. doi:https://doi.org/10.1016/ j.seares.
          <year>2024</year>
          .
          <volume>102487</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>X.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          , D.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yeung</surname>
          </string-name>
          , W.-k. Wong, W.-c. Woo,
          <article-title>Convolutional lstm network: A machine learning approach for precipitation nowcasting (</article-title>
          <year>2015</year>
          ). URL: https://arxiv.org/abs/1506. 04214. doi:
          <volume>10</volume>
          .48550/ARXIV.1506.04214.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>