<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Observation, November</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>UNet for the Future Weather Prediction: Weather4cast 2021</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sungbin Choi</string-name>
          <email>sungbin.choi.1@gmail.com</email>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>1</volume>
      <issue>2021</issue>
      <abstract>
        <p>This paper describes our experiments of UNet based deep convolutional neural network model applied on Weather4cast challenge 2021 stage 1. This challenge's task is to predict future weather patterns, which is composed of four target variables per pixel: temperature, convective rainfall rate, probability of occurrence of tropopause folding, and cloud mask. Given equal size of input and output weather image, we trained UNet shaped neural network model to make predictions for each target variable. Evaluation results show competitive performance compared to the baseline method.</p>
      </abstract>
      <kwd-group>
        <kwd>Weather forecast</kwd>
        <kwd>deep learning</kwd>
        <kwd>neural network</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In this paper, we describe our methods on Weather4cast
2021’ stage 1 challenge. This challenge’s task is to
predict future 8 hours weather image given prior 1 hour’s
weather data [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Both input and output weather image
have 256 x 256 pixels, and each pixel represents
approximately 3km x 3km spatial area. Input weather image
contains various weather-related variable measurements
(For detailed explanation with regard to the input data
variables, please see [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]). Target output variable to
predict is temperature, convective rainfall rate, probability
of occurrence of tropopause folding, and cloud mask.
      </p>
      <p>
        Training data contains weather images acquired from 3
distinct regions worldwide, named R1, R2 and R3 (which
corresponds to Nile region, Eastern Europe, and South
West Europe, respectively) in this challenge. In the core
challenge task, both training data and held out test data
came from R1, R2 and R3 regions. In the transfer learning
task, training data is same as core challenge task, but held
out test data is coming from diferent regions, named R4,
shown efective performance in similar tasks [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], [7]. We trained neural network model having UNet
based architecture. Each convolution block is densely
connected with subsequent layers like a DenseNet [8].
      </p>
      <p>Evaluation results show competitive performance
compared to the baseline methods for this task.
2.</p>
    </sec>
    <sec id="sec-2">
      <title>Methods</title>
      <sec id="sec-2-1">
        <title>2.1. Preprocessing data</title>
        <p>images.
vector.</p>
        <p>Prediction takes prior 1 hour’s weather data as input,
and should produce prediction ranging to next 8 hours.
Each weather image has 15 minutes interval, so input has
4 weather images, and output should have 32 weather</p>
        <p>Input weather image contains various weather-related
variables data per pixel. There are 9 continuous
variables, and 16 discrete variables. Continuous variables are
normalized to minimum 0.0, maximum 1.0 range, and
discrete variables are converted into one-hot encoding</p>
        <p>
          In this study, recurrent model such as LSTM (long
short term memory network) [9] was not utilized. So,
time dimension is simply merged into feature channel
dimension.
2.2. Model
We implemented UNet [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] based model, as described in
layers, which are densely connected as depicted in
Figure 2 [8].
        </p>
        <p>Each blue box represents dense convolution layers
block with average pooling layer. Each orange box
represents deconvolution layers. Green arrow represents
skip connections between downsampling path and
upsampling path.</p>
        <p>We trained model for each of 4 target variables
separately, but all of them have identical UNet with DenseNet
model structure described above. For temperature,
convective rainfall rate, probability of occurrence of tropopause
folding, mean squared error is used as loss function. For
cloud mask, since it can have only binary output (0 or 1),
sigmoid loss is used instead.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.3. Training</title>
        <p>Training data has 316 days weather images from R1, R2
and R3 region, time period ranging 2019-2020. Held-out
test data is captured from separate time period
(20202021).</p>
        <p>In this study, we haven’t tried to train region-specific
model. All training weather images from diferent
regions were merged and used together to train general
model that can be used to apply on all regions.</p>
        <p>We trained 1 model for the temperature prediction,
2 models for the convective rainfall rate prediction, 2
models for the probability of occurrence of tropopause
folding prediction, and 3 models for the cloud mask
prediction. When multiple models are trained for the same
target variable, output prediction values are simply
averaged to produce final output. Our experimentation code
is publicly available at https://github.com/sungbinchoi/
w4c_st1.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <p>Oficial evaluation metric for this challenge was mean
squared error divided by mean squared error of the
baseline prediction, which is taking last previous weather
image as prediction output on all out timeframe, so
baseline prediction output is scored as 1.0.</p>
      <p>In this study, we trained only region-agnostic model.
Same models used to produce prediction for the core
task is used for the transfer learning task without any
change. Our best evaluation results from held out test
set was 0.507325 in the core challenge task and 0.465760
in the transfer learning task, which means roughly loss
is halved compared to the baseline.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>In this experiment, we used UNet for the weather forecast
task. It showed competitive performance compared to
the baseline.</p>
      <p>In our study, we trained only region-agnostic model.
In our preliminary experiment, it was not clear whether
a region-specific model is more efective than the
general model for the core challenge task. As a layperson,
this was quite counterintuitive, because weather from R1
(Nile region) seems very diferent from R3 (South West
Europe). We will explore more various research ideas
and hope to find more efective methods in the next stage
of the challenge.
[7] H. Huang, L. Lin, R. Tong, H. Hu, Q. Zhang,
Y. Iwamoto, X. Han, Y. Chen, J. Wu, Unet 3+: A
full-scale connected unet for medical image
segmentation, in: 2020 IEEE International Conference on
Acoustics, Speech and Signal Processing, ICASSP
2020, Barcelona, Spain, May 4-8, 2020, IEEE, 2020,
pp. 1055–1059.
[8] G. Huang, Z. Liu, L. Van Der Maaten, K. Q.
Weinberger, Densely connected convolutional networks,
in: Proceedings of the IEEE conference on computer
vision and pattern recognition, 2017, pp. 4700–4708.
[9] S. Hochreiter, J. Schmidhuber, Long short-term
memory, Neural Computation 9 (1997) 1735–1780.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>IARAI</surname>
          </string-name>
          , Weather4cast, https://www.iarai.ac.at/ weather4cast/,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>O.</given-names>
            <surname>Ronneberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Brox</surname>
          </string-name>
          , U-net:
          <article-title>Convolutional networks for biomedical image segmentation</article-title>
          , in: International Conference on
          <article-title>Medical image computing and computer-assisted intervention</article-title>
          , Springer,
          <year>2015</year>
          , pp.
          <fpage>234</fpage>
          -
          <lpage>241</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Dou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Heng</surname>
          </string-name>
          , Hdenseunet:
          <article-title>Hybrid densely connected unet for liver and tumor segmentation from CT volumes</article-title>
          ,
          <source>IEEE Trans. Medical Imaging</source>
          <volume>37</volume>
          (
          <year>2018</year>
          )
          <fpage>2663</fpage>
          -
          <lpage>2674</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Weng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <article-title>Nas-unet: Neural architecture search for medical image segmentation</article-title>
          ,
          <source>IEEE Access 7</source>
          (
          <year>2019</year>
          )
          <fpage>44247</fpage>
          -
          <lpage>44257</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>D-unet: A dimension-fusion U shape network for chronic stroke lesion segmentation</article-title>
          ,
          <source>IEEE ACM Trans. Comput. Biol. Bioinform</source>
          .
          <volume>18</volume>
          (
          <year>2021</year>
          )
          <fpage>940</fpage>
          -
          <lpage>950</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Qamar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Usama</surname>
          </string-name>
          ,
          <article-title>A variant form of 3d-unet for infant brain segmentation</article-title>
          ,
          <source>Future Gener. Comput. Syst</source>
          .
          <volume>108</volume>
          (
          <year>2020</year>
          )
          <fpage>613</fpage>
          -
          <lpage>623</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>