<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Intitcesr,naantidonEanlgCinoenefreinregn.cCezoefstYoecahrolywaR,eJpaonrtusaorny noise.</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>A Fast Progressive U-Net For MRI Denoising In The K-space</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lorenzo Di Luccio</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Renato Giamba</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adriano Puglisi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefano Giagu</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Physics, Sapienza University of Rome</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>1</volume>
      <fpage>4</fpage>
      <lpage>16</lpage>
      <abstract>
        <p>Magnetic Resonance Imaging (MRI) is a prevalent non invasive imaging technique that produces high contrast anatomical images without ionizing radiation. However, due to long acquisition times, MRI scans are prone to noise and artifacts corruption. In this paper, we address the denoising problem by leveraging the intrinsic nature of the K-space domain where the noise naturally occurs during acquisition. We propose a light, eficient U-Net architecture that is specifically tailored to operate directly in the K-space. The model considers a residual learning-based estimate of the noise component across a range of noise levels and distributions and in particular for additive Gaussian noise. We also propose a SNR degradation based progressive training scheme that greatly improves performance across a wide range of noise levels. The network is computationally cost-efective and can be run on CPU in acceptable inference time, making it suitable for real time or resource constrained applications.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>MRI is a powerful medical imaging technique that
provides detailed images of the internal structures of the
body of a patient. The main benefit of MRI is the
possibility to acquire images in multiple planes (sagittal, coronal,
and axial) without repositioning the patient. MRI
provides a very high contrast images of soft tissues, using
their water content and molecular properties. It is also
non invasive because it doesn’t require X-rays or other
radiations that could be harmful for the patient.</p>
      <p>When a patient is placed inside an MRI scanner, the
machine generates a magnetic field that aligns the
protons in fat and water molecules along the direction of
the field. Once the alignment is established, the scanner
applies radio frequency pulses at the natural frequency
of the protons. The energy stored in the pulses excites
the alignment, and the protons precess along the axis of
the magnetic field. When the pulses stop, the protons
begin to realign themselves to their original
configuration. As they realign, they release energy in the form of
radio frequency signals, which are picked up by special
receiving coils. They do not appear directly as an image;
they are in K-space, a domain in which data is recorded in
terms of spatial frequencies. To obtain a doctor-readable
image, an inverse Fourier transform is then performed
on this complex valued data, resulting in high-resolution
grayscale images that relate to the anatomical structure
of the tissues being examined.</p>
      <p>Although magnetic resonance imaging (MRI) is a
valuable tool for medical imaging, the resulting scans are
inherently noisy and present some challenges. The
acquisition process is time-consuming and expensive, so
the patient must remain still throughout the scan,
otherwise the resulting images will contain motion artifacts.</p>
      <p>Additionally, the MRI machine is sensitive to external
factors such as temperature and electrical pulses, which
can cause additional noise to be introduced. The thermal
mobility of protons can potentially introduce noise into
the image background, especially in low-signal regions.</p>
      <p>Mathematically, certain probability distributions can
be used to characterize noise in MRI scans. The
equation 1 describes the bell-shaped symmetric probability
distribution of Gaussian noise, often known as normal
noise. This type of noise is additive, defined by a
constant standard deviation across the image, and is typically
attributed to thermal motion and electrical noise.</p>
      <p>,
() =</p>
      <p>1
 √2
− (−  )2
 2 2</p>
      <p>Rician noise is prominent in the low signal-to-noise
ratio (SNR) regions of the scans. It follows a Rician
distribution (equation 2) and emerges when the magnitude of
the MRI signal is nonnegative and follows a Rayleigh
distribution, while the phase is uniformly distributed. This
type of noise is particularly dificult to handle compared
to other types of noise.</p>
      <p>,
() =
 − (−  )2
 2  2 2
0
︁(  )︁
 2
(1)
(2)
creating variability in the returned signal. It follows a They reduce noise caused by rapid change or motion but
Poisson distribution (equation 3). occasionally perturb content as well.</p>
      <p>Non-linear filters are applied when the noise is not
 () =  −  (3) uniformly distributed. They rely on adaptive kernels
! or other rules of convolution. They are more advanced</p>
      <p>
        Finally, the T1 and T2 relaxation phases are usually and can function diferently depending on local intensity
subject to exponential noise. This type of noise reflects patterns.
the intrinsic unpredictability of the decay of MRI signals The second is anisotropic difusion filtering, which is a
over time and has an exponential distribution (equation variation of standard spatial filtering in that it adapts its
4). behavior to the local image properties. The algorithm
ad () =  −  (4) justs the number of filtering iterations depending on the
intensity gradient in the neighborhood: more iterations
A modern and popular approach to deal with long ac- where there are smooth transitions, fewer where there
quisition processes is to take a shorter one and capture are sharp transitions. This technique preserves edges
less details, this involves acquiring only a portion of the more efectively than linear filters, but certain details
K-space data which corresponds to a subset of spatial along boundaries are still lost.
frequencies. With this process, the generated images The second technique is the non-local means filter,
contains more artifacts and more noise, but reliable and which better exploits image redundancy. In this case,
powerful deep learning models can enhance these scans. the algorithm treats larger structures as meaningful
feaThis paper is focused on the implementation of a model tures and smaller patterns as noise that should be
refor denoising that can work with diferent levels of noise; moved. Each pixel is adjusted by averaging values from
in particular the focus is on denoising MRI scans with an a broader neighborhood area weighted by structure
simiadditive Gaussian noise at the K-space level. The model, larity. This technique overcomes some of the deficiencies
which follow a residual approach, should adapt also to of anisotropic difusion but at the cost of more
compuother noise probability distributions. tation, since it involves a search in large neighborhoods
for each pixel.
2. Related Works Finally, methods that combine domain and range
filtering consider both spatial proximity and intensity
similarFrom image pixel analysis to contemporary deep learn- ity. A highly successful algorithm in this category is the
ing models, a number of methods have been developed bilateral filter, initially derived by Tomasi and Manduchi
recently to solve the issue of denoising in magnetic reso- [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], subsequently generalized to MRI denoising by Xie
nance imaging (MRI). Using the subdivision suggested by et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. This filter averages spatial and photometric
J. Mohan et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we follow an outline of the primary information to some extent by averaging neighbor pixel
MRI denoising techniques in this section, beginning with values and diminishing the influence of those with
exthose that are not based on deep learning. treme intensity diferences, thereby actually preserving
edges. Its derivative, the trilateral filter, incorporates a
time element as well, which causes it to respond more
2.1. Filtering favorably in dynamic or multi-frame situations. These
These algorithms make use of a weighting kernel on the techniques, while efective in noise reduction and
strucnoisy image to reduce its variance. Damping of noise is tural detail preservation, can produce visible halo
artidone with this operation. Filters are basically of linear facts along sharp edges, where the filtered to unfiltered
and non linear category. boundary is evident. Domain techniques address image
      </p>
      <p>Linear filters are ideal if the noise is evenly distributed. denoising by first converting the image from the
spaThey are realized with fixed smoothing kernels and are tial domain to a new representation, where noise and
not dependent on the content. There is a subgroup that appropriate image structures can be separated more
easacts in the spatial domain: the kernel acts on each pixel of ily. Standard filters or operations specifically designed
a 2D neighborhood. The goal is to smooth, detect edges, for the operation can then be applied to the new domain.
or sharpen. Traditional examples include the Gaussian The main classes within this class difer depending on the
iflter, the Sobel operator, and the Laplacian kernel. Linear type of transform used and the character of the resulting
iflters remove Gaussian noise, but blur details as well. representation.</p>
      <p>The second type is temporal. There, the kernel is used One of the simplest methods is frequency domain
dein the time domain. There, the kernel is applied along noising, which applies the Fourier transform to represent
time, on the same spatial pixels across two frames. It the image in terms of spatial frequency. Noise, being
detects temporal changes, the most widely used ones high frequency by nature, can be minimized in this
doare temporal averaging, diferencing, and optical flow. main without losing low-frequency data. This method is
usually efective when the noise intensity is moderate, coder presented by Hinton and Salakhutdinov [19]. The
but breaks down when the noise increases. main application is to learn a latent and more compact</p>
      <p>The wavelet transform is an improvement over this representation of an image by downgrading it in the first
method, as it decomposes the image into sets of frequency half of the model and then reconstructing it back in the
bands at diferent scales and allows simultaneous local- second half; this process can be adapted to denoising by
ization in both space and frequency. The multi resolution passing noisy images as input and expecting as output
structure allows for better separation between noise and their denoised version. Similar architectures have also
signal at diferent frequency levels. However, the wavelet been applied in other areas of the medical sector, such
basis can be challenged when representing curved or as in the processing of speech signals for the automatic
highly textured structures, which limits its applicability detection of speech disorders [20]. Some years later,
Ronin some medical imaging scenarios. neberger et al. [21] proposed a new architecture, the</p>
      <p>This drawback was addressed with the creation of the U-Net, consisting in a fusion between the Autoencoder
curvelet transform, a generalization of wavelets specif- and the ResNet, that achieved good results in biomedical
ically designed to produce a coarser representation of image segmentation. Also the U-Net can be adapted to
curves and edges. It is well suited to handle long and perform general image denoising and so in particular
anisotropic features, and is ideal for images with highly MRI denoising.
directional textures. Despite this advantage, the trans- Although the deep learning MRI denoising
perforform produces artifacts in smooth regions, where the mance was adequate, a new problem arose. MRI scans
redundancy of the representation is lower. with noise distributions other than Gaussian could not</p>
      <p>The contourlet transform takes this concept a step be accommodated by the networks. This drawback stems
further by combining multiscale and directional decom- from the straightforward methodology of the models,
position. It starts with a Laplacian pyramid for coarse which learned to produce a clean image from a noisy
imimage structures and then uses directional filtering to age. This indicated that the models had only worked on
capture edges at diferent orientations. This produces a denoising one distribution, or perhaps a limited number
more expressive and flexible representation, especially of related distributions. Zhang et al. [22] proposed a new
for contours and geometric shapes. However, the in- approach that they called residual learning in which the
creased complexity of this method leads to increased model learns to predict the residual image, i.e. the noise
computational costs, especially for the fine-grained edge from the noisy image, with the possibility of focusing on
preservation required at multiple scales. more and diferent noise distributions. Residual learning
also introduced regularization in the training process and
boosted the image denoising performances.</p>
      <sec id="sec-1-1">
        <title>2.2. Deep Learning approaches</title>
        <p>
          In recent years, machine learning and then deep learning
started to become popular, because of the better
computational performance ofered by modern GPUs and the
brilliant results obtained. The first improvement in the
use of deep learning for images was the introduction of
hybrid neural networks models [
          <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7 ref8">4, 5, 6, 7, 8</xref>
          ] and
convolutional neural networks (CNNs) [
          <xref ref-type="bibr" rid="ref10 ref11 ref9">9, 10, 11</xref>
          ] with the
architecture LeNet proposed by LeCun et al. [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. This
model was taken as inspiration by subsequent and
popular networks like AlexNet [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], GoogLeNet [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], VGG
[
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], transformers [16, 17] .
        </p>
        <p>Afterwords, researchers believed that increasing the
number of convolutional layers in the networks would
have lead to better performances, but it was not true
because the models sufered from the vanishing gradient
problem. To overcome this issue, He et al. [18] proposed
ResNet, a new network with the so called residual layers
where the new outputs are computed with an additive
update from the previous inputs. Moreover, ResNet
allowed to obtain better performances and a more stable
training with less convolutional layers.</p>
        <p>ResNet was also one of first deep learning models used
for general image denoising, together with the
Autoen</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>3. Methodology</title>
      <sec id="sec-2-1">
        <title>This work introduces a new method for MRI denoising that try to exploit and merge the benefits of previous works, keeping also an eye on speed.</title>
        <sec id="sec-2-1-1">
          <title>3.1. Data Acquisition</title>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>The dataset is the Information eXtraction from Images</title>
        <p>(IXI) [23], which includes approximately 600 MRI scans
in NIFTI format. Data were acquired from healthy
subjects using diferent acquisition protocols (T1, T2, PD,
MRA, DTI) at three hospitals in London, namely
Hammersmith Hospital (Philips 3T), Guy’s Hospital (Philips
1.5T), and the Institute of Psychiatry (GE 1.5T). Each scan
is associated with protocol specific parameters, publicly
available on the dataset website.</p>
        <sec id="sec-2-2-1">
          <title>3.2. MRI Scan Preprocessing</title>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>The main preprocessing step is to extract 2D slices from the T1-weighted volumetric MRI of each subject. This</title>
        <p>procedure ofers several advantages. 2D image process- niques were not used. This is because the U-Net model
ing is faster and more suitable for real-time situations, already tends to adjust the activations thanks to the
conas it uses less processing resources than full-volume pro- catenation mechanism between the old and new features
cessing. A more focused investigation is also possible, in the upsampling phase. Another reason is related to
as regions of interest are often located in a single plane, eficiency, the network remains faster and, in the tests
such as axial, coronal or sagittal. This makes it possible carried out, the denoised images with or without
normalto focus on the relevant anatomical elements and elimi- ization did not present relevant diferences. PReLU was
nates the need to consider the entire volume. 2D images adopted as the activation function, since the presence of
are easier to visualise and evaluate and provide a recog- a trainable parameter improves the overall performance
nisable and comprehensible representation of the data. of the network according to the metrics used.
Another advantage is the reduction in the overall size The network can be described with simple equations
of the data, which facilitates its transmission or storage, following its division in layers.
especially in resource limited environments. The down part can be seen as the function</p>
        <p>Despite these advantages, relying solely on 2D slices DownConv2d () defined as in equation 6, in which the
has some disadvantages. It may not accurately represent  parameter represents the features computed by the
complex structures or dynamic processes and may result previous layer.
in the loss of 3D or temporal information. In some
situations, combining both modalities and examining selected  = PReLU (Conv2d ()) (6)
slices in the context of the full volume may be more use- +1 = PReLU (Conv2d ())
ful. For each of the 581 subjects, for a total of 43,575 The up part can be seen as the function
images, the 25 central slices in the coronal, axial, and UpConv2d (, − − 1) defined as in equation 7,
sagittal planes are selected to avoid completely or mostly in which the  parameter represents the features
comblack images. A padding technique involving continuous puted by the previous layer and the − − 1 parameter
zero-filling is used to uniformly scale all photos to the represents the correspondent features computed by the
maximum size, since slices acquired from diferent planes down part to be concatenated.
have variable sizes.
ingA(f7te5r%p),rveaplriodcaetisosnin(g1,5t%he) adnadtatseestt(i1s0s%pl)i;t into the train-  = cat(+P1R=eLUPR(eCLoUnv(TC2odnv(2d)(),))− − 1) (7)</p>
        <p>
          Before loading the images in the dataset they are
normalized in the [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ] range, then the noisy images are Since the model (figure 1 for a visual representation)
generated by adding Gaussian noise in the K-space ob- uses the residual approach, the entire network can be
tained by the Fourier transform. seen as the function  = Net (), where the input  is
        </p>
        <p>In particular, the levels of added noise is based on the a batch of noisy MRI scans in the K-space and the output
Signal-to-Noise Ratio (SNR), with the standard deviation  is the estimated batch of noise also in the K-space.
 computed using equation 5, where |ℱ []| is the average
of the magnitude of the Fourier transform of the image. 3.3.1. Loss Function</p>
      </sec>
      <sec id="sec-2-4">
        <title>Image denoising in general can be considered a regression</title>
        <p>(5) task, so a well suited loss function is the mean squared
error (MSE) defined in equation 8.</p>
        <sec id="sec-2-4-1">
          <title>3.3. Proposed Model</title>
        </sec>
      </sec>
      <sec id="sec-2-5">
        <title>The model proposed in this work is a modified version</title>
        <p>of the classic U-Net. The main modifications concern In particular we use the MSE applied to noisy scans
several aspects of the network. In both the descending and free noise scans is called pixel loss (PL) (equation 9),
and ascending parts, five convolutional layers are present. which is the main part of the total loss function.
Although additional layers can be added, the size of the Some works like Zhao et al. [24] or Mustafa et al.
preprocessed MRI images is 256×256, so adding more [25] showed that in image regression tasks, the MSE
layers would lead to activation maps that are too small. succeeded in performing the task with good results, but it
Furthermore, the network is faster and lighter by avoid- blurred the resulting images. A popular way to overcome
ing additional layers. this problem is to add other parts in the total loss function,</p>
        <p>Convolutions do not include bias, since the introduc- focusing on the features [24], or to design from scratch a
tion of bias parameters has shown a drastic worsening specific and more complex loss function [25].
of the network performance, generating images with For the proposed network, we integrates other two
anomalous visual artifacts. Similarly, normalization tech- parts in the total loss function, the frequency loss (FL)
MSE (, ) =</p>
        <p>1 ∑︁ (, − )2
 =1</p>
        <p>(8)
pling and 5 upsampling layers, connected via skip connections.
(equation 10) that is the sum of MSE applied to real and
imaginary part of the Fourier transform of the noisy scans
and the free noise ones, and the edge loss (EL) (equation
11 that is the MSE applied to some edge features of the
noisy scans and the free noisy ones; in particular the 
function used are the Laplacian and/or the Sobel filters.</p>
        <p>PL (, ) = MSE (, )
(9)</p>
        <p>EL (, ) = MSE ( () ,  ())</p>
      </sec>
      <sec id="sec-2-6">
        <title>The total loss function L is the simple sum of the three</title>
        <p>previously introduced parts: the pixel loss (PL), the
frequency loss (FL) and the edge loss (EL).</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Experiments &amp; Results</title>
      <sec id="sec-3-1">
        <title>The experiments were conducted on a Google Colab en</title>
        <p>vironment with an NVIDIA T4 GPU (CUDA 11.8), 12 GB</p>
      </sec>
      <sec id="sec-3-2">
        <title>RAM, and Intel Xeon CPU at 2.20 GHz. Python version</title>
      </sec>
      <sec id="sec-3-3">
        <title>3.10, and the PyTorch version 2.0.1 with CUDA.</title>
        <p>We used ADAM optimizer, the starting learning rate of
0.001 and batch size of 64. A learning rate scheduler was
used to reduce the learning rate by a factor of 0.1 at every
instance that the validation loss failed to decrease for two
successive epochs. Training was also monitored using
an early stopping criterion that terminated if validation
loss failed to drop for five epochs.</p>
      </sec>
      <sec id="sec-3-4">
        <title>Model comparison was based on two common im</title>
        <p>age reconstruction evaluation metrics,namely the Peak</p>
      </sec>
      <sec id="sec-3-5">
        <title>Signal-to-Noise Ratio (PSNR) and Structural Similarity</title>
      </sec>
      <sec id="sec-3-6">
        <title>Index Measure (SSIM). The PSNR is defined as</title>
        <p>PSNR(,  ) = 10 log10
︂(
max()2 )︂
MSE(,  )
(13)
where  and  are the reference and reconstructed
images, respectively, and MSE is the mean squared error.</p>
      </sec>
      <sec id="sec-3-7">
        <title>The SSIM is computed as</title>
        <p>SSIM(,  ) =</p>
        <p>(2    + 1)(2  + 2)
( 2 +  2 + 1)( 2 +  2 + 2)
, (14)
where   ,   are the local means,  2 ,  2 the variances,
and   the covariance between  and  , with 1 and
2 being constants to stabilize the division. Two pairs of
training experiments can be distinguished. The first uses</p>
      </sec>
      <sec id="sec-3-8">
        <title>MSE loss, while the second uses MAE loss plus another</title>
        <p>type of normalization. The studies vary in how noisy
scans are produced within each pair. Specifically, the
second uses a progressive technique, while the first uses
a random one.</p>
        <sec id="sec-3-8-1">
          <title>4.1. Evaluation under MSE and MAE</title>
        </sec>
      </sec>
      <sec id="sec-3-9">
        <title>The training phase was structured around two main</title>
        <p>configurations, each evaluated under diferent
conditions. The first group of experiments employed the Mean</p>
      </sec>
      <sec id="sec-3-10">
        <title>Squared Error (MSE) as loss function, while the second</title>
        <p>
          used the Mean Absolute Error (MAE), along with a difer- reduction. For this reasons, the other two experiments
ent input normalization strategy based on centering the are made with the same loss function, but substituting
data with zero mean and unit standard deviation, instead each instance of MSE with MAE.
of scaling it to the [
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ][
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ] interval. Within each group, Another change is to have the normalized MRI scans
two noise injection strategies were considered: random centered with mean 0 and standard deviation 1 instead
and progressive. of having them normalized in the interval between 0 and
        </p>
        <p>In the first experiment, the training set was corrupted 1. This because this type of normalization is widely used
by Gaussian noise with a signal to noise ratio (SNR) ran- in lots of works and tends to have better performances
domly sampled from a discrete uniform distribution be- in most of the cases.
tween -5 and 5. For the validation set, the SNR was fixed
at -5 to maintain a challenging evaluation scenario. The 4.3. Results
model was trained with early stopping based on
validation loss. From Table 1 and Figures 2a–3d, the following trends are</p>
        <p>The second experiment introduced noise in a progres- easily observable. The most obvious is that progressive
sive manner, the training began with images at SNR = 5 noise injection consistently outperforms random noise
and gradually moved down to SNR = -5. At each level, the injection. With both loss functions, MSE and MAE, the
model was trained separately using a validation set with progressive setting improves the average by about 4.8 dB
matching noise level. This setup was designed to let the for PSNR and from 0.14 to 0.15 for SSIM. In addition to
model adapt incrementally to increasing noise intensity. achieving higher scores, this training approach is also</p>
        <p>The third and fourth experiments mirrored the struc- more stable, with lower variance at various noise levels.
ture of the first and second, respectively, but used the Comparing the two loss functions, the model trained with
MAE loss instead of MSE, and applied the aforementioned MSE has slightly better SSIM values. The diference is
centered normalization. These variations aimed to evalu- small, about 0.01 on average, but constant. This indicates
ate the impact of a loss function less sensitive to outliers that the model trained with MSE is better at keeping
and a normalization scheme more common in deep learn- structural details intact. In contrast, the model trained
ing workflows. Detailed results for all experiments are with MAE is superior in PSNR by about 0.2 dB on average.
reported in Table 1. This holds true for all noise levels and suggests that MAE
may be better at removing noise, albeit at the expense of
4.2. The MAE and the centered standard slightly worse structural fidelity These results highlight
the behavior of the proposed architecture under diferent
normalization training conditions. To further contextualize its
performance, we compare it against existing state-of-the-art
denoising approaches.</p>
      </sec>
      <sec id="sec-3-11">
        <title>Although the random approach introduces stochasticity</title>
        <p>in the training procedure allowing the model to be more
robust to diferent levels of noise earlier, the progressive
approach has faster convergence time because the model
only has to adapt to the next level of noise from the
previous without the abrupt shifts.</p>
        <p>Moreover, MSE can be sensitive to the presence of
noise or outliers in the data because it strongly
penalizes large deviations from the true values, due to the
squared function. So the resulting denoised images could
be overly influenced by noise. Instead, the Mean
Absolute Error (MAE) (equation 15) is more robust to noise
and outliers.</p>
        <p>MAE (, ) =</p>
        <p>1 ∑︁ |, − |
 =1
(15)</p>
        <p>It treats errors uniformly, due to the absolute value
function making it less susceptible to extreme values in
the data. Consequently, the denoised images produced
tend to be more noise resistant. Also, MAE tends to
produce images that are visually crisper, but with better
preservation of fine structures. MSE instead tends to
suppress high-frequency details and edges in favor of noise</p>
        <sec id="sec-3-11-1">
          <title>4.4. Comparisons with other methods</title>
        </sec>
      </sec>
      <sec id="sec-3-12">
        <title>To evaluate the denoising performance of the networks,</title>
        <p>they are compared with other four denoising methods.
In particular, with Optimized NLM by Coupé et al. [26],
WSM by Coupé et al. [27], MCDnCNNg and MCDnCNNs
by Jiang et al. [28].</p>
        <p>It is useful to point out that the four chosen methods
work with diferent datasets and have a diferent noise
generation procedure. Therefore, the collected metrics
to make comparisons are diferent (a kind of average on
the level of noise) from the real ones.</p>
        <p>Analyzing the table 2, the two fast U-Net trained with
the progressive approach have overall very similar
metrics with the other methods. In particular they are better
on MRI scans with low noise, while they lose some
performances on scans with high noise. This is due to the
trade-ofs made to keep the network fast.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Conclusions</title>
      <p>improve its speed while maintaining good performance.</p>
      <p>The noise generation procedure, based on the K-space
In this work, a new approach to MRI denoising is pre- and the signal-to-noise ratio (SNR), allows to obtain more
sented, introducing several novelties in both the noise realistic noisy scans. This realism is also due to the
depengeneration phase and the training strategy, as well as dence of the noise on the mean of the Fourier transform
an architectural modification of the classical U-Net to modulus. However, this type of generation makes the
(a) MSE, random
(b) MSE, SNR=5
(c) MSE, SNR=0
(d) MSE, SNR=-5
comparison with other methods more complex, since
in most works a simulated noise based on predefined
percentages is preferred.</p>
      <p>The progressive training approach, where the SNR
varies from 5 to − 5, represents a turning point for the
improvement of the results both in terms of PSNR and
SSIM, as well as leading to a clear visual quality in the
denoised images. Furthermore, the adaptive nature of this
strategy allows the network to maintain the denoising
performance already acquired along all the considered
noise levels.</p>
      <p>The proposed U-Net, composed of only five layers and
without normalizations, can run in short times even on
CPU. This compromise in terms of speed does not lead
to a significant degradation of the metrics, which remain
comparable with those obtained by other methods based
on deep learning. It should also be noted that, although
processing time is rarely analyzed in the literature, today
in the medical field it is increasingly important to obtain
reliable results in short times.</p>
    </sec>
    <sec id="sec-5">
      <title>6. Declaration on Generative AI</title>
      <sec id="sec-5-1">
        <title>During the preparation of this work, the authors used</title>
        <p>ChatGPT, Grammarly in order to: Grammar and spelling
check, Paraphrase and reword. After using this
tool/service, the authors reviewed and edited the content as
needed and take full responsibility for the publication’s
content.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Krishnaveni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <article-title>A survey on the magnetic resonance image denoising methods</article-title>
          .,
          <source>Biomed. Signal Process. Control</source>
          .
          <volume>9</volume>
          (
          <year>2014</year>
          )
          <fpage>56</fpage>
          -
          <lpage>69</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Tomasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Manduchi</surname>
          </string-name>
          ,
          <article-title>Bilateral filtering for gray and color images</article-title>
          ,
          <source>6th Int. Conf. Comput. Vis</source>
          ., Bombay, India. (
          <year>1998</year>
          )
          <fpage>839</fpage>
          -
          <lpage>846</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.-A.</given-names>
            <surname>Heng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>Image difusion using saliency bilateral filter</article-title>
          .,
          <source>IEEE Trans. Information Technology in Biomedicine 12</source>
          (
          <year>2008</year>
          )
          <fpage>768</fpage>
          -
          <lpage>771</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Paternò</surname>
          </string-name>
          ,
          <article-title>An innovative hybrid neuro-wavelet method for reconstruction of missing data in astronomical photometric surveys</article-title>
          ,
          <source>in: Lecture Notes in Computer Science (in- 2014. cluding subseries Lecture Notes in Artificial Intel-</source>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Alfarano</surname>
          </string-name>
          , G. De Magistris,
          <string-name>
            <given-names>L.</given-names>
            <surname>Mongelli</surname>
          </string-name>
          , S. Russo,
          <source>ligence and Lecture Notes in Bioinformatics)</source>
          , vol- J. Starczewski,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <source>A novel convmixer transume 7267 LNAI</source>
          ,
          <year>2012</year>
          , p.
          <fpage>21</fpage>
          -
          <lpage>29</lpage>
          . doi:
          <volume>10</volume>
          .
          <article-title>1007/ former based architecture for violent behavior de978-3-642-29347-4_3</article-title>
          . tection, in: Lecture Notes in Computer Science
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Pappalardo</surname>
          </string-name>
          , E. Tramontana,
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Now-</surname>
          </string-name>
          (
          <source>including subseries Lecture Notes in Artificial Inicki</source>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Starczewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Woźniak</surname>
          </string-name>
          ,
          <source>Toward work telligence and Lecture Notes in Bioinformatics)</source>
          ,
          <source>volgroups classification based on probabilistic neu- ume 14126 LNAI</source>
          ,
          <year>2023</year>
          , p.
          <fpage>3</fpage>
          -
          <lpage>16</lpage>
          . doi:
          <volume>10</volume>
          .1007/ ral network approach,
          <source>in: Lecture Notes in Ar- 978-3-031-42508-0_1. tificial Intelligence (Subseries of Lecture Notes</source>
          <volume>in</volume>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fiani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , Remote eye movement Computer Science), volume
          <volume>9119</volume>
          ,
          <year>2015</year>
          , p.
          <fpage>79</fpage>
          -
          <lpage>89</lpage>
          . desensitization and reprocessing treatment of longdoi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -19324-
          <issue>3</issue>
          _8. covid- and
          <string-name>
            <surname>post-</surname>
          </string-name>
          covid
          <article-title>-related traumatic disorders:</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          , G. Capizzi,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <article-title>Some remarks An innovative approach</article-title>
          ,
          <source>Brain Sciences</source>
          <volume>14</volume>
          (
          <year>2024</year>
          ).
          <article-title>on the application of rnn and prnn for the charge</article-title>
          - doi:10.3390/brainsci14121212.
          <article-title>discharge simulation of advanced lithium-ions bat-</article-title>
          [18]
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Deep residual learntery energy storage</article-title>
          ,
          <source>in: SPEEDAM</source>
          <year>2012</year>
          -
          <article-title>21st In- ing for image recognition</article-title>
          ,
          <year>2015</year>
          . ternational Symposium on Power Electronics, Elec- [19]
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Salakhutdinov</surname>
          </string-name>
          , Reducing the ditrical Drives,
          <source>Automation and Motion</source>
          ,
          <year>2012</year>
          , p.
          <article-title>941 mensionality of data with neural networks</article-title>
          ,
          <source>Science - 945. doi:10</source>
          .1109/SPEEDAM.
          <year>2012</year>
          .
          <volume>6264500</volume>
          . 313 (
          <year>2006</year>
          )
          <fpage>504</fpage>
          -
          <lpage>507</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. L.</given-names>
            <surname>Sciuto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tramontana</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          [20]
          <string-name>
            <given-names>L.</given-names>
            <surname>Corvitto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Faiella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Puglisi</surname>
          </string-name>
          , S. Russo,
          <article-title>multithread nested neural network architecture to</article-title>
          et al.,
          <article-title>Speech and language impairment detection model surface plasmon polaritons propagation, Mi- by means of ai-driven audio-based techniques</article-title>
          ,
          <source>in: cromachines 7</source>
          (
          <year>2016</year>
          ).
          <source>doi:10.3390/mi7070110. CEUR WORKSHOP PROCEEDINGS</source>
          , volume
          <volume>3869</volume>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lo Sciuto</surname>
          </string-name>
          , G. Capizzi,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shikler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <surname>Or-</surname>
          </string-name>
          CEUR-WS,
          <year>2024</year>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>31</lpage>
          .
          <article-title>ganic solar cells defects classification by using a [21] O</article-title>
          .
          <string-name>
            <surname>Ronneberger</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Fischer</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Brox</surname>
          </string-name>
          , U-net:
          <article-title>Convolunew feature extraction algorithm and an ebnn with tional networks for biomedical image segmentation, an innovative pruning algorithm</article-title>
          ,
          <source>International 2015. Journal of Intelligent Systems</source>
          <volume>36</volume>
          (
          <year>2021</year>
          )
          <fpage>2443</fpage>
          -
          <lpage>2464</lpage>
          . [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , W. Zuo,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Meng</surname>
          </string-name>
          , L. Zhang, Bedoi:
          <volume>10</volume>
          .1002/int.22386.
          <article-title>yond a gaussian denoiser: Residual learning of deep</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F.</given-names>
            <surname>Fiani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ponzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <article-title>Keeping eyes on the cnn for image denoising</article-title>
          .,
          <source>CoRR abs/1608</source>
          .03981 road:
          <article-title>Understanding driver attention and its role (2016). in safe driving</article-title>
          , in: CEUR Workshop Proceedings, [
          <volume>23</volume>
          ]
          <string-name>
            <given-names>Biomedical</given-names>
            <surname>Image Analysis Group</surname>
          </string-name>
          , Imperial College volume
          <volume>3695</volume>
          ,
          <year>2023</year>
          , p.
          <fpage>85</fpage>
          -
          <lpage>95</lpage>
          . London, Ixi dataset, https://brain-development.org/
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.</given-names>
            <surname>Iacobelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pelella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ponzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          , C. Napoli, ixi-dataset/,
          <year>2006</year>
          . RRID:
          <article-title>SCR_005839. A fast and accessible neural network based eye-</article-title>
          [24]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Gallo</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Frosio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kautz</surname>
          </string-name>
          ,
          <article-title>Loss functions tracking system for real-time psychometric and hci for image restoration with neural networks., IEEE applications</article-title>
          ,
          <source>in: CEUR Workshop Proceedings, Trans. Computational Imaging</source>
          <volume>3</volume>
          (
          <year>2017</year>
          )
          <fpage>47</fpage>
          -
          <lpage>57</lpage>
          . volume
          <volume>3870</volume>
          ,
          <year>2024</year>
          , p.
          <fpage>32</fpage>
          -
          <lpage>41</lpage>
          . [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mustafa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mikhailiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Iliescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Babbar</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Fiani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <article-title>A fully automatic visual R. K. Mantiuk, Training a task-specific image reattention estimation support system for a safer driv- construction loss, in: Proceedings of the IEEE/CVF ing experience</article-title>
          ,
          <source>in: CEUR Workshop Proceedings, Winter Conference on Applications of Computer</source>
          volume
          <volume>3695</volume>
          ,
          <year>2023</year>
          , p.
          <fpage>40</fpage>
          -
          <lpage>50</lpage>
          . Vision,
          <year>2022</year>
          , pp.
          <fpage>2319</fpage>
          -
          <lpage>2328</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>LeCun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Boser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Denker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Henderson</surname>
          </string-name>
          , R. E. [26]
          <string-name>
            <given-names>P.</given-names>
            <surname>Coupé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Yger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Prima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hellier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kervrann</surname>
          </string-name>
          , Howard,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hubbard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. D.</given-names>
            <surname>Jackel</surname>
          </string-name>
          , Backpropagation C.
          <article-title>Barillot, An optimized blockwise nonlocal means applied to handwritten zip code recognition, Neural denoising filter for 3-d magnetic resonance images</article-title>
          .,
          <source>Computation</source>
          <volume>1</volume>
          (
          <year>1989</year>
          )
          <fpage>541</fpage>
          -
          <lpage>551</lpage>
          .
          <source>IEEE Trans. Med. Imaging</source>
          <volume>27</volume>
          (
          <year>2008</year>
          )
          <fpage>425</fpage>
          -
          <lpage>441</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          , I. Sutskever,
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          , Imagenet [27]
          <string-name>
            <given-names>P.</given-names>
            <surname>Coupé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hellier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Prima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kervrann</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Barclassification with deep convolutional neural net- illot, 3d wavelet subbands mixing for image deworks, in: Advances in neural information process- noising</article-title>
          .,
          <source>Int. J. Biomedical Imaging</source>
          <year>2008</year>
          (
          <year>2008</year>
          )
          <article-title>ing systems</article-title>
          ,
          <year>2012</year>
          , pp.
          <fpage>1097</fpage>
          -
          <lpage>1105</lpage>
          . 590183:
          <fpage>1</fpage>
          -
          <lpage>590183</lpage>
          :
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sermanet</surname>
          </string-name>
          , S. Reed, [28]
          <string-name>
            <given-names>D.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. P. J.</given-names>
            <surname>Vosters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Anguelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vanhoucke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rabi- T. Tan</surname>
          </string-name>
          ,
          <article-title>Denoising of 3d magnetic resonance imnovich, Going deeper with convolutions, 2015. ages with multi-channel residual learning of con-</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Simonyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          ,
          <article-title>Very deep convolu- volutional neural network</article-title>
          .,
          <source>CoRR abs/1712</source>
          .
          <article-title>08726 tional networks for large-scale image recognition, (</article-title>
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>