<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>S. Fonio);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Benchmarking Federated Learning Frameworks' Scalability</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gianluca Mittone</string-name>
          <email>gianluca.mittone@unito.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Samuele Fonio</string-name>
          <email>samuele.fonio@unito.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Robert Birke</string-name>
          <email>robert.birke@unito.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Aldinucci</string-name>
          <email>marco.aldinucci@unito.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Artificial Intelligence, Machine Learning, Federated Learning, Scalability, High-Performance Computing</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Turin</institution>
          ,
          <addr-line>Turin</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Federated learning (FL) is a distributed machine learning paradigm allowing cooperative model training between multiple parties while maintaining local data privacy. FL can be deployed at various scales, ranging from thousands of low-end devices (e.g., smartphones) to just a few high-performance infrastructures (e.g., HPCs), raising critical concerns about the scalability of state-of-the-art FL frameworks. This preliminary study evaluates the scaling performance of a representative FL framework, i.e., Flower, in high-performance controlled environments, providing insights on such frameworks from a computational performance point of view. Two public Top500 preexascale HPC infrastructures are exploited to obtain reliable and comparable results: Leonardo and MareNostrum5. Our findings suggest that the design of current FL frameworks, and especially their communication backends, may overlook computational performance, leading to poor scaling in large-scale scenarios.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Federated learning (FL) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is a machine learning (ML) paradigm allowing the cooperative training of
ML models while enforcing local data privacy. Such property is crucial when working with datasets
protected by strict privacy policies, such as medical or financial data, which are firmly ruled by the
General Data Protection Regulation (GDPR). Many communities actively investigate FL, from the ML
one to the parallel and distributed systems ones [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Much research has been done on all its aspects,
such as communication eficiency [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], model aggregation [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and cross-facility deployment [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. From a
computational point of view, training state-of-the-art ML models is becoming increasingly demanding,
with large language models (LLMs) being a well-known example [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In this regard, FL has been recently
regarded as the future way to train LLMs by many works [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], resulting in an eficient strategy to
scale computation while addressing privacy concerns. Many FL frameworks (FLFs) are available for
deploying FL systems; most are open-source, and each one targets diferent situations, from local FL
simulations to large-scale device ones. Despite many of them adopting very similar design choices,
their diferent implementation can lead to very diferent computational performance [
      </p>
      <sec id="sec-1-1">
        <title>2]. This work</title>
        <p>focuses on the single-data-center scalability of a widespread, representative FLF (i.e., Flower) on two
High-Performance Computing (HPC) infrastructures (i.e., Leonardo and MareNostrum5), providing
valuable insights into the scaling performance of current FLFs design and implementation.</p>
        <p>The contribution of this work is threefold and prone to further development. We i) present
experimental results on the strong and weak scalability of Flower on two pre-exascale HPC infrastructures,
ii discuss the design and implementation of current FLF, and iii) provide usability insights on the
Top500 HPC infrastructures in Europe. Future works will delve into larger ML models, stressing the
(M. Aldinucci)</p>
        <p>CEUR
Workshop</p>
        <p>ISSN1613-0073
communication backend of current FLFs, exploiting other HPCs with diferent microarchitectures for a
broader performance comparison, and testing the scaling of other major FLFs.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>This work provides insights into significant computation and communication bottlenecks of current
FLFs. Thanks to abundant, powerful, and reliable hardware, HPC infrastructures allow obtaining sound
and reliable performance measurements at diferent scales. At the same time, insights into the diferent
HPC characteristics are gathered by observing how they behave under the same FL workload.</p>
      <sec id="sec-2-1">
        <title>2.1. Computational Performance Metrics</title>
        <p>In HPC, strong scalability refers to a system’s capability to decrease the runtime of a fixed-size problem
as additional computational resources (e.g., processors or nodes) are allocated. Achieving perfect strong
scalability is often impeded by inter-process communication overhead, synchronization requirements,
and resource contention, which introduce ineficiencies as resources scale. In this context, the strong
scalability is evaluated by increasing the number of clients involved in the federation while keeping
the global data pool size fixed. This enables the training workload (i.e., the dataset) to be partitioned
across multiple processes, reducing training time as each client processes a smaller subset of the data.
On the other hand, weak scalability measures the system’s eficiency as workload and resources grow
proportionally. In an ideal scenario, a system with good weak scalability should maintain consistent
performance (e.g., constant runtime) as the workload per client remains steady, even as the total problem
size and resources grow. In this context, the weak scalability is evaluated by increasing the number of
clients involved in the federation while keeping the local data pool size fixed.</p>
        <p>Although this approach does not reflect a real-world FL setting where data is found locally on each
client and its size cannot be manipulated, it is still a valid representative and controlled stress test for
the FLF’s design and implementation, especially for its communication backend. Thus, learning metrics
are not considered since these experiments are not representative of any meaningful ML problem.
Furthermore, by isolating the communication component under maximum load conditions, insights
into the system’s handling of high-throughput communication demands can be gathered, which are
essential for understanding its scalability limits.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. HPC Infrastructures</title>
        <p>In the following, the main information of each HPC used in the reported experiments is provided:
• Leonardo is a pre-exascale system hosted by CINECA, Italy. It provides a performance peak of
238.70 PFlop/s (LINPACK Rmax). It consists of 4992 liquid-cooled compute nodes interconnected
through an NVIDIA Mellanox network, with Dragon Fly+, with a maximum bandwidth of
200Gbit/s between each pair of nodes. The Booster nodes used in the experiments are equipped
with a custom BullSequana X2135 ”Da Vinci” blade with an Intel Xeon 8358 Ice Lake CPU (32
cores, 2.6GHz), 512GB DDR4 RAM (8 x 64GB, 3200MHz), 4 NVIDIA custom A100 GPUs (64GB,
HBM2e), and 2 NVIDIA InfiniBand HDR 2×100Gb/s network cards.
• MareNostrum5 (accelerated partition, ACC) is a pre-exascale system hosted by BSC, Spain. It
provides a performance peak of 175.30 PFlop/s (LINPACK Rmax). It comprises 1120 compute
nodes interconnected through an Infiniband NDR fat-tree network topology. Each compute node
is equipped with 2x Intel Sapphire Rapids 8460Y+ at 2.3GHz and 40 cores each (80 cores node),
512 GB of Main memory, using DDR5, 4x Nvidia Hopper GPU with 64 GB HBM2 memory, 480GB
on NVMe storage, and 4x NDR200 (BW per node 800Gb/s).</p>
        <p>At the time of writing, Leonardo ranks 9th in the Top500 HPC ranking, while MareNostrum5 ACC
ranks 11th. Both HPCs are pre-exascale systems adopting the standard Intel+NVIDIA architecture,
exposing a similar software stack despite the diferent hardware characteristics.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Flower: A Friendly Federated AI Framework</title>
        <p>
          Flower [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] is a widespread FLF used for simulation and real-world deployment, widely accepted by the
research community and industry. It is open-source, fully implemented in Python, exploits gRPC as
a communication backend, and is based on a client-server architecture. Flower focuses on ofering a
smooth and straightforward user experience, which is in line with its design: the full Python
implementation allows ML practitioners a fast and seamless interaction with the library, and the client-server
architecture provides a well-known distributed framework with which the developer can easily interact.
gRPC as a communication backend is suficiently flexible to accommodate the dynamicity in client
participation that is typical of FL. Furthermore, Flower exhibits better computational performance
when compared with its competitors [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. Flower’s broad adoption, top-tier performance, and technical
characteristics matching most state-of-the-art FLFs (e.g., Intel OpenFL, NVIDIA Flare, FedML) make it
ideal as a representative FLF for this study.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experiments and Results</title>
      <p>
        Experiments focus on training the whole range of ResNets [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] available in PyTorch on the
CIFAR10 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] dataset. ResNets range from the smallest, ResNet-18, to the largest, ResNet-152, providing
increasingly heavier requirements from both the computational (more FLOPs required for the training)
and communication (more bits to be transmitted) perspectives. It is worth mentioning that the kernel
size for the skip connections was 1 × 1. Table 1 reports the diferent characteristics of the various
ResNets used in the reported experiments. CIFAR-10 provides 50,000 32 × 32 pixel images for training
and 10000 for testing, organized in 10 classes. The batch size used in the experiments is 256, and the
results presented are the average over 5 rounds of communication for soundness, which constitute
the entire training, with one epoch per round. The number of clients used increases by powers of 2,
e.g., 2, 4, 8, 16, 32, 64 clients. Federated averaging is used as an aggregation strategy, without using any
outer optimizer; SGD is used as a local optimizer, and no learning rate schedule is set. Each client
is provided with a full, exclusive computational node (as a consequence, with 4 GPUs). Such 1-to-1
mapping between clients and nodes is a crucial aspect of this study: by providing each client with more
than enough resources to complete its ML task, we cannot encounter computing bottlenecks on the
client side. At the same time, for simplicity, no intra-node distributed training approaches are adopted.
In this way, each client exploits just one GPU, leaving the intra-node communication times and the
diferent intra-node DNN training distribution techniques out of the equation, making the comparison
between diferent infrastructures as fair as possible.
      </p>
      <p>Figure 1a provides MareNostrum5 strong scalability results; only the smallest and biggest model
results are plotted for readability. It is noticeable that the computing workload diminishes as the number
of clients grows, as expected; however, it is interesting to see that communication soon becomes the
bottleneck of the entire computation. It is worth noting that with 16 clients, the communication time
equals the training time for ResNet18 and overcomes it for ResNet152. On the other hand, Figure 1b
25
20 ()se
m
i</p>
      <p>T
15 iton
a
c
i
n
10 umm
o</p>
      <p>C
5
0
1
2</p>
      <p>Number3of Clients (lo4g scale)
5
6
1
2</p>
      <p>Number3of Clients (lo4g scale)
5
6
(a) Strong scaling
(b) Weak scaling scenario
provides Marenostrum5 weak scalability results. While the training time varies slightly with a growing
number of clients, the communication cost exhibits the same growing behaviour as the strong scaling
scenario. This fact highlights a critical performance issue in communication management that is
unrelated to the computational aspect and is directly linked to the number of clients participating in
the training. Such issues may not be just found in Flower but also in similar FLFs.</p>
      <p>The computing capabilities these experiments require are not particularly demanding for the
underlying architectures, while the communication infrastructure is more stressed. Table 2 allows comparing
the two HPC infrastructures from both points of view for the proposed use case. Results show that
Marenostrum5’s training time is generally lower, as each node has more powerful CPUs and GPUs.
The communication time on Marenostrum5 also consistently outperforms Leonardo’s, indicating better
interconnection performances, which is expected given the diferent network topologies (fat-tree vs.
DragonFly+) and interconnection bandwidth.</p>
      <sec id="sec-3-1">
        <title>3.1. Conclusions</title>
        <p>A performance measurement of a representative FLF (i.e., Flower), on two European pre-exascale
Top500 HPC systems (Leonardo and MareNostrum5) is conducted. The same codebase is tested with
growing model sizes and growing client populations. Experimental results highlight that current FLFs
may overlook computational performance in their design, especially from the communication point
of view, which can be particularly detrimental in large-scale deployments. However, to confirm this,
a more comprehensive benchmark examining more FLFs is required, and an in-depth study of their
communication backend is necessary to establish whether such issues are derived from software design
or from the backend itself.
This work is partially supported by the EuroHPC-JU funding under grant No. 101093441, with support
from the Horizon-EuroHPC-JU-2021-COE-01 (SPACE CoE), and by the Spoke ”FutureHPC &amp; BigData”
of the ICSC - Centro Nazionale di Ricerca in ”High-Performance Computing, Big Data and Quantum
Computing”, funded by the European Union - NextGenerationEU.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Declaration on Generative AI</title>
      <sec id="sec-4-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>McMahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hampson</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. A. y Arcas</surname>
          </string-name>
          ,
          <article-title>Communication-eficient learning of deep networks from decentralized data</article-title>
          ,
          <source>in: Proceedings of the 20th International Conference on Artificial Intelligence and Statistics</source>
          , AISTATS, volume
          <volume>54</volume>
          <source>of Proceedings of Machine Learning Research, PMLR</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1273</fpage>
          -
          <lpage>1282</lpage>
          . URL: http://proceedings.mlr.
          <source>press/v54/mcmahan17a.html.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Mittone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tonci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Birke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Colonnelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Medic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bartolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Esposito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Parisi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Beneventi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Polato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Torquati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Benini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aldinucci</surname>
          </string-name>
          ,
          <article-title>Experimenting with emerging RISC-V systems for decentralised machine learning</article-title>
          ,
          <source>in: Proceedings of the 20th ACM International Conference on Computing Frontiers</source>
          ,
          <string-name>
            <surname>CF</surname>
          </string-name>
          , ACM,
          <year>2023</year>
          , pp.
          <fpage>73</fpage>
          -
          <lpage>83</lpage>
          . doi:
          <volume>10</volume>
          .1145/3587135.3592211.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          , Fedkd:
          <article-title>Communication eficient federated learning via knowledge distillation</article-title>
          ,
          <source>CoRR abs/2108</source>
          .13323 (
          <year>2021</year>
          ). URL: https://arxiv.org/abs/2108.13323. arXiv:
          <volume>2108</volume>
          .
          <fpage>13323</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S. P.</given-names>
            <surname>Karimireddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mohri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Reddi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. U.</given-names>
            <surname>Stich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Suresh</surname>
          </string-name>
          ,
          <article-title>SCAFFOLD: stochastic controlled averaging for federated learning</article-title>
          ,
          <source>in: Proceedings of the 37th International Conference on Machine Learning</source>
          , ICML, volume
          <volume>119</volume>
          <source>of Proceedings of Machine Learning Research, PMLR</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>5132</fpage>
          -
          <lpage>5143</lpage>
          . URL: http://proceedings.mlr.press/v119/karimireddy20a.html.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>I. Colonnelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Birke</surname>
          </string-name>
          , G. Malenza,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mittone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mulone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Galjaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bassini</surname>
          </string-name>
          , G. Scipione,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martinovič</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vondrák</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aldinucci</surname>
          </string-name>
          ,
          <article-title>Cross-facility federated learning</article-title>
          ,
          <source>Procedia Computer Science</source>
          <volume>240</volume>
          (
          <year>2024</year>
          )
          <fpage>3</fpage>
          -
          <lpage>12</lpage>
          . doi:https://doi.org/10.1016/j.procs.
          <year>2024</year>
          .
          <volume>07</volume>
          .003,
          <article-title>proceedings of the First EuroHPC user day</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Sevilla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Heim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Besiroglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hobbhahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Villalobos</surname>
          </string-name>
          ,
          <article-title>Compute trends across three eras of machine learning</article-title>
          ,
          <source>in: International Joint Conference on Neural Networks</source>
          ,
          <string-name>
            <surname>IJCNN</surname>
          </string-name>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . doi:
          <volume>10</volume>
          .1109/IJCNN55064.
          <year>2022</year>
          .
          <volume>9891914</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Sani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Iacob</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Marino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Paulik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. F.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Aleksandrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Lane</surname>
          </string-name>
          ,
          <article-title>The future of large language model pre-training is federated</article-title>
          ,
          <source>CoRR abs/2405</source>
          .10853 (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .48550/ARXIV.2405.10853. arXiv:
          <volume>2405</volume>
          .
          <fpage>10853</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Beutel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Topal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mathur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Parcollet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Lane</surname>
          </string-name>
          ,
          <article-title>Flower: A friendly federated learning research framework</article-title>
          , CoRR abs/
          <year>2007</year>
          .14390 (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2007</year>
          .14390. arXiv:
          <year>2007</year>
          .14390.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fonio</surname>
          </string-name>
          ,
          <article-title>Benchmarking federated learning frameworks for medical imaging tasks</article-title>
          ,
          <source>in: Image Analysis and Processing - ICIAP Proceedings</source>
          ,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , volume
          <volume>14366</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2023</year>
          , pp.
          <fpage>223</fpage>
          -
          <lpage>232</lpage>
          . doi:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>031</fpage>
          - 51026- 7\_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Deep residual learning for image recognition</article-title>
          ,
          <source>in: 2016 IEEE Conference on Computer Vision</source>
          and Pattern Recognition,
          <string-name>
            <surname>CVPR</surname>
          </string-name>
          , IEEE Computer Society,
          <year>2016</year>
          , pp.
          <fpage>770</fpage>
          -
          <lpage>778</lpage>
          . doi:
          <volume>10</volume>
          .1109/CVPR.
          <year>2016</year>
          .
          <volume>90</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          ,
          <article-title>Learning multiple layers of features from tiny images</article-title>
          ,
          <year>2009</year>
          . URL: https://www.cs. toronto.edu/~kriz/learning-features-2009
          <source>-TR.pdf.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>