<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Workshop on Software Quality Analysis, Monitoring, Improvement, and Applications, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Data Partitioning Efects in Federated Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mirwais Ahmadzai</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giang Nguyen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Informatics and Information Technologies, STU in Bratislava</institution>
          ,
          <addr-line>Ilkovičova 2, Bratislava 84216</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>1</volume>
      <fpage>0</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>Federated learning is a potential ML approach that promotes cooperative learning among many distributed systems while ensuring data privacy. In this study, we present a wide review of the design and evaluation of FL, with a particular focus on data partitioning. We discuss the challenges and solutions associated with FL implementation and demonstrate the design and execution of our proposed FL architecture. The main contribution of this paper is an investigation of data partitioning in FL and its impact on system performance. Using real-world public opinion data, we evaluate our proposed FL architecture and investigate performance measures such as binary accuracy, F1 score, loss, communication overhead, and data transmission between the server and clients. The experimental results provide useful information on the efective use of FL in various contexts. We underline the distinct advantages of various data partitioning algorithms based on data distribution and privacy requirements. Our findings contribute to the creation of successful FL systems that protect privacy.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Data Partitioning</kwd>
        <kwd>Federated Learning</kwd>
        <kwd>Architecture</kwd>
        <kwd>Design</kwd>
        <kwd>Implementation</kwd>
        <kwd>Evaluation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>suggesting best practices for selecting and implementing data partitioning techniques in FL for
public opinion survey data.</p>
      <p>Although a variety of model performance evaluation metrics are discussed, such as
communication eficiency, model performance, privacy, system performance, system and statistics
heterogeneity, and motivitibility, our experimental evaluation only focuses on model performance
(accuracy and F1 score) and communication overhead using public opinion data. Furthermore,
this work does not address the ethical issues of FL, which would require further investigation.
An additional study is required to investigate the influence of data partitioning strategies on
other areas of FL system performance in order to provide an improved understanding of their
efects and potential best practices.</p>
      <p>In this context, the remainder of this paper is structured as follows: it starts with a brief
review of related work and highlights the diferences and contributions of our paper in Section 2.
The contribution and motivation of the research in the context of data from the public opinion
survey are described in Section 2.2. The proposed design of the FL architecture is described in
Section 3. Data partitioning in FL is discussed in Section 4. The performance evaluation of the
FL architecture using public opinion data and presenting the metrics and techniques used for
this evaluation is done in Section 5, and Section 5.1. Finally, the article concludes the work and
suggests potential research directions in Section 6.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Recent years have seen a considerable increase in the level of research on FL, and many studies
have been conducted on its application in various domains. In this section, a systematic
literature review is applied to select and highlight relevant work on its design, application, and
evaluation. The review is targeted searches in reputable databases using topic keywords to
ensure completeness. Table 1 summarizes the review according to the focus area of the study,
the method used to determine the main findings and limitations.</p>
      <p>
        The paper [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] presents a scalable production system for FL on mobile devices, with an
emphasis on the dificulties of privacy, security, and communication. Although the study
provides useful insights into the practical implementation of FL, it does not evaluate or compare
the performance of the system with other systems. The paper [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] looks at the latest developments
and problems in FL, including ways to mitigate privacy risks, without focusing on their limits.
The paper [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] proposes a practical FL method that reduces communication costs and is robust
to non-IID data distributions, but its limitations include experiments conducted on a limited
number of data sets and model architectures, as well as a lack of consideration for privacy
preservation. The paper [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] proposes an algorithm that minimizes learning loss within a given
resource budget, although it has constraints such as focusing on a certain class of ML models
and conducting experiments in a simulated environment. The paper [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] presents a complete
study of current research in the management of non-Independent and Identically Distributed
(non-IID) data on ML models in FL, but no concrete conclusions are presented.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Challenges and Solutions in Federated Learning Implementation</title>
        <p>
          Implementing FL can be dificult because it requires balancing the privacy and utility of local
data with the efectiveness of the ML process. Due to its distributed nature, FL encounters a
variety of challenges during training, including problems with communication, heterogeneity
of data and systems, and data privacy and security. In general, it requires careful consideration
when designing an FL system [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>Table 2, describes the main issues in FL architecture due to privacy requirements and data
volume, leading to limited communication in FL networks. Local updating, compression
approaches, decentralized training, and importance-based updating are some of the solutions
suggested by researchers. These strategies are designed to maintain the balance between
effective communication, convergence, and accuracy of the model. Federated networks face
the challenge of system heterogeneity, which causes participants with various
communication, processing, and storage capacities. To address this issue, asynchronous communication,
client participation, and fault tolerance are used. Client participation selects devices based on
their resources and data quality, while fault tolerance adds algorithmic redundancy or coded
computation to handle device failures.</p>
        <p>
          The existence of non-IID data throughout the network causes challenges in statistical
heterogeneity in FL. To solve this, the researchers propose employing multitask learning, measuring
heterogeneity with measures such as local dissimilarity, and representing user preferences
with personalization layers. Recent research has revealed that FL may not always provide
adequate privacy guarantees during model updates and may be vulnerable to two types of
attack, including poisoning attacks and inference attacks [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. Poisoning attacks can be
carried out during the model’s training phase or on the data. Inference attacks can occur during
model updates and expose participants’ private information to the adversary [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. There are
various privacy-preserving mechanisms, such as Secure Multiparty Computing (SMPC),
Diferential Privacy (DP), and Homomorphic Encryption (HE), that can be used in FL. By integrating
numerous parties, the SMPC maintains security. To preserve individual privacy, DP adds noise
to the data, while HE modifies the encryption parameters to protect user data.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Motivation and Contribution</title>
        <p>Federated learning has limited use in the real world despite its benefits, such as improved model
accuracy and privacy preservation. These issues can be resolved and its practical adoption
improved by looking into the efects of data partitioning in FL. This article’s goal is to investigate
how data partitioning techniques afect FL system performance with our following contributions:
1. An examination of data partitioning techniques in FL, focusing on how they afect system
performance and communication efectiveness.
2. A novel approach to choosing and putting into practice the best data partitioning strategies
for certain use cases.
3. Evaluating the efectiveness of the proposed methodology in increasing model accuracy
and decreasing communication overhead using data from public opinion surveys.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Federated Learning Architecture Design and Implementation</title>
      <p>
        The hardware and software requirements for implementing the FL architecture can vary
depending on the individual use case and the scale of the devices involved. However, in general,
clients, servers, models, and algorithms are components of the FL architecture [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Hardware
requirements include a collection of distributed devices (such as smartphones, laptops, and the
Internet of Things (IoT) devices) with enough processing power to locally train an ML model,
a server that meets certain criteria, and a reliable network connection that can interact with
the central server and other participating devices. Each client (local device) has its own data
set, which is used to train the ML model. The FL process is coordinated by the server, which
sends model updates to the clients and aggregates their updates. The server also keeps the
global model safe and secure. The model is an ML model that has been trained using the FL
method. It is usually a deep neural network that is trained and decentralized across clients. The
optimization technique used to train the model is called an algorithm. In terms of software
requirements, they are as follows [
        <xref ref-type="bibr" rid="ref22 ref23">22, 23</xref>
        ].
      </p>
      <p>Client side model evaluation and deployment</p>
      <p>No
Yes</p>
      <p>ML frameworks that support the FL process such as TensorFlow or PyTorch. A central server
software that manages the process, including model aggregation and device synchronization.
which can be built with technologies like Apache Kafka, RabbitMQ, and Redis. A client-side
software library that allows devices to participate in FL and communicate securely with the
central server. Protocols for secure and encrypted communication to protect the privacy of
data on participating devices. FL architecture design workflow for the public opinion survey
example is depicted in Fig. 1, the central server distributes the initial model parameters to all
clients. Clients train their local models with initial parameters and exchange the results with
the central server. The central server aggregates the local models and distributes the global
model to the clients.</p>
      <p>
        Depending on the particular use case and privacy restrictions, many methodologies can be
utilized to evaluate and deploy the model. Clients do a local evaluation of the model and send
the results to the central server for aggregation. The performance of the global model is then
evaluated in general by the server. As an alternative, the server validation data set can be used
for evaluation. When privacy is an issue, clients can also receive the global model that has been
aggregated for local predictions. The global model, on the other hand, can be hosted by the
server and made available as a service to clients, who can then send their data for predictions.
Data privacy, resource limitations, and the complexity of the model management process are a
few examples of considerations that impact the decision to choose client-side or server-side
evaluation and deployment [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Data Partitioning in Federated Learning</title>
      <p>
        FL partitioning distributes data across multiple parties who collaborate to increase the usefulness
of their combined data. This method overcomes the limitations of domain-specific data and
makes it easier for clients with various interests to work together. Based on data flow between
parties, FL data partitioning can include transfer learning, vertical partitioning, and horizontal
partitioning (Fig. 2). It takes careful preparation to bring together the interested parties and
partition the data in a way that produces an FL environment, as proper data partitioning is
crucial for the FL process [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
      </p>
      <p>Horizontal FL (HFL) combines data from entities with similar features but diferent samples.
In the HFL example, two research organizations (regions A and B) collect data from a public
opinion survey but are only able to share limited information because of privacy concerns. The
purpose is to develop an ML model that uses parameters such as age, gender, and service type
to predict how satisfied clients are with government services. Each organization first trains
a local model using its own data, then shares model updates with a central server to create a
global model, and then deploys the aggregated model to clients.</p>
      <p>Vertical FL (VFL) combines data from entities with the same sample IDs but distinct features.
VFL allows diferent respondents to share demographic data while maintaining the privacy of
survey responses. Each client trains a local model using local data and survey results and then
shares model updates with the server, which constructs a global model capable of generating
predictions across all attributes.</p>
      <p>Transfer FL (TFL) involves the use of a previously trained model on a similar task to improve
the performance of a new model on a new task. TFL can be applied for both VFL and HFL. TFL
involves training a model on one set of data and then fine-tuning it on another. When one
region has more data than the other, the model is trained on the bigger data set first and then
ifne-tuned on the smaller data set.</p>
      <p>Data in Region A</p>
      <p>ParticipantID
001
002
003
...</p>
      <p>Age
35
42
55
...</p>
      <p>...</p>
      <p>Data in Region B
ParticipantID
001
002
003
...</p>
      <p>Data in Region B
ParticipantID
001
002
003
...</p>
      <p>serviceType</p>
      <p>A
B
C
...
serviceType</p>
      <p>A
B
C
...</p>
      <p>Satisfactory_Level
1
4
5
...</p>
      <p>Satisfactory_Level
1
4
5
...</p>
    </sec>
    <sec id="sec-5">
      <title>5. Performance Evaluation of Data Partitioning in Federated</title>
    </sec>
    <sec id="sec-6">
      <title>Learning Architectures</title>
      <p>
        Evaluation of FL architecture is a crucial component because it allows us to measure the
eficiency of the model and make additional improvements. Evaluation metrics, methods, and
best practices for FL architecture will be discussed in this part. The metrics shown in Table 3
are frequently used to assess the efectiveness and eficiency of the FL approach. These metrics
include communication costs, model performance, system scalability and performance, attack
rates, computation and energy costs, convergence rates, statistical and system heterogeneity,
client motivation, and data and device security [
        <xref ref-type="bibr" rid="ref26 ref27">26, 27</xref>
        ].
      </p>
      <sec id="sec-6-1">
        <title>5.1. Experimental Results and Discussion</title>
        <p>The performance of HFL, VFL, and TFL was compared using data collected by the Asia
Foundation. It was a public opinion survey to obtain civilian thoughts and impressions on a variety
of Afghanistan-related issues. The data set includes survey questions and responses related to
security, governance, and country development.</p>
        <p>Due to the FL nature, clients send their local model updates to the server, which aggregate
these updates to improve the global model. Communication is a substantial bottleneck in the
FL process, especially if the network bandwidth is limited or a large number of clients are
participating. In our experiments, the quantization technique was investigated as a substantial
solution to this problem for three FL architectures (HFL, VFL, and TFL). The findings showed
in Fig. 3 that the transmission overhead without quantization was HFL:14.44 MB, VFL:0.12
MB, TFL:17.32 MB, while, the communication overhead for the three techniques decreased
significantly after applying quantization to model updates as HFL:7.22, VFL:0.06, TFL:8.70.</p>
        <p>These findings indicate the eficiency of quantification in reducing communication overhead
in FL systems. Reduced overhead can result in faster convergence, improved scalability, and
lower communication costs. However, it is critical to assess the impact of quantization on the
accuracy and loss metrics of the model. In this research, we also conducted tests for binary
classification models in horizontal, vertical, and transfer FL setups. The results showed that
the use of quantization maintained acceptable levels of accuracy and loss, making it a feasible
solution to reduce communication overhead in FL systems.</p>
        <p>Table 4, summarizes the experimental methodology and analysis performed in our study. It
includes key elements such as the description of the data set, the comparison of diferent data
partitioning methods, the specific federated learning model used, the hyperparameters chosen
for training, the evaluation metrics used, and an outline of the experimental procedure, details
about the analysis process, and the source of the results shown in Fig. 3, Fig. 4, Fig. 5, and Fig. 6.
Mentioned and presented figures represent the author’s contribution in this research.</p>
        <p>Accuracy</p>
        <p>F1 Score</p>
        <p>Loss</p>
        <p>Accuracy</p>
        <p>F1 Score</p>
        <p>Loss
0.0 0
20
4R0ounds
60
80</p>
        <p>4R0ounds
20
60
80</p>
        <p>The performance of three diferent FL approaches for HFL, VFL, and TFL is studied. The
performance of each approach is evaluated using three metrics: test loss, accuracy, and F1 score.
The findings of the HFL experiment are indicated in Fig. 4.
theTtheestinacitciualralcoyssiso0f.5t3h,eanHdFtLheteFs1t sicsor0e.6i9s, 0.90 Accuracy F1 Score Loss
0.54. Model performance improved consis- 0.88
tently throughout 90 rounds, with the test re
loss dropped to 0.32, the accuracy increased coS10.86
to 0.85, and the F1 score increased to 0.83. As /Fy
shown in Fig. 5, the first test loss for verti- rccau0.84
cal FL is 0.66, the accuracy is 0.55, and the cA
F1 score is 0.0. Over 90 rounds, the model 0.82
improved in all metrics, with the test loss
decreased to 0.30, the accuracy increased to 0.85, 0.80 0 20 40 60 80
and the F1 score increased to 0.84. Rounds
i0n.8Ft4hi6ne1a,ilnlayni,tdifaotlrhdTeaFFtLa1issnectoFirigse. 0i6s.,30t2h.,8e2th9ine2i.taiTcahcluetermsatcolyodseissl Figure 6: tThFoLr’MscoodnetlriPbeurtfioornm. ance. Source:
auimproved during 90 rounds, with the test loss
dropped to 0.30, and the accuracy and F1 score slightly increased to 0.8480, 0.8329 respectively.</p>
        <p>In summary, during 90 rounds, the three FL approaches showed a continuous improvement in
performance in all evaluation metrics. The findings show that FL can be eficiently applied to a
variety of situations, each strategy providing unique benefits based on specific data distribution
and privacy needs.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>6. Conclusion</title>
      <p>The paper provides a thorough investigation of the architecture of FL, with a particular emphasis
on data partitioning. The importance of FL has been emphasized and the problems and
limitations of existing FL techniques have been studied. The design concepts and factors necessary to
establish an FL architecture have also been investigated. FL architectures have been evaluated
using metrics and approaches related to data partitioning strategies. The implementation and
evaluation of the FL architecture was carried out using various data partitioning architectures
and the results were thoroughly explained. By evaluating the FL system using new measures,
future development of more eficient, efective, and privacy-preserving FL systems can be helped.
These measures should address statistical and system heterogeneity, system performance, client
motivation, system scalability, and data privacy, in particular. Taking these factors into account,
we can improve our understanding of FL, leading to the development of more eficient, efective,
and secure FL learning systems.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This publication has been written thanks to the support of the Operational Programme Integrated
Infrastructure for the project: International Center of Excellence for Research on Intelligent
and Secure Information and Communication Technologies and Systems – Phase II (ITMS code:
313021W404), co-funded by the European Regional Development Fund. It is also supported by
the Operational Program Integrated Infrastructure for the project: National infrastructure for
supporting technology transfer in Slovakia II – NITT SK II, co-funded by the European Regional
Development Fund, and the AI4EOSC project under grant number 101058593.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mathews</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramaswamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Beaufays</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Augenstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Eichner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kiddon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <article-title>Federated learning for mobile keyboard prediction</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1811</year>
          .03604.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bonawitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Eichner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Grieskamp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Huba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ingerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ivanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kiddon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Konečný</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mazzocchi</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>T. V.</given-names>
          </string-name>
          <string-name>
            <surname>Overveldt</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Petrou</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramage</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Roselander</surname>
          </string-name>
          ,
          <article-title>Towards federated learning at scale: System design</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1902</year>
          .01046.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kairouz</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Avent</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bellet</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bennis</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          <string-name>
            <surname>Bhagoji</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Bonawitz</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Charles</surname>
            , G. Cormode,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cummings</surname>
          </string-name>
          , et al.,
          <article-title>Advances and open problems in federated learning</article-title>
          ,
          <source>Foundations and Trends® in Machine Learning</source>
          <volume>14</volume>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>210</lpage>
          . doi:
          <volume>10</volume>
          .1561/ 2200000083.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>McMahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hampson</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. A. y Arcas</surname>
          </string-name>
          ,
          <article-title>Communication-eficient learning of deep networks from decentralized data</article-title>
          ,
          <source>in: Artificial intelligence and statistics</source>
          , PMLR,
          <year>2017</year>
          , pp.
          <fpage>1273</fpage>
          -
          <lpage>1282</lpage>
          . URL: http://proceedings.mlr.
          <source>press/v54/mcmahan17a/ mcmahan17a.pdf.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tuor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Salonidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Leung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Makaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chan</surname>
          </string-name>
          ,
          <article-title>Adaptive federated learning in resource constrained edge computing systems</article-title>
          ,
          <source>IEEE journal on selected areas in communications 37</source>
          (
          <year>2019</year>
          )
          <fpage>1205</fpage>
          -
          <lpage>1221</lpage>
          . doi:
          <volume>10</volume>
          .1109/JSAC.
          <year>2019</year>
          .
          <volume>2904348</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <article-title>Federated learning on non-iid data: A survey</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>465</volume>
          (
          <year>2021</year>
          )
          <fpage>371</fpage>
          -
          <lpage>390</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.neucom.
          <year>2021</year>
          .
          <volume>07</volume>
          .098.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Applications of federated learning in smart cities: recent advances, taxonomy</article-title>
          , and open challenges,
          <source>Connection Science</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
          . doi:
          <volume>10</volume>
          .1080/09540091.
          <year>2021</year>
          .
          <volume>1936455</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. M.</given-names>
            <surname>Mammen</surname>
          </string-name>
          ,
          <source>Federated learning: Opportunities and challenges</source>
          ,
          <year>2021</year>
          . arXiv:
          <volume>2101</volume>
          .
          <fpage>05428</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Shafahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. R.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Najibi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Suciu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Studer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Dumitras</surname>
          </string-name>
          , T. Goldstein,
          <article-title>Poison frogs! targeted clean-label poisoning attacks on neural networks</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>31</volume>
          (
          <year>2018</year>
          ). URL: https://proceedings.neurips.cc/paper_files/ paper/2018/file/22722a343513ed45f14905eb07621686-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Bhagoji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Calo</surname>
          </string-name>
          ,
          <article-title>Analyzing federated learning through an adversarial lens</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>634</fpage>
          -
          <lpage>643</lpage>
          . URL: http://proceedings.mlr.press/v97/bhagoji19a/bhagoji19a.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Melis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Song</surname>
          </string-name>
          , E. De Cristofaro,
          <string-name>
            <given-names>V.</given-names>
            <surname>Shmatikov</surname>
          </string-name>
          ,
          <article-title>Exploiting unintended feature leakage in collaborative learning</article-title>
          ,
          <source>in: 2019 IEEE symposium on security and privacy (SP)</source>
          , IEEE,
          <year>2019</year>
          , pp.
          <fpage>691</fpage>
          -
          <lpage>706</lpage>
          . doi:
          <volume>10</volume>
          .1109/SP.
          <year>2019</year>
          .
          <volume>00029</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. E.</given-names>
            <surname>Choromanska</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <article-title>LeCun, Deep learning with elastic averaging sgd</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>28</volume>
          (
          <year>2015</year>
          ). URL: https://proceedings.neurips.cc/ paper_files/paper/2015/file/d18f655c3fce66ca401d5f38b48c89af-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Dally</surname>
          </string-name>
          ,
          <article-title>Deep compression: Compressing deep neural networks with pruning, trained quantization</article-title>
          and hufman coding,
          <year>2016</year>
          . arXiv:
          <volume>1510</volume>
          .
          <fpage>00149</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>T.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. U.</given-names>
            <surname>Stich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jaggi</surname>
          </string-name>
          ,
          <article-title>Don't use large mini-batches</article-title>
          ,
          <source>use local sgd</source>
          ,
          <year>2020</year>
          . arXiv:
          <year>1808</year>
          .07217.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>esgd: Commutation eficient distributed deep learning on the edge</article-title>
          ,
          <source>HotEdge</source>
          (
          <year>2018</year>
          )
          <article-title>6</article-title>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2010/file/ abea47ba24142ed16b7d8fbf2c740e0d-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zinkevich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Weimer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Smola</surname>
          </string-name>
          ,
          <article-title>Parallelized stochastic gradient descent</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>23</volume>
          (
          <year>2010</year>
          ). URL: https://proceedings. neurips.cc/paper_files/paper/2010/file/abea47ba24142ed16b7d8fbf2c740e0d-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>T.</given-names>
            <surname>Nishio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Yonetani</surname>
          </string-name>
          ,
          <article-title>Client selection for federated learning with heterogeneous resources in mobile edge</article-title>
          , in: ICC 2019
          <article-title>-2019 IEEE international conference on communications (ICC)</article-title>
          , IEEE,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICC.
          <year>2019</year>
          .
          <volume>8761315</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>V.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-K. Chiang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Sanjabi</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Talwalkar</surname>
          </string-name>
          ,
          <article-title>Federated multi-task learning</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ). URL: https://proceedings.neurips.cc/ paper_files/paper/2017/file/6211080fa89981f66b1a0c9d55c61d0f-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>I. I.</given-names>
            <surname>Eliazar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. M.</given-names>
            <surname>Sokolov</surname>
          </string-name>
          ,
          <article-title>Measuring statistical heterogeneity: The pietra index</article-title>
          ,
          <source>Physica A: Statistical Mechanics and its Applications</source>
          <volume>389</volume>
          (
          <year>2010</year>
          )
          <fpage>117</fpage>
          -
          <lpage>125</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.physa.
          <year>2009</year>
          .
          <volume>08</volume>
          .006.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Arivazhagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Choudhary</surname>
          </string-name>
          ,
          <article-title>Federated learning with personalization layers</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1912</year>
          .00818.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bonawitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ivanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kreuter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marcedone</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Patel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramage</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Segal</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Seth</surname>
          </string-name>
          ,
          <article-title>Practical secure aggregation for privacy-preserving machine learning</article-title>
          ,
          <source>in: proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1175</fpage>
          -
          <lpage>1191</lpage>
          . doi:
          <volume>10</volume>
          .1145/3133956.3133982.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Behera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Otter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shetty</surname>
          </string-name>
          , et al.,
          <article-title>Federated learning using distributed messaging with entitlements for anonymous computation and secure delivery of model (</article-title>
          <year>2020</year>
          ). URL: https://www.academia.edu/download/70178855/25661363.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á. Morell</surname>
          </string-name>
          , E. Alba,
          <article-title>Dynamic and adaptive fault-tolerant asynchronous federated learning using volunteer edge devices</article-title>
          ,
          <source>Future Generation Computer Systems</source>
          <volume>133</volume>
          (
          <year>2022</year>
          )
          <fpage>53</fpage>
          -
          <lpage>67</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.future.
          <year>2022</year>
          .
          <volume>02</volume>
          .024.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yang</surname>
          </string-name>
          , G. Andrew,
          <string-name>
            <given-names>H.</given-names>
            <surname>Eichner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Beaufays</surname>
          </string-name>
          , Applied federated learning:
          <source>Improving google keyboard query suggestions</source>
          ,
          <year>2018</year>
          . arXiv:
          <year>1812</year>
          .02903.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>V.</given-names>
            <surname>Mothukuri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Parizi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pouriyeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dehghantanha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <article-title>A survey on security and privacy of federated learning</article-title>
          ,
          <source>Future Generation Computer Systems</source>
          <volume>115</volume>
          (
          <year>2021</year>
          )
          <fpage>619</fpage>
          -
          <lpage>640</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.future.
          <year>2020</year>
          .
          <volume>10</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>G.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Verifynet: Secure and verifiable federated learning</article-title>
          ,
          <source>IEEE Transactions on Information Forensics and Security</source>
          <volume>15</volume>
          (
          <year>2019</year>
          )
          <fpage>911</fpage>
          -
          <lpage>926</lpage>
          . doi:
          <volume>10</volume>
          .1109/ TIFS.
          <year>2019</year>
          .
          <volume>2929409</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          , H.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Paik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <article-title>A systematic literature review on federated machine learning: From a software engineering perspective</article-title>
          ,
          <source>ACM Computing Surveys (CSUR) 54</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>39</lpage>
          . doi:
          <volume>10</volume>
          .1145/3450288.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>