<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Rennes, France, June</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>on green deployment for Edge AI - Abstract</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Santiago del Rey</string-name>
          <email>santiago.del.rey@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Silverio Martínez-Fernández</string-name>
          <email>silverio.martinez@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xavier Franch</string-name>
          <email>xavier.franch@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Eriksson, B. Penzenstadler, AK. Peters, C. C. Venters. Joint Proceedings of ICT4S 2023 Doctoral Symposium</institution>
          ,
          <addr-line>Demonstrations</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>In: B. Combemale</institution>
          ,
          <addr-line>G. Mussbacher, S. Betz, A. Friday, I. Hadar, J. Sallou, I. Groher, H. Muccini, O. Le Meur, C. Herglotz, E</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universitat Politècnica de Catalunya</institution>
          ,
          <addr-line>Barcelona</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>0</volume>
      <fpage>5</fpage>
      <lpage>09</lpage>
      <abstract>
        <p>The convergence of edge computing and Artificial Intelligence, namely Edge AI, ofers many opportunities to the industry for building competitive and innovative business models. However, this new paradigm has its own challenges in terms of latency, privacy, and energy. The latter is relevant considering that current AI requires expensive computation that is hard to achieve in existing edge devices. This work reviews 20 studies published between December 2018 and March 2023 on the subject of energy eficiency for the deployment of Edge AI. Most of the publications are devoted to improving the eficient deployment of Edge AI, while only a few focus on measuring the carbon footprint and energetic impact. Our work can help researchers quickly understand the state-of-the-art and learn which topics need more research.</p>
      </abstract>
      <kwd-group>
        <kwd>edge computing</kwd>
        <kwd>energy-eficiency</kwd>
        <kwd>deployment</kwd>
        <kwd>edge AI</kwd>
        <kwd>green AI</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>(X. Franch)
CEUR
Workshop
Proceedings</p>
      <p>Energetic impact of deploying ML models
3 studies: gondi2021,dodge2022, yousefpour2023</p>
      <p>Quantify energy consumption</p>
      <p>Impact of design decisions (region, hardware, etc.)
Improve on-device energy consumption
7 studies: bateni2018, jayakodi2020, wan2020,
yang2021, wang2021, abreu2022, wang2022</p>
      <p>Optimize usage of device resources
Optimize model or hardware architectures</p>
      <p>Improve network energy consumption
8 studies: mohammed2020, manasi2020, kim2020,
yang2020, sun2020, guler2021, kim2021, yosuf2021</p>
      <p>Where and how to perform the training/inference</p>
      <p>Worker scheduling
Challenges in green deployment for Edge AI
2 studies: tao2020, fraga-lamas2021</p>
      <p>
        Review of current techniques
Formulate challenges from a sustainable point of view
1.1. Energetic impact of deploying ML models (3 studies)
We find that little work has been done on analyzing the energetic impact of deploying ML
models. Dodge et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] show that the two most impactful factors on carbon footprint are
geographical location and time of day, in this order. Hence, they propose two scheduling
methods to optimize cloud workloads based on the time of day. Gondi and Pratap [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] evaluate
the energy-accuracy trade-of of Automatic Speech Recognition (ASR) transformer models on
an edge device. Their results show an exponential growth in CPU energy consumption as the
word error rate (WER) improves linearly. Yousefpour et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] quantify the carbon footprint of
Federated Learning (FL). They find that asynchronous FL is faster than synchronous FL, but has
higher carbon emissions. Moreover, they find that the overall benefits of higher concurrency
(i.e., number of devices), considering resource consumption, do not scale linearly.
      </p>
    </sec>
    <sec id="sec-2">
      <title>1.2. Improve network energy consumption (8 studies)</title>
      <p>
        A significant portion of the studies reviewed proposes new frameworks to reduce energy by
optimizing where and when is the training/inference performed [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Yosuf et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] study how
to place DNN inference models in a Cloud Fog Network architecture for energy eficiency. Their
results show that significant savings can be achieved by the full utilization of edge devices.
They also found that fog servers are bypassed in favor of cloud data centers. They argue this
is caused due to the processing ineficiency and high Power Usage Efectiveness (PUE) of the
fog servers. Kim and Wu [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] propose AutoScale, a tool that can select the optimal execution
scaling decision based on the DNN characteristics, QoS and accuracy targets, underlying system
profiles, and stochastic runtime variance. They improve inference energy eficiency by 9.8 × and
1.6× compared to the baseline settings of mobile CPU and cloud ofloading.
1.3. Improve on-device energy consumption (7 studies)
Many of the studies reviewed focus on optimizing the energy consumption in the edge
devices [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Wang et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] implement an online optimization framework connecting the
asynchronous execution of federated training with application co-running to minimize energy
consumption on mobile devices. By designating the training process to run in the background
while an application is running, they can save over 60% of energy with three times faster
convergence speed compared to previous schemes. Abreu et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] present a framework to
facilitate the exploration of dedicated decision trees (DTs) and random forests (RFs) accelerators.
The proposed framework translates tree-based structures to hardware description languages.
Their approach achieves 10× power reduction compared to prior works.
      </p>
    </sec>
    <sec id="sec-3">
      <title>1.4. Current challenges (2 studies)</title>
      <p>
        Only two papers study the challenges of deploying ML models on the edge. Tao et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
review the challenges of training DNNs with FPGA. They find these challenges mainly lie in
the complexity of resource management and the requirements of both software and hardware
design knowledge. Moreover, they propose an evaluation workflow and performance metric
to consider on-chip resource usage, training eficiency, energy eficiency, and model accuracy.
Fraga-Lamas et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] provide a more general view and review the essential concepts related
to the development of Edge AI Green IoT systems and their carbon footprint, and make a list of
twelve open challenges.
      </p>
      <sec id="sec-3-1">
        <title>2. Relevance and Novelty</title>
        <p>With increased bandwidth and lower latency, edge computing promises to decentralize cloud
applications. Meanwhile, the current AI methods assume computations are conducted in a
powerful computational infrastructure, such as data centers with substantial computing and data
storage capabilities. One of the main challenges of bringing edge computing and AI together
remains in the energy constraints of edge devices.</p>
        <p>
          This poster provides a brief overview of the state-of-the-art in green deployment for Edge AI.
We find that some papers focus on very specific application areas, such as ASR [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], FL [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], or
DTs [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], while some works are more general-purpose [
          <xref ref-type="bibr" rid="ref10 ref5">5, 10</xref>
          ]. In addition, excluding the two
studies reporting current challenges, we find that 14 out of 18 papers report empirical results
and the hardware used. We find that four papers report only using mobile phones or SoCs (e.g.,
Raspberry Pi, Nvidia Jetson), and two use a combination of both. While mobile phones vary
greatly, we find that are the most commonly used devices for experimentation. Overall, we
ifnd that while most of the research is focused on improving energy eficiency by optimizing
the edge devices’ workload and communication, little work has been done on understanding
the factors impacting energy consumption and carbon footprint (e.g., time of day, underlying
hardware). This calls for putting more efort into understanding what elements contribute to
increasing energy consumption and how. This can help to tackle the problem more accurately.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Acknowledgments</title>
        <p>This work is part of the GAISSA project (TED2021-130923B-I00), which is funded by
MCIN/AEI/10.13039/501100011033 and by the European Union ”NextGenerationEU”/PRTR.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Han</surname>
          </string-name>
          , V.
          <string-name>
            <surname>C. M. Leung</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Niyato</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Edge</surname>
            <given-names>AI</given-names>
          </string-name>
          :
          <source>Convergence of Edge Computing and Artificial Intelligence</source>
          ,
          <year>2020</year>
          . doi:
          <volume>10</volume>
          .1007/
          <fpage>978</fpage>
          -981-15-6186-3.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dustdar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Zomaya</surname>
          </string-name>
          ,
          <source>Edge Intelligence: The Confluence of Edge Computing and Artificial Intelligence</source>
          ,
          <source>IEEE IoT Journal</source>
          <volume>7</volume>
          (
          <year>2020</year>
          )
          <fpage>7457</fpage>
          -
          <lpage>7469</lpage>
          . doi:
          <volume>10</volume>
          .1109/JIOT.
          <year>2020</year>
          .
          <volume>2984887</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Verdecchia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sallou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Cruz</surname>
          </string-name>
          ,
          <article-title>A systematic review of green ai</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2301</volume>
          .
          <fpage>11047</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Wohlin</surname>
          </string-name>
          ,
          <article-title>Guidelines for snowballing in systematic literature studies and a replication in software engineering</article-title>
          ,
          <source>in: Proceedings of the 18th International Conference on Evaluation and Assessment in Software Engineering, EASE '14</source>
          ,
          <year>2014</year>
          . doi:
          <volume>10</volume>
          .1145/2601248.2601268.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dodge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Prewitt</surname>
          </string-name>
          , R. Tachet des Combes, E. Odmark,
          <string-name>
            <given-names>R.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Strubell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Luccioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>DeCario</surname>
          </string-name>
          , W. Buchanan,
          <article-title>Measuring the Carbon Intensity of AI in Cloud Instances</article-title>
          , in: FAccT,
          <year>2022</year>
          , pp.
          <fpage>1877</fpage>
          -
          <lpage>1894</lpage>
          . doi:
          <volume>10</volume>
          .1145/3531146.3533234.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gondi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Pratap</surname>
          </string-name>
          ,
          <article-title>Performance and eficiency evaluation of ASR inference on the edge</article-title>
          ,
          <source>Sustainability</source>
          <volume>13</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .3390/su132212392.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yousefpour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shenoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Stock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Maeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-W.</given-names>
            <surname>Krüger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rabbat</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-J. Wu</surname>
            ,
            <given-names>I. Mironov</given-names>
          </string-name>
          ,
          <source>Green Federated Learning</source>
          ,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2303.14604.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mohammed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Albeshri</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Katib</surname>
          </string-name>
          , R. Mehmood, UbiPriSEQ-Deep Reinforcement Learning to Manage Privacy, Security, Energy, and
          <article-title>QoS in 5G IoT HetNets, Appl</article-title>
          . Sci.
          <volume>10</volume>
          (
          <year>2020</year>
          )
          <article-title>7120</article-title>
          . doi:
          <volume>10</volume>
          .3390/app10207120.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Yosuf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Alenazi</surname>
            ,
            <given-names>T. E. H.</given-names>
          </string-name>
          <string-name>
            <surname>El-Gorashi</surname>
            ,
            <given-names>J. M. H.</given-names>
          </string-name>
          <string-name>
            <surname>Elmirghani</surname>
          </string-name>
          ,
          <article-title>Energy-Eficient AI over a Virtualized Cloud Fog Network</article-title>
          , in: e-Energy,
          <year>2021</year>
          , pp.
          <fpage>328</fpage>
          -
          <lpage>334</lpage>
          . doi:
          <volume>10</volume>
          .1145/3447555.3465378.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y. G.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-J. Wu</surname>
          </string-name>
          ,
          <article-title>AutoScale: Energy Eficiency Optimization for Stochastic Edge Inference Using Reinforcement Learning</article-title>
          ,
          <source>in: MICRO-53</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1082</fpage>
          -
          <lpage>1096</lpage>
          . doi:
          <volume>10</volume>
          . 1109/MICRO50266.
          <year>2020</year>
          .
          <volume>00090</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bateni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          , C. Liu,
          <article-title>PredJoule: A Timing-Predictable Energy Optimization Framework for Deep Neural Networks</article-title>
          , in: RTSS,
          <year>2018</year>
          , pp.
          <fpage>107</fpage>
          -
          <lpage>118</lpage>
          . doi:
          <volume>10</volume>
          .1109/RTSS.
          <year>2018</year>
          .
          <volume>00020</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Energy Minimization for Federated Asynchronous Learning on Battery-Powered Mobile Devices via Application Co-running</article-title>
          ,
          <source>in: ICDCS</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>939</fpage>
          -
          <lpage>949</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICDCS54860.
          <year>2022</year>
          .
          <volume>00095</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>B.</given-names>
            <surname>Abreu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grellert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bampi</surname>
          </string-name>
          ,
          <article-title>A framework for designing power-eficient inference accelerators in tree-based learning applications</article-title>
          ,
          <source>Engineering Applications of Artificial Intelligence</source>
          <volume>109</volume>
          (
          <year>2022</year>
          )
          <article-title>104638</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.engappai.
          <year>2021</year>
          .
          <volume>104638</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tao</surname>
          </string-name>
          , R. Ma,
          <string-name>
            <surname>M.-L. Shyu</surname>
          </string-name>
          , S.-C.
          <article-title>Chen, Challenges in Energy-Eficient Deep Neural Network Training With FPGA</article-title>
          , in: CVPR Workshops,
          <year>2020</year>
          , pp.
          <fpage>400</fpage>
          -
          <lpage>401</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Fraga-Lamas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. I.</given-names>
            <surname>Lopes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Fernández-Caramés</surname>
          </string-name>
          ,
          <article-title>Green IoT and Edge AI as Key Technological Enablers for a Sustainable Digital Transition towards a Smart Circular Economy: An Industry 5.0 Use Case</article-title>
          ,
          <source>Sensors</source>
          <volume>21</volume>
          (
          <year>2021</year>
          )
          <article-title>5745</article-title>
          . doi:
          <volume>10</volume>
          .3390/s21175745.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>