<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>neural network applications in training and inference - Some approaches and practices - Extended Abstract</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Angela Ciocan Voinea</string-name>
          <email>angela.ciocan@univ-lr.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vincent Courboulay</string-name>
          <email>vincent.courboulay@univ-lr.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>La Rochelle University - L3i</institution>
          ,
          <addr-line>Avenue Michel Crépeau, La Rochelle, 17042</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>La Rochelle University - L3i</institution>
          ,
          <addr-line>Avenue Michel Crépeau, La Rochelle, 17042</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Pôle Emploi - DSI Nantes</institution>
          ,
          <addr-line>Rue Konrad Adenauer, Nantes, 44000</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A Deep Neural Networks (DNNs) have gained tremendous importance for companies in recent times. This has become possible with the emergence of big data and the great possibility of storage of these data, on the one hand, and the computing resources which allowed the development of deeper deep learning algorithms with a considerable number of parameters for the other hand [1],[2],[3],[4]. The development of more efficient hardware, such as GPUs and TPUs, has made it possible to train DNNs faster and more efficiently, which has made them more accessible to companies of all sizes [5],[6],[7],[8],[9]. The ability of DNNs to learn from vast amounts of data and extract useful insights has made them a powerful tool for companies across various industries, and as a result, DNNs have become a key tool for companies looking to gain a competitive edge in their respective industries [10],[11],[12],[13],[14]. As DNNs continue to advance in their ability to perform complex tasks, concerns have arisen about their Overall, considering the environmental impact of DNNs development requires an exhaustive approach that considers the entire lifecycle of a model, from data collection to model deployment [19], [20],[21]. Adopting sustainable practices and technologies can help reduce the environmental impact of DNN development while still achieving high performance on various tasks. Two key phases in the lifecycle of DNNs applications that can significantly impact its development and deployment are the training and inference. Both require significant computational resources, resulting in high energy consumption and carbon emissions [22],[23],[24],[25]. Considering the environmental impact of both training and inference, and adopting sustainable practices and technologies, can help reduce the environmental footprint of DNN models throughout their entire lifecycle. This holistic approach to DNN development can contribute to a more sustainable future for artificial intelligence and technology in general. In this paper, we discuss techniques such as Model Compression, Knowledge Distillation, Transfer Learning and Data Management practices that can contribute to a more sustainable approach to DNN development ICT4S Workshop: Towards Sustainable and Energy-Aware Video Communications (SEViC), Month 06, 2023, Rennes, France</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        and
exploring
strategies for reducing
energy consumption
without compromising
performance
[
        <xref ref-type="bibr" rid="ref22">26</xref>
        ],[
        <xref ref-type="bibr" rid="ref23">27</xref>
        ],[
        <xref ref-type="bibr" rid="ref24">28</xref>
        ],[
        <xref ref-type="bibr" rid="ref25">29</xref>
        ],[
        <xref ref-type="bibr" rid="ref26">30</xref>
        ].
      </p>
      <p>This paper aims to highlight the importance of sustainability in DNNs development and to inspire further
research into more sustainable approaches to training and inference.</p>
      <p>2023 Copyright for this paper by its authors.</p>
      <p>Tim Dettmers. "The impact of convolutional neural networks on the field of computer vision." arXiv preprint arXiv:1704.06904, 2017.
Sam S. Stone, Robert D. Kirchhoff, and Marc Snir. "Evaluating the Potential of Intel's Next Unit of Computing (NUC) for Deep Learning
Research." International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA), 2019.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>OpenAI, AI</article-title>
          and Compute, Dario Amodei and Danny Hernandez, May
          <volume>16</volume>
          ,
          <year>2018</year>
          , https://openai.com/blog/ai-and-compute/ Gadepally et al.,
          <year>2019</year>
          .
          <source>AI Enabling Technologies: A Survey</source>
          ,
          <year>2019</year>
          , https://arxiv.org/abs/
          <year>1905</year>
          .03592 Yann LeCun, Yoshua Bengio, and
          <string-name>
            <given-names>Geoffrey</given-names>
            <surname>Hinton</surname>
          </string-name>
          .
          <source>"Deep learning." Nature</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Ian</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          , Yoshua Bengio, and
          <string-name>
            <given-names>Aaron</given-names>
            <surname>Courville</surname>
          </string-name>
          .
          <article-title>"Deep Learning."</article-title>
          MIT Press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Tim</given-names>
            <surname>Dettmers</surname>
          </string-name>
          , Greg Crosswhite, and
          <string-name>
            <given-names>Greg</given-names>
            <surname>Farquhar</surname>
          </string-name>
          .
          <article-title>"Theoretical and Empirical Speedup of Deep Learning on GPUs versus CPUs</article-title>
          .
          <source>" International Conference on Learning Representations (ICLR)</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Norman P.</given-names>
            <surname>Jouppi</surname>
          </string-name>
          , Cliff Young, Nishant Patil,
          <string-name>
            <surname>David Patterson</surname>
          </string-name>
          , et al.
          <article-title>"</article-title>
          <source>In-Datacenter Performance Analysis of a Tensor Processing Unit." International Symposium on Computer Architecture (ISCA)</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Volodymyr</given-names>
            <surname>Mnih</surname>
          </string-name>
          , Nicolas Heess, Alex Graves,
          <string-name>
            <given-names>Timothy</given-names>
            <surname>Lillicrap</surname>
          </string-name>
          , et al.
          <article-title>"Asynchronous Methods for Deep Reinforcement Learning."</article-title>
          <source>International Conference on Machine Learning (ICML)</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          , et al.
          <article-title>"A Survey on Continual Learning in Deep Neural Networks."</article-title>
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xie</surname>
          </string-name>
          , et al.
          <article-title>"Progressive Learning for Deep Neural Networks: A Review."</article-title>
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dong</surname>
          </string-name>
          , et al.
          <article-title>"Deep Learning in Internet of Things: A Survey."</article-title>
          <source>IEEE Communications Surveys &amp; Tutorials</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          , J. Cheng, X.
          <string-name>
            <surname>Xu</surname>
          </string-name>
          , et al.
          <article-title>"A Review of Deep Learning Based Speech Synthesis."</article-title>
          <source>IEEE Transactions on Cognitive and Developmental Systems</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Shervin</surname>
            <given-names>Minaee</given-names>
          </string-name>
          , Amirali Abdolrashidi, and
          <string-name>
            <given-names>Yao</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>"Deep Learning in Medical Imaging: A Review." arXiv preprint arXiv:</article-title>
          <year>1902</year>
          .04208,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Muhammad</given-names>
            <surname>Abdullah</surname>
          </string-name>
          <string-name>
            <given-names>Jamal</given-names>
            , John B.
            <surname>Hill</surname>
          </string-name>
          , and
          <string-name>
            <surname>Tariq</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>King</surname>
          </string-name>
          .
          <article-title>"</article-title>
          <source>Towards Sustainable Deep Learning: Perspective and Survey." arXiv preprint arXiv:2105.01653</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Vivienne</surname>
            <given-names>Sze</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu-Hsin</surname>
            <given-names>Chen</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tien-Ju Yang</surname>
          </string-name>
          , et al.
          <source>"Hardware Challenges of Deep Learning." arXiv preprint arXiv:1903.05734</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Payal</given-names>
            <surname>Dhar</surname>
          </string-name>
          .
          <source>The carbon impact of artificial intelligence</source>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>2</volume>
          ,
          <fpage>423</fpage>
          -
          <lpage>425</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Wu</surname>
          </string-name>
          et al.
          <source>Sustainable AI: Environmental Implications, Challenges and Opportunities</source>
          ,
          <year>2022</year>
          , https://arxiv.org/abs/2111.00364.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Kavya</given-names>
            <surname>Srinet</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gaurav</given-names>
            <surname>Sharma</surname>
          </string-name>
          .
          <article-title>"Greening Data Science: Methodologies and Metrics for Sustainable Machine Learning."</article-title>
          <source>In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT*)</source>
          , pp.
          <fpage>581</fpage>
          -
          <lpage>591</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Tegan</surname>
            <given-names>Maharaj</given-names>
          </string-name>
          , Nicholas Lane,
          <string-name>
            <given-names>Dong</given-names>
            <surname>Yin</surname>
          </string-name>
          , et al.
          <article-title>"Towards Energy Efficient Deep Learning: A Survey."</article-title>
          <source>arXiv preprint arXiv:2104</source>
          .
          <year>02009</year>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Dally</surname>
          </string-name>
          .
          <article-title>"Deep Compression: Compressing Deep Neural Networks with Pruning</article-title>
          ,
          <source>Trained Quantization and Huffman Coding." arXiv preprint arXiv:1510.00149</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Shoeybi</surname>
          </string-name>
          et al.
          <article-title>"Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism."</article-title>
          <source>In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          et al.
          <article-title>"Stochastic Gradient Descent with Hyperbolic-Tangent Learning Rates."</article-title>
          <source>In Proceedings of the 37th International Conference on Machine Learning (ICML)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ma</surname>
          </string-name>
          et al.
          <article-title>"Efficient Stochastic Gradient Descent with Horizontally Updated Adaptive Learning Rates."</article-title>
          <source>In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          et al.
          <article-title>"Attention Is All You Need."</article-title>
          <source>In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS)</source>
          , pp.
          <fpage>5998</fpage>
          -
          <lpage>6008</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          et al.
          <article-title>"Learning Efficient Object Detection Models with Knowledge Distillation."</article-title>
          <source>In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>D. N.</given-names>
            <surname>Domingo</surname>
          </string-name>
          et al.
          <article-title>"Large Scale Pre-training for Computer Vision: A Study on When and Why."</article-title>
          <source>In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          et al.
          <article-title>"Efficient Large-Scale Federated Learning with Differential Privacy and Weight Sparsification."</article-title>
          <source>In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          et al.
          <article-title>"Meta-Pruning: Meta Learning for Automatic Neural Network Channel Pruning."</article-title>
          <source>In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>E.</given-names>
            <surname>Chen</surname>
          </string-name>
          et al.
          <article-title>"Dropout Distillation for Resource-Efficient Inference in Large Neural Networks."</article-title>
          <source>In Proceedings of the 2021 International Conference on Learning Representations (ICLR)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>