<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Deep Transfer Hashing for Adaptive Learning on Federated Streaming Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sample. Hash. Adapt. Repeat.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manuel Röder</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Frank-Michael Schleif</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Artificial Intelligence and Robotics Würzburg</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Computer Science and Business Information Systems, TUAS Würzburg-Schweinfurt</institution>
          ,
          <addr-line>Würzburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Faculty of Technology, Bielefeld University</institution>
          ,
          <addr-line>Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>7</fpage>
      <lpage>11</lpage>
      <abstract>
        <p>This extended abstract explores the integration of federated learning with deep transfer hashing for distributed prediction tasks, emphasizing resource-eficient client training from evolving data streams. Federated learning allows multiple clients to collaboratively train a shared model while maintaining data privacy - by incorporating deep transfer hashing, high-dimensional data can be converted into compact hash codes, reducing data transmission size and network loads. The proposed framework utilizes transfer learning, pre-training deep neural networks on a central server, and fine-tuning on clients to enhance model accuracy and adaptability. A selective hash code sharing mechanism using a privacy-preserving global memory bank further supports client fine-tuning. This approach addresses challenges in previous research by improving computational eficiency and scalability. Practical applications include Car2X event predictions, where a shared model is collectively trained to recognize trafic patterns, aiding in tasks such as trafic density assessment and accident detection. The research aims to develop a robust framework that combines federated learning, deep transfer hashing and transfer learning for eficient and secure downstream task execution.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Federated Learning</kwd>
        <kwd>Streaming Data</kwd>
        <kwd>Deep Transfer Hashing</kwd>
        <kwd>Real World Deployment</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction and Background</title>
      <p>
        The rapid growth of data and the increasing emphasis on privacy-preserving machine learning
techniques have spurred significant interest in federated learning (FL) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This extended abstract explores
the integration of FL with deep transfer hashing (DTH) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] methods for distributed downstream
classification and retrieval tasks, focusing on resource-aware FL client training from evolving data streams [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
and leveraging transfer learning through pre-training deep neural network models on the FL server
and employ the learned model weights for client model initialization. In addition, the client fine-tuning
process is further supported by a selective hash code sharing mechanism through the use of a globally
available but privacy preserving memory bank. The overarching goal of this concept paper is to present
and elaborate on a novel idea that addresses challenges identified in previous research [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ] and to
initiate further discussions.
      </p>
      <p>FL is a distributed machine learning paradigm where multiple clients collaboratively train a shared
model while keeping their data decentralized and secure. This approach is particularly beneficial for
applications that require strict data protection and security measures. By combining federated learning
with deep transfer hashing techniques, we aim to eficiently convert high-dimensional data into compact,
low-dimensional hash codes, significantly decreasing data transmission size between the FL server and
clients, reducing network transfer loads, and potentially improving client inference eficiency. Deep
transfer hashing methods have proven to be highly efective in reducing the dimensionality of data
while preserving its intrinsic structure. This capability is crucial for classification tasks, where the
highdimensional nature of data often poses significant computational challenges. In this context, locality
traffic scene under observation
traffic input
stream
server
feature
extractor
hashing layer
server
&lt;&lt; feedback &gt;&gt;
&lt;&lt; update &gt;&gt;
feature
extractor
hashing layer
local memory bank</p>
      <p>client
&lt;&lt; share &gt;&gt;
sensitive hashing and learning to hash approaches have been widely used. However, traditional locality
sensitive hashing methods require the construction and administration of numerous hash tables, which
can be impractical for distributed optimisation tasks such as those observed in FL. Learning to hash
potentially provides a more scalable and eficient solution by leveraging the powerful representation
capabilities of deep learning models to learn complex hash functions in an end-to-end manner. To further
enhance the performance of our proposed framework, we employ transfer learning by pre-training the
DNN model on the high-performance server. This pre-trained model can then be fine-tuned on client
devices using their local data streams, ensuring that the model adapts to the specific characteristics of
each client’s data. This approach not only accelerates the training process but also improves the overall
accuracy and generalization capability of the model.</p>
      <p>
        We aim to integrate our approach in practical scenarios that involve raw or pre-processed data points
from monitoring sensors installed at various locations, such as trafic cameras and surveillance cameras
as seen in Figure 1. In these scenarios, models can learn to recognize patterns, objects, or anomalies.
For example, in a Car2X [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] driven application, our concept can support various use-case areas such as
Intersection Movement Assist, Intersection Collision Avoidance or Green Light Optimal Speed Advisory
by enabling models to distinguish between vehicle types, assess real-time trafic density and detect and
alert about accidents. In summary, our research aims to combine the strengths of FL, DTH and transfer
learning to develop a robust and eficient framework for downstream classification and retrieval tasks,
while adhering to the constraints imposed by FL.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>We consider a FL environment composed of a central, resource-heavy server  tasked with network
orchestration and multiple resource-restricted clients indexed by , where  = 1, . . .  as outlined in
Fig. 1. Additionally, client  inspects sample , seen only once, from the non-I.I.D. data stream  at
time , in which the occurrence of objects is not evenly distributed, using an arbitrary sensing device.</p>
      <p>
        In the preparation phase, a task-specific deep neural network model is pre-trained on the server to
learn a task-specific hash function ℎ that maps an input  to a binary code , facilitating the nearest
neighbor search used for prediction inference. A well-designed hash function should preserve the
relative distances between items in the original space, meaning that items close to a specific query
in Hamming space should also be close to the query in the original space [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Subsequently, the FL
server distributes the learned model weights at the start of a new FL round and also initializes the
global memory bank with learned hash codes, raising Open Question 1. The establishment of a global
memory bank, which is fed with hash codes by the FL server in a sophisticated manner, ofers enhanced
data protection on the one hand and a reduced network data flow on the other, as only data that is
intended to support local model training is made available. Each FL client participating in the distributed
learning process follows a simple distributed SHAR pattern for local model adaptation as outlined in
Algorithm 1:
Sample. Recall that the FL client  samples data point  from the data stream  at time . The selection
of the sampling algorithm depends on various parameters like the downstream task, the quality of the
streamed data, the underlying model and the cost of sample labeling. The authors of this work already
proposed an Active Learning-based sampling method “that identifies relevant stream observations to
optimize the underlying client model, given a local labeling budget, and performs instantaneous labeling
decisions without relying on any memory bufering strategies” [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Hence, the framework of this paper
is not bound or limited to intelligent sampling strategies and also works with heuristic approaches like
for example the selection of samples based on fixed time intervals.
      </p>
      <p>
        Hash. Participating clients receive the pre-trained model and have on-demand access to the global
memory bank , utilizing these resources to perform localized fine-tuning with the sampled data set.
This process involves each client further optimizing the hash function ℎ to better adapt to their specific
data, thereby enhancing the model’s performance for their particular fine-tuning tasks. To achieve this,
we aim to employ a pointwise-based hashing method. Recent advancements typically construct the
classification loss within the Hamming space. Specifically, these methods generate a set of central hash
codes, each associated with a specific class label. The objective is to enforce the network outputs to
converge towards their respective hash centers using various loss terms, thereby ensuring that the hash
function preserves the relative distances between items efectively [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] Our intention is to support the
local hash code learning step by enriching the adaptation phase with relevant information obtained
from the global memory bank , raising Open Question 2.
      </p>
      <p>Adapt. An important consideration in deploying a FL model for critical real-world applications like
Car2X is the phenomenon of concept drift, where the statistical properties of data points sampled from
the client data stream change over time. This drift can result from evolving user preferences, seasonal
variations, or other dynamic factors influencing the data distribution. To maintain the eficacy of our
transferred hashing algorithm, it is crucial to evaluate and implement strategies for detecting concept
drift. Integrating mechanisms to handle the changing data problem into our incremental hash code
learning process ensures that the model adapts to new patterns and continues to generate accurate and
relevant hash codes. Efective detection and adaptation techniques will help to maintain the performance
and reliability of the model, even as the underlying data distributions change over time, raising Open
Question 3.</p>
      <p>Repeat. The adaptation on local clients is repeated until the global model converges. Updates received
on the server (hash codes, model parameters) from the clients after each round of federated training
must be integrated into both the server model and the global memory bank and thereby maintaining
data integrity and data privacy, raising Open Question 4.</p>
      <p>Overall, evaluation and benchmarking of the proposed method is essential, with a clear justification
for its preference over existing techniques. A major challenge is the availability of ground truth data,
especially to assess the model’s handling of concept drift, raising Open Question 5.
Algorithm 1 Sample. Hash. Adapt. Repeat.</p>
      <p>Require: Pre-trained hash function ℎ on client , unlabeled stream of samples , global memory
bank 
1: Initialize global memory bank 
2: for each federated learning round do
3: Initialize  = 1
4: Initialize  = {} ◁ Initialize local hash code storage
5: for  ∈  do ◁ Sample from data stream
6:  = calculate_hash() ◁ Hash code generation
7:  ←  ∪ {}
8:  ←  + 1
9: end for
10: ℎ = adapt(, ) ◁ Adapt client model
11: Send update to server 
12: end for ◁ Repeat until convergence
13: return Converged global learning model, educated global memory bank</p>
    </sec>
    <sec id="sec-3">
      <title>3. Conclusion and Open Questions</title>
      <p>In this work, we presented the integration of federated learning with deep transfer hashing methods for
distributed classification and retrieval tasks. Our approach focuses on resource-aware client training
from evolving data streams, leveraging transfer learning through pre-trained models on the FL server,
and utilizing selective hash code sharing via a privacy-preserving global memory bank. This integration
is supposed to eficiently convert high-dimensional data into compact hash codes, reducing network
data transmission and improving client training and inference eficiency.</p>
      <p>
        This concept paper outlines a foundational idea aimed at addressing challenges in data privacy,
computational eficiency, and scalability in FL prediction tasks. To underline the importance and
relevance of our research, we identified and outlined a real-world use-case that enhances several areas
of Car2X application. We seek to initiate further discussions and collaborations to refine these concepts
and advance privacy-preserving machine learning techniques. By attending this workshop, we hope to
gain educated insights from the discussions on the following open questions:
Open Question 1: What hash codes to include in the global memory bank (class prototypes) and
what is the best type and structure of the memory layout (map, tree, graph)?
Open Question 2: How can external memory banks improve distributed FL learning while sticking to
all FL constraints (data privacy, communication eficiency, . . . ) ?
Open Question 3: How to properly integrate the incremental aspect in deep transfer hashing to take
account for concept drift as proposed in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] within a non-FL environment?
Open Question 4: How to adapt and integrate external memory management strategies (e.g. from
continual learning, reservoir sampling)?
Open Question 5: How can we evaluate the proposed approach to demonstrate its unique advantages,
and what specific conditions must be met for its successful application??
      </p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <p>MR is supported through the Bavarian HighTech Agenda, specifically by the Würzburg Center for
Artificial Intelligence and Robotics (CAIRO) and the ProPere THWS scholarship.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>McMahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hampson</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. A.</surname>
          </string-name>
          y. Arcas,
          <article-title>Communication-eficient learning of deep networks from decentralized data</article-title>
          , in: A.
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          Zhu (Eds.),
          <source>Proceedings of the 20th international conference on artificial intelligence and statistics</source>
          , volume
          <volume>54</volume>
          <source>of Proceedings of machine learning research</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1273</fpage>
          -
          <lpage>1282</lpage>
          . URL: https://proceedings.mlr.
          <source>press/v54/mcmahan17a.html.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. S. M.</given-names>
            <surname>Goh</surname>
          </string-name>
          ,
          <article-title>Transfer hashing: From shallow to deep</article-title>
          ,
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          <volume>29</volume>
          (
          <year>2018</year>
          )
          <fpage>6191</fpage>
          -
          <lpage>6201</lpage>
          . doi:
          <volume>10</volume>
          .1109/TNNLS.
          <year>2018</year>
          .
          <volume>2827036</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O.</given-names>
            <surname>Marfoq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Neglia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kameni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Vidal</surname>
          </string-name>
          ,
          <article-title>Federated learning for data streams</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2301</volume>
          .
          <fpage>01542</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Röder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Heller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Münch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.-M.</given-names>
            <surname>Schleif</surname>
          </string-name>
          ,
          <article-title>Eficient cross-domain federated learning by mixstyle approximation</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2312</volume>
          .
          <fpage>07064</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Röder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Münch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Raab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.-M.</given-names>
            <surname>Schleif</surname>
          </string-name>
          ,
          <article-title>Crossing domain borders with federated fewshot adaptation</article-title>
          ,
          <source>in: Proceedings of the 13th International Conference on Pattern Recognition Applications and Methods -</source>
          Volume
          <volume>1</volume>
          : ICPRAM, INSTICC, SciTePress,
          <year>2024</year>
          , pp.
          <fpage>511</fpage>
          -
          <lpage>521</lpage>
          . doi:
          <volume>10</volume>
          .5220/0012351900003654.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>C2C-CC</surname>
            ,
            <given-names>Car</given-names>
            2 Car
          </string-name>
          <string-name>
            <surname>Communication</surname>
            <given-names>Concortium</given-names>
          </string-name>
          ,
          <year>2024</year>
          . URL: https://www.car-2
          <article-title>-car</article-title>
          .org/.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>X.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.-S.</given-names>
            <surname>Hua</surname>
          </string-name>
          ,
          <article-title>A survey on deep hashing methods</article-title>
          ,
          <source>ACM Trans. Knowl. Discov. Data</source>
          <volume>17</volume>
          (
          <year>2023</year>
          ). URL: https://doi.org/10.1145/3532624. doi:
          <volume>10</volume>
          .1145/ 3532624.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Röder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.-M.</given-names>
            <surname>Schleif</surname>
          </string-name>
          ,
          <source>Sparse Uncertainty-Informed Sampling from Federated Streaming Data</source>
          ,
          <year>2024</year>
          . Accepted for publication
          <source>in proceedings of European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning</source>
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>X.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. W. Y.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>Deep incremental hashing for semantic image retrieval with concept drift</article-title>
          ,
          <source>IEEE Transactions on Big Data</source>
          <volume>9</volume>
          (
          <year>2023</year>
          )
          <fpage>1102</fpage>
          -
          <lpage>1115</lpage>
          . doi:
          <volume>10</volume>
          .1109/TBDATA.
          <year>2022</year>
          .
          <volume>3233457</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>