=Paper=
{{Paper
|id=Vol-3770/paper2
|storemode=property
|title=Deep Transfer Hashing for Adaptive Learning on Federated Streaming Data
|pdfUrl=https://ceur-ws.org/Vol-3770/paper2.pdf
|volume=Vol-3770
|authors=Manuel Röder,Frank-Michael Schleif
|dblpUrl=https://dblp.org/rec/conf/ial/RoderS24
}}
==Deep Transfer Hashing for Adaptive Learning on Federated Streaming Data==
Deep Transfer Hashing for Adaptive Learning on
Federated Streaming Data
Sample. Hash. Adapt. Repeat.
Manuel Röder1,2,3,* , Frank-Michael Schleif1
1
Faculty of Computer Science and Business Information Systems, TUAS Würzburg-Schweinfurt, Würzburg, Germany
2
Faculty of Technology, Bielefeld University, Bielefeld, Germany
3
Center for Artificial Intelligence and Robotics Würzburg, Germany
Abstract
This extended abstract explores the integration of federated learning with deep transfer hashing for distributed
prediction tasks, emphasizing resource-efficient client training from evolving data streams. Federated learning
allows multiple clients to collaboratively train a shared model while maintaining data privacy - by incorporating
deep transfer hashing, high-dimensional data can be converted into compact hash codes, reducing data trans-
mission size and network loads. The proposed framework utilizes transfer learning, pre-training deep neural
networks on a central server, and fine-tuning on clients to enhance model accuracy and adaptability. A selective
hash code sharing mechanism using a privacy-preserving global memory bank further supports client fine-tuning.
This approach addresses challenges in previous research by improving computational efficiency and scalability.
Practical applications include Car2X event predictions, where a shared model is collectively trained to recognize
traffic patterns, aiding in tasks such as traffic density assessment and accident detection. The research aims to
develop a robust framework that combines federated learning, deep transfer hashing and transfer learning for
efficient and secure downstream task execution.
Keywords
Federated Learning, Streaming Data, Deep Transfer Hashing, Real World Deployment
1. Introduction and Background
The rapid growth of data and the increasing emphasis on privacy-preserving machine learning tech-
niques have spurred significant interest in federated learning (FL) [1]. This extended abstract explores
the integration of FL with deep transfer hashing (DTH) [2] methods for distributed downstream classifi-
cation and retrieval tasks, focusing on resource-aware FL client training from evolving data streams [3]
and leveraging transfer learning through pre-training deep neural network models on the FL server
and employ the learned model weights for client model initialization. In addition, the client fine-tuning
process is further supported by a selective hash code sharing mechanism through the use of a globally
available but privacy preserving memory bank. The overarching goal of this concept paper is to present
and elaborate on a novel idea that addresses challenges identified in previous research [4, 5] and to
initiate further discussions.
FL is a distributed machine learning paradigm where multiple clients collaboratively train a shared
model while keeping their data decentralized and secure. This approach is particularly beneficial for
applications that require strict data protection and security measures. By combining federated learning
with deep transfer hashing techniques, we aim to efficiently convert high-dimensional data into compact,
low-dimensional hash codes, significantly decreasing data transmission size between the FL server and
clients, reducing network transfer loads, and potentially improving client inference efficiency. Deep
transfer hashing methods have proven to be highly effective in reducing the dimensionality of data
while preserving its intrinsic structure. This capability is crucial for classification tasks, where the high-
dimensional nature of data often poses significant computational challenges. In this context, locality
IAL@ECML-PKDD’24: 8th Intl. Worksh. & Tutorial on Interactive Adaptive Learning, Sep. 9th , 2024, Vilnius, Lithuania
*
Corresponding author.
$ manuel.roeder@thws.de (M. Röder); frank-michael.schleif@thws.de (F. Schleif)
0009-0003-4907-3999 (M. Röder); 0000-0002-7539-1283 (F. Schleif)
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
7
Manuel Röder et al. CEUR Workshop Proceedings 7–11
traffic scene under observation traffic input
stream
feature
extractor local memory bank
hashing layer
server
client
<< feedback >>
<< share >>
0 1 0 1 0 1 0 1
1 1 1 1 1 1 0 1
feature << update >> 0 1 0 1 1 1 1 1
extractor 0 1 0 0 0 0 0 1
1 1 0 0 0 0 0 1
hashing layer
global memory bank
server
Figure 1: Car2x scenario: FL clients with monitoring sensors (camera, radar) sample from a traffic input stream
at different geographic locations. The central server instance maintains a global memory bank that is accessible
for all client devices to support local client model fine-tuning. The feature extractor and hashing layer are
pre-trained on the large-scale server data set 𝐷𝑆 . The central server administrates a global memory bank.
sensitive hashing and learning to hash approaches have been widely used. However, traditional locality
sensitive hashing methods require the construction and administration of numerous hash tables, which
can be impractical for distributed optimisation tasks such as those observed in FL. Learning to hash
potentially provides a more scalable and efficient solution by leveraging the powerful representation
capabilities of deep learning models to learn complex hash functions in an end-to-end manner. To further
enhance the performance of our proposed framework, we employ transfer learning by pre-training the
DNN model on the high-performance server. This pre-trained model can then be fine-tuned on client
devices using their local data streams, ensuring that the model adapts to the specific characteristics of
each client’s data. This approach not only accelerates the training process but also improves the overall
accuracy and generalization capability of the model.
We aim to integrate our approach in practical scenarios that involve raw or pre-processed data points
from monitoring sensors installed at various locations, such as traffic cameras and surveillance cameras
as seen in Figure 1. In these scenarios, models can learn to recognize patterns, objects, or anomalies.
For example, in a Car2X [6] driven application, our concept can support various use-case areas such as
Intersection Movement Assist, Intersection Collision Avoidance or Green Light Optimal Speed Advisory
by enabling models to distinguish between vehicle types, assess real-time traffic density and detect and
alert about accidents. In summary, our research aims to combine the strengths of FL, DTH and transfer
learning to develop a robust and efficient framework for downstream classification and retrieval tasks,
while adhering to the constraints imposed by FL.
2. Methodology
We consider a FL environment composed of a central, resource-heavy server 𝑆 tasked with network
orchestration and multiple resource-restricted clients indexed by 𝑖, where 𝑖 = 1, . . . 𝐼 as outlined in
Fig. 1. Additionally, client 𝑖 inspects sample 𝑥𝑡 , seen only once, from the non-I.I.D. data stream 𝑈𝑖 at
time 𝑡, in which the occurrence of objects is not evenly distributed, using an arbitrary sensing device.
8
Manuel Röder et al. CEUR Workshop Proceedings 7–11
In the preparation phase, a task-specific deep neural network model is pre-trained on the server to
learn a task-specific hash function ℎ that maps an input 𝑥 to a binary code 𝑏, facilitating the nearest
neighbor search used for prediction inference. A well-designed hash function should preserve the
relative distances between items in the original space, meaning that items close to a specific query
in Hamming space should also be close to the query in the original space [7]. Subsequently, the FL
server distributes the learned model weights at the start of a new FL round and also initializes the
global memory bank with learned hash codes, raising Open Question 1. The establishment of a global
memory bank, which is fed with hash codes by the FL server in a sophisticated manner, offers enhanced
data protection on the one hand and a reduced network data flow on the other, as only data that is
intended to support local model training is made available. Each FL client participating in the distributed
learning process follows a simple distributed SHAR pattern for local model adaptation as outlined in
Algorithm 1:
Sample. Recall that the FL client 𝑖 samples data point 𝑥𝑡 from the data stream 𝑈𝑖 at time 𝑡. The selection
of the sampling algorithm depends on various parameters like the downstream task, the quality of the
streamed data, the underlying model and the cost of sample labeling. The authors of this work already
proposed an Active Learning-based sampling method “that identifies relevant stream observations to
optimize the underlying client model, given a local labeling budget, and performs instantaneous labeling
decisions without relying on any memory buffering strategies” [8]. Hence, the framework of this paper
is not bound or limited to intelligent sampling strategies and also works with heuristic approaches like
for example the selection of samples based on fixed time intervals.
Hash. Participating clients receive the pre-trained model and have on-demand access to the global
memory bank 𝑀𝑆 , utilizing these resources to perform localized fine-tuning with the sampled data set.
This process involves each client further optimizing the hash function ℎ𝑖 to better adapt to their specific
data, thereby enhancing the model’s performance for their particular fine-tuning tasks. To achieve this,
we aim to employ a pointwise-based hashing method. Recent advancements typically construct the
classification loss within the Hamming space. Specifically, these methods generate a set of central hash
codes, each associated with a specific class label. The objective is to enforce the network outputs to
converge towards their respective hash centers using various loss terms, thereby ensuring that the hash
function preserves the relative distances between items effectively [7] Our intention is to support the
local hash code learning step by enriching the adaptation phase with relevant information obtained
from the global memory bank 𝑀𝑆 , raising Open Question 2.
Adapt. An important consideration in deploying a FL model for critical real-world applications like
Car2X is the phenomenon of concept drift, where the statistical properties of data points sampled from
the client data stream change over time. This drift can result from evolving user preferences, seasonal
variations, or other dynamic factors influencing the data distribution. To maintain the efficacy of our
transferred hashing algorithm, it is crucial to evaluate and implement strategies for detecting concept
drift. Integrating mechanisms to handle the changing data problem into our incremental hash code
learning process ensures that the model adapts to new patterns and continues to generate accurate and
relevant hash codes. Effective detection and adaptation techniques will help to maintain the performance
and reliability of the model, even as the underlying data distributions change over time, raising Open
Question 3.
Repeat. The adaptation on local clients is repeated until the global model converges. Updates received
on the server (hash codes, model parameters) from the clients after each round of federated training
must be integrated into both the server model and the global memory bank and thereby maintaining
data integrity and data privacy, raising Open Question 4.
Overall, evaluation and benchmarking of the proposed method is essential, with a clear justification
for its preference over existing techniques. A major challenge is the availability of ground truth data,
especially to assess the model’s handling of concept drift, raising Open Question 5.
9
Manuel Röder et al. CEUR Workshop Proceedings 7–11
Algorithm 1 Sample. Hash. Adapt. Repeat.
Require: Pre-trained hash function ℎ𝑖 on client 𝑖, unlabeled stream of samples 𝑈𝑖 , global memory
bank 𝑀𝑆
1: Initialize global memory bank 𝑀𝑆
2: for each federated learning round do
3: Initialize 𝑡 = 1
4: Initialize 𝐵𝑖 = {} ◁ Initialize local hash code storage
5: for 𝑥𝑡 ∈ 𝑈𝑖 do ◁ Sample from data stream
6: 𝑏𝑡 = calculate_hash(𝑥𝑡 ) ◁ Hash code generation
7: 𝐵𝑖 ← 𝐵𝑖 ∪ {𝑏𝑡 }
8: 𝑡←𝑡+1
9: end for
10: ℎ𝑖 = adapt(𝐵𝑖 , 𝑀𝑆 ) ◁ Adapt client model
11: Send update to server 𝑆
12: end for ◁ Repeat until convergence
13: return Converged global learning model, educated global memory bank 𝑀𝑆
3. Conclusion and Open Questions
In this work, we presented the integration of federated learning with deep transfer hashing methods for
distributed classification and retrieval tasks. Our approach focuses on resource-aware client training
from evolving data streams, leveraging transfer learning through pre-trained models on the FL server,
and utilizing selective hash code sharing via a privacy-preserving global memory bank. This integration
is supposed to efficiently convert high-dimensional data into compact hash codes, reducing network
data transmission and improving client training and inference efficiency.
This concept paper outlines a foundational idea aimed at addressing challenges in data privacy,
computational efficiency, and scalability in FL prediction tasks. To underline the importance and
relevance of our research, we identified and outlined a real-world use-case that enhances several areas
of Car2X application. We seek to initiate further discussions and collaborations to refine these concepts
and advance privacy-preserving machine learning techniques. By attending this workshop, we hope to
gain educated insights from the discussions on the following open questions:
Open Question 1: What hash codes to include in the global memory bank (class prototypes) and
what is the best type and structure of the memory layout (map, tree, graph)?
Open Question 2: How can external memory banks improve distributed FL learning while sticking to
all FL constraints (data privacy, communication efficiency, . . . ) ?
Open Question 3: How to properly integrate the incremental aspect in deep transfer hashing to take
account for concept drift as proposed in [9] within a non-FL environment?
Open Question 4: How to adapt and integrate external memory management strategies (e.g. from
continual learning, reservoir sampling)?
Open Question 5: How can we evaluate the proposed approach to demonstrate its unique advantages,
and what specific conditions must be met for its successful application??
Acknowledgments
MR is supported through the Bavarian HighTech Agenda, specifically by the Würzburg Center for
Artificial Intelligence and Robotics (CAIRO) and the ProPere THWS scholarship.
10
Manuel Röder et al. CEUR Workshop Proceedings 7–11
References
[1] B. McMahan, E. Moore, D. Ramage, S. Hampson, B. A. y. Arcas, Communication-efficient learning
of deep networks from decentralized data, in: A. Singh, J. Zhu (Eds.), Proceedings of the 20th
international conference on artificial intelligence and statistics, volume 54 of Proceedings of machine
learning research, 2017, pp. 1273–1282. URL: https://proceedings.mlr.press/v54/mcmahan17a.html.
[2] J. T. Zhou, H. Zhao, X. Peng, M. Fang, Z. Qin, R. S. M. Goh, Transfer hashing: From shallow
to deep, IEEE Transactions on Neural Networks and Learning Systems 29 (2018) 6191–6201.
doi:10.1109/TNNLS.2018.2827036.
[3] O. Marfoq, G. Neglia, L. Kameni, R. Vidal, Federated learning for data streams, 2023.
arXiv:2301.01542.
[4] M. Röder, L. Heller, M. Münch, F.-M. Schleif, Efficient cross-domain federated learning by mixstyle
approximation, 2023. arXiv:2312.07064.
[5] M. Röder, M. Münch, C. Raab, F.-M. Schleif, Crossing domain borders with federated few-
shot adaptation, in: Proceedings of the 13th International Conference on Pattern Recogni-
tion Applications and Methods - Volume 1: ICPRAM, INSTICC, SciTePress, 2024, pp. 511–521.
doi:10.5220/0012351900003654.
[6] C2C-CC, Car 2 Car Communication Concortium, 2024. URL: https://www.car-2-car.org/.
[7] X. Luo, H. Wang, D. Wu, C. Chen, M. Deng, J. Huang, X.-S. Hua, A survey on deep hashing methods,
ACM Trans. Knowl. Discov. Data 17 (2023). URL: https://doi.org/10.1145/3532624. doi:10.1145/
3532624.
[8] M. Röder, F.-M. Schleif, Sparse Uncertainty-Informed Sampling from Federated Streaming Data,
2024. Accepted for publication in proceedings of European Symposium on Artificial Neural Networks,
Computational Intelligence and Machine Learning 2024.
[9] X. Tian, W. W. Y. Ng, H. Xu, Deep incremental hashing for semantic image retrieval with concept
drift, IEEE Transactions on Big Data 9 (2023) 1102–1115. doi:10.1109/TBDATA.2022.3233457.
11