Roadmap of Federated Learning: from Motivation to Practice

Presented by:
Dr. Jiayu Zhou , University of Michigan
Dr. Steve Drew, University of Calgary
Guojun Tang, University of Calgary

Join us to explore how federated learning enables private, collaborative model training across distributed data

Federated learning (FL), unlike traditional centralized learning methods, provides a distributed machine learning (ML) paradigm that enables different clients to train a global model collaboratively without directly uploading or sharing their private datasets. Due to its feasibility, FL has achieved outstanding success in a wide range of applications, from edge computing and the Internet of Things (IoT) to healthcare and finance.

In this tutorial, we will present a comprehensive roadmap of federated learning, including (1) the motivation of FL, (2) current challenges of FL, (3) practical applications of FL, and (4) hands-on experiments on FL. By this tutorial, the intended audience may learn about the history of federated learning and the cutting-edge algorithms in FL.

Tutorial Outline

We first dive into FL based on a real-world scenario: how we train a keyboard input prediction model among tremendous users’ devices. In traditional centralized machine learning methods, we require users to upload their data to participate in the training, which will violate the data privacy and cause significant communication overhead. To address the aforementioned issue, Google proposed a pioneering decentralized machine learning framework entitled Federated Learning that only requires users to upload the model parameters instead of raw data to participate in the model training. We will introduce the details of FedAvg, the most typical FL algorithm, and its efficacy in this scenario. In the rest of this section, we will talk about other concepts of FL, such as the taxonomy of FL and the current popular FL software frameworks.

In the second section, we will discuss the current challenges in FL and their corresponding solutions as follows:

  1. Data heterogeneity: Clients hold data drawn from different distributions. For example, some specific classes of samples concentrate on few clients, while other clients only hold very limited data of them. Data heterogeneity can often slow down or even prevent model convergence. The typical mitigations of this issue include using a regularization term [5], knowledge distillation [12], model contrastive loss [3], and control-variate correction [4].
  2. System heterogeneity: Clients in FL usually differ in computing resources, power, and network quality. It leads to the drop-off and even bias against weaker devices in the system. We may alleviate it by setting up a proper system design [6] or empowering the weaker devices to train a thinner sub-layer of the model [11].
  3. Efficiency: FL is a distributed system, which requires considering the communication efficiency. There are some cutting-edge algorithms to cut off the communication overhead by using the prototype training [7] and one-shot FL [15] so that clients and servers only interact with each other in one round of communication.
  4. There is still a probability that the malicious user may obtain the original data from the model uploaded by clients [8]. The typical methods used to enhance security include homomorphic encryption [9] and secure aggregation [10].

In the last section, there are some practical applications of FL. In the first case study we will present how FL improves the productivity of edge computing. In the following case, we will talk about a real-world application of FL in finance, how to utilize federate learning to assist the finance crime detection. The last case will demonstrate the FL application in healthcare scenarios.

Timeline

  1. Introduction (60 mins)
    1. Motivation of FL
    2. Introduction of FedAvg [1]
    3. Taxonomy of FL (PFL vs global model, Cross-devices vs Cross-silos, and VFL and HFL)
    4. Popular FL software frameworks
  2. Challenges (60 mins)
    1. Data heterogeneity [2,3,4,5,12]
    2. System heterogeneity [6, 7, 11]
    3. Efficiency [7, 15]
    4. Privacy [8,9,10]
  3. Practical study cases (30 mins)
    1. IoT [13]
    2. Finance [14]
    3. Healthcare (cross-silos with multi-modal case)

References

[1] McMahan B, Moore E, Ramage D, et al. Communication-efficient learning of deep networks from decentralized data[C]//Artificial intelligence and statistics. PMLR, 2017: 1273-1282.
[2] Zhao Y, Li M, Lai L, et al. Federated learning with non-iid data[J]. arXiv preprint arXiv:1806.00582, 2018.
[3] Li Q, He B, Song D. Model-contrastive federated learning[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021: 10713-10722.
[4] Karimireddy S P, Kale S, Mohri M, et al. Scaffold: Stochastic controlled averaging for federated learning[C]//International conference on machine learning. PMLR, 2020: 5132-5143.
[5] Li T, Sahu A K, Zaheer M, et al. Federated optimization in heterogeneous networks[J]. Proceedings of Machine learning and systems, 2020, 2: 429-450.
[6] Bonawitz K, Eichner H, Grieskamp W, et al. Towards federated learning at scale: System design[J]. Proceedings of machine learning and systems, 2019, 1: 374-388.
[7] Tan Y, Long G, Liu L, et al. Fedproto: Federated prototype learning across heterogeneous clients[C]//Proceedings of the AAAI conference on artificial intelligence. 2022, 36(8): 8432-8440.
[8] Zhu L, Liu Z, Han S. Deep leakage from gradients[J]. Advances in neural information processing systems, 2019, 32.
[9] Zhang C, Li S, Xia J, et al. {BatchCrypt}: Efficient homomorphic encryption for {Cross-Silo} federated learning[C]//2020 USENIX annual technical conference (USENIX ATC 20). 2020: 493-506.
[10] Bonawitz K, Ivanov V, Kreuter B, et al. Practical secure aggregation for privacy-preserving machine learning[C]//proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security. 2017: 1175-1191.
[11] Diao E, Ding J, Tarokh V. Heterofl: Computation and communication efficient federated learning for heterogeneous clients[J]. arXiv preprint arXiv:2010.01264, 2020.
[12] Zhu Z, Hong J, Zhou J. Data-free knowledge distillation for heterogeneous federated learning[C]//International conference on machine learning. PMLR, 2021: 12878-12889.
[13] Wu J, Dong F, Leung H, et al. Topology-aware federated learning in edge computing: A comprehensive survey[J]. ACM Computing Surveys, 2024, 56(10): 1-41.
[14] Zhang H, Hong J, Dong F, et al. A privacy-preserving hybrid federated learning framework for financial crime detection[J]. arXiv preprint arXiv:2302.03654, 2023.
[15] Jhunjhunwala D, Wang S, Joshi G. Fedfisher: Leveraging fisher information for one-shot federated learning[C]//International Conference on Artificial Intelligence and Statistics. PMLR, 2024: 1612-1620.


Presenters

Dr. Jiayu Zhou
Dr. Jiayu Zhou
University of Michigan, United States

Biography

Jiayu Zhou is an Associate Professor in the School of Information at the University of Michigan. Before joining Michigan, he was a professor of Computer Science at Michigan State University. He received his Ph.D. in Computer Science from Arizona State University in 2014. Jiayu’s research spans large-scale machine learning, data mining, and biomedical informatics, addressing both foundational methods and their impactful applications. He has served as a technical program committee member for leading international conferences, including NeurIPS, ICML, and SIGKDD, and as an Associate Editor for journals including ACM Transactions on Computing for Healthcare, ACM SIGKDD Explorations, Neurocomputing (Elsevier), and the Journal of Alzheimer’s Disease (JAD). His research is generously supported by prestigious institutions, including the National Science Foundation (NSF), the National Institutes of Health (NIH), and the Office of Naval Research (ONR). He is notably recognized with the NSF CAREER Award (2018).

Jiayu’s work has received notable recognitions, including the Best Student Paper Award at the IEEE International Conference on Data Mining (ICDM 2014) and the International Symposium on Biomedical Imaging (ISBI 2016), as well as the Best Paper Award at the IEEE International Conference on Big Data (BigData 2016). Additionally, his team received the Best Paper Award in the Health Track at the 2022 SIGKDD Conference on Knowledge Discovery and Data Mining. Most recently, Jiayu was recognized as one of the winners of the NSF/NIST Privacy-Enhancing Technologies Challenge, showcasing his innovations in privacy-preserving machine learning at the Summit for Democracy, demonstrating a commitment to reinforcing democratic values.

Dr. Steve Drew
Dr. Steve Drew
University of Calgary, Canada

Biography

Dr. Steve Drew is an Assistant Professor at the Department of Electrical and Software Engineering, University of Calgary. His research areas include distributed systems, machine learning, cloud/edge computing, and blockchain. He has over ten years of experience in the industry. He worked for Cisco Systems on cloud and edge service orchestration. He was one of the final winners of the NSF/NIST Privacy-Enhancing Technologies Challenges, where his privacy-preserving machine learning innovation was showcased at the Summit of Democracy as a testament to reinforcing democratic values.

Guojun Tang
Guojun Tang
University of Calgary, Canada

Biography

Guojun Tang is a PhD student from the University of Calgary under the supervision of Dr. Steve Drew. His research interests include federated learning, private computing, and data mining.

IEEE SWC 2025 Colocated Conferences

IEEE Digital Twin 2025

The 2025 IEEE Digital Twin Conference
View

IEEE MVS 2025

The IEEE International Conference on Metaverse 2025
View

IEEE ScalCom 2025

The 25th IEEE International Conference on Scalable Computing and Communications
View

IEEE UIC 2025

The 22nd IEEE International Conference on Ubiquitous Intelligence and Computing
View

IEEE ATC 2025

The 22nd IEEE Autonomous and Trusted Vehicle Conference
View

Sponsors