posted by organizer: yuhanpanda || 537 views || tracked by 2 users: [display]

FL@FM-ICME 2024 : International Workshop on Federated Learning and Foundation Models for Multi-Media


When Jul 15, 2024 - Jul 19, 2024
Where Niagara Falls, ON, Canada
Submission Deadline Mar 23, 2024
Notification Due Apr 15, 2024
Categories    artificial intelligence   machine learning   federated learning   foundation models

Call For Papers

[Call for Papers]
Foundation models (FMs) are typically associated with large language models (LLMs), like ChatGPT, and are characterized by their scale and broad applicability. While these models provide transformative capabilities, they also introduce significant challenges, particularly concerning distributed model management and related data privacy, efficiency, and scalability. The training of foundation models is data and resource intensive and the conventional methods are typically centralized; this creates significant challenges including regulatory and privacy concerns in real-world use cases. These include distributed training data, computational resources to manage distributed data repositories, and development of and alignment with regulatory guidelines (e.g., GDPR) that restrict sharing sensitive data.

Federated learning (FL) is an emerging paradigm that can mitigate these challenges by training a global but distributed model using distributed data. The extensive application of machine learning to analyze and draw insight from real-world, distributed, and sensitive data necessitates familiarity with and adoption of this relevant and timely topic within the general scientific community. As FL allows self-interested data owners to collaboratively train models, end-users can become co-creators of AI solutions. By adopting federated learning approaches, we can leverage distributed data and computing power available across different sources while respecting user privacy.

The rise of FMs amplifies the importance and relevance of FL as a crucial research direction. With FMs becoming the norm in machine learning development, the focus shifts from model architecture design to tackling the issues surrounding privacy-preserving and distributed learning. Advancements in FL methods have the potential to unlock the use of FMs, enabling efficient and scalable training while safeguarding sensitive data.

With this in mind, we invite original research contributions, position papers, and work-in-progress reports on various aspects of federated learning in the era of foundation models. Since the emergence of foundation models has been a relatively recent phenomenon, their full impact on federated learning has not yet been well explored or understood. We hope to provide a platform to facilitate interaction among students, scholars, and industry professionals from around the world to discuss the latest advancements, share insights, and identify future directions in this exciting field. The workshop topics include but are not limited to:

Theory and algorithmic foundations:
-Impact of heterogeneity in FL of large models
-Multi-stage model training (e.g., base model + fine tuning)
-Optimization advances in FL (e.g., beyond first-order and local methods)
-Prompt tuning in federated settings
-Self-supervised learning in federated settings

Leveraging foundation models to improve federated learning:
-Adaptive aggregation strategies for FL in heterogeneous environments
-Foundation model enhanced FL knowledge distillation
-Overcoming data interoperability challenges using foundation models
-Personalization of FL with foundation models

Federated learning for training and tuning foundation models:
-Fairness, bias, and interpretability challenges in FL with foundation models
-Federated transfer learning with foundation models
-FL techniques for traning large-scale foundation models
-Hardware for FL with foundation models
-Optimization algorithms for federated training of foundation models
-Privacy-preserving mechanisms in FL with foundation models
-Resource-efficient FL with foundation models
-Security and robustness considerations in FL with foundation models
-Systems and infrastructure for FL with foundation models
-Vertical federated learning with foundation models
-Vulnerabilities of FL with foundation models

[Submission Instructions]
Submitted papers must be written in English, with a maximum length limit of 6 printed pages. Papers that do not comply with the length limit will not be reviewed. Use the standard IEEE Transactions templates for Microsoft Word or LaTeX formats found at: All submitted papers will be under a single-blinded peer review for their novelty, technical quality and impact. The submissions can contain author details. Submission will be accepted via the Easychair submission website.

Easychair submission site:

For enquiries, please email to:

[Organizing Committee]
General Co-Chairs:
-Xiaoxiao Li (The University of British Columbia, Canada)
-Randy Goebel (University of Alberta, Canada)

Program Co-Chairs:
-Ross Mitchell (University of Alberta, Canada)
-Jane Z. Wang (The University of British Columbia, Canada)
-Han Yu (Nanyang Technological University, Singapore)

Related Resources

ECAI 2024   27th European Conference on Artificial Intelligence
FL-ICME 2024   ICME'24 Special Session on Trustworthy Federated Learning for Multimedia
ICANN 2024   33rd International Conference on Artificial Neural Networks
FL@FM-TheWebConf 2024   International Workshop on Federated Foundation Models for the Web 2024
JCICE 2024   2024 International Joint Conference on Information and Communication Engineering(JCICE 2024)
ICMLA 2024   23rd International Conference on Machine Learning and Applications
FL@FM-IJCNN 2024   IJCNN'24 Special Session on Trustworthy Federated Learning: in the Era of Foundation Models
ACM-Ei/Scopus-CCISS 2024   2024 International Conference on Computing, Information Science and System (CCISS 2024)
FLLM 2024   The 2nd International Conference on Foundation and Large Language Models
ECCV 2024   European Conference on Computer Vision