Goals
Recent advances in deep learning, especially the rise of large language models, have greatly impacted fields such as urban planning and healthcare. Data now come from many sources, including structured tabular records and graphs that capture complex relationships between samples. This diversity has increased the need for scalable and privacy-preserving learning methods. Federated Learning (FL) is a powerful tool for aggregating knowledge from distributed data while protecting privacy. However, applying FL to new areas like large-scale language modeling, multi-agent systems, and AI-driven scientific discovery requires new methods, as these fields face challenges in scalability, trustworthiness, and robustness. This workshop aims to encourage discussion on distributed data mining and graph analytics in this evolving technological landscape, and it invites researchers and practitioners to share innovative algorithms, system designs, applications, and evaluation strategies that address these challenges and drive the development of trustworthy intelligent systems.
Organizers
Program Committee Members
.
- Yue Tan (University of Technology Sydney)
- Lun Wang (Google)
- Arun Ganesh (Google)
- Jian Xu (Tsinghua University)
- Jingtao Li (Sony AI)
- Xuefeng Jiang (Institute of Computing Technology, Chinese Academy of Sciences)
- Weiming Zhuang (Sony Research)
- Guangjing Wang (Michigan State University)
- Zhaozhuo Xu (Stevens Institute of Technology)
- Sebastian U Stich (CISPA Helmholtz Center for Information Security)
- Ruixuan Liu (Emory University)
- Yuyang Deng (Pennsylvania State University)
- Siqi Liang (Michigan State University)
- Krishna Kanth Nakka (Huawei Technologies Ltd.)
- Bing Luo (Duke Kunshan University)
- Shuyang Yu (Michigan State University)
- Graham Cormode (Facebook)
- Andrew Hard (Google)
- Yuhang Yao (Carnegie Mellon University)
- Haobo Zhang (Michigan State University)