Data Heterogeneity
Data heterogeneity, the variability in data distributions across different sources in federated learning, significantly hinders model accuracy and convergence. Current research focuses on mitigating this challenge through various techniques, including personalized model architectures (e.g., using adaptors or subnetworks), robust aggregation methods (e.g., weighted averaging based on client performance or data characteristics), and innovative training strategies (e.g., warmup phases, loss decomposition, and adversarial training). Addressing data heterogeneity is crucial for advancing federated learning's applicability in diverse real-world scenarios, particularly in healthcare, industrial IoT, and other domains with decentralized, privacy-sensitive data.
Papers
FedSiam-DA: Dual-aggregated Federated Learning via Siamese Network under Non-IID Data
Ming Yang, Yanhan Wang, Xin Wang, Zhenyong Zhang, Xiaoming Wu, Peng Cheng
Personalized Federated Learning for Multi-task Fault Diagnosis of Rotating Machinery
Sheng Guo, Zengxiang Li, Hui Liu, Shubao Zhao, Cheng Hao Jin
Rethinking Data Heterogeneity in Federated Learning: Introducing a New Notion and Standard Benchmarks
Mahdi Morafah, Saeed Vahidian, Chen Chen, Mubarak Shah, Bill Lin
Momentum Tracking: Momentum Acceleration for Decentralized Deep Learning on Heterogeneous Data
Yuki Takezawa, Han Bao, Kenta Niwa, Ryoma Sato, Makoto Yamada