Horizontal Federated Learning

Horizontal Federated Learning (HFL) is a machine learning approach enabling collaborative model training across multiple devices without directly sharing their data. Current research emphasizes improving HFL efficiency through techniques like low-rank approximations and optimized communication strategies to reduce computational and communication costs, while also addressing vulnerabilities like backdoor attacks and developing incentive mechanisms to encourage participation. HFL's significance lies in its ability to train high-quality models on decentralized data while preserving privacy, with applications ranging from recommendation systems to e-health and wireless network optimization.

Papers