Abstract
One of the key challenges in federated learning (FL) is local data distribution heterogeneity across clients, which may cause inconsistent feature spaces across clients. To address this issue, we propose Federated Feature Matching (FedFM), which guides each client's features to match shared category-wise anchors (landmarks in feature space). This method attempts to mitigate the negative effects of data heterogeneity in FL by aligning each client's feature space. We tackle the challenge of varying objective functions in theoretical analysis and provide convergence guarantee for FedFM. In FedFM, to mitigate the phenomenon of overlapping feature spaces across categories and enhance the effectiveness of feature matching, we propose a feature matching loss called contrastive-guiding (CG), which guides each local feature to match with the corresponding anchor while keeping away from non-corresponding anchors. Additionally, to achieve higher efficiency and flexibility, we propose a FedFM variant, called FedFM-Lite, which enables flexible trade-off between algorithm utility and communication bandwidth cost. Through extensive experiments, we demonstrate that FedFM with CG outperforms seven classical and representative works by quantitative and qualitative comparisons. FedFM-Lite can achieve better performance than state-of-the-art methods with five to ten times less communication costs.
Original language | English |
---|---|
Pages (from-to) | 4224-4239 |
Number of pages | 16 |
Journal | IEEE Transactions on Signal Processing |
Volume | 71 |
DOIs | |
State | Published - 16 Oct 2023 |
Keywords
- Federated learning
- contrastive-guiding
- data heterogeneity
- feature matching
All Science Journal Classification (ASJC) codes
- Signal Processing
- Electrical and Electronic Engineering