Relaxed Contrastive Learning for Federated Learning

33citations
arXiv:2401.04928
33
citations
#852
in CVPR 2024
of 2716 papers
4
Top Authors
4
Data Points

Abstract

We propose a novel contrastive learning framework to effectively address the challenges of data heterogeneity in federated learning. We first analyze the inconsistency of gradient updates across clients during local training and establish its dependence on the distribution of feature representations, leading to the derivation of the supervised contrastive learning (SCL) objective to mitigate local deviations. In addition, we show that a naïve adoption of SCL in federated learning leads to representation collapse, resulting in slow convergence and limited performance gains. To address this issue, we introduce a relaxed contrastive learning loss that imposes a divergence penalty on excessively similar sample pairs within each class. This strategy prevents collapsed representations and enhances feature transferability, facilitating collaborative training and leading to significant performance improvements. Our framework outperforms all existing federated learning approaches by huge margins on the standard benchmarks through extensive experimental results.

Citation History

Jan 28, 2026
0
Feb 13, 2026
33+33
Feb 13, 2026
33
Feb 13, 2026
33