Demajh logo Demajh, Inc.

FedWCM: Unleashing the Potential of Momentum-based Federated Learning in Long-Tailed Scenarios: what it means for business leaders

FedWCM rescues momentum-based federated learning from severely long-tailed, non-IID data by dynamically re-weighting global momentum, restoring convergence and accelerating accuracy gains at scale without extra compute overhead.

1. What the method is

FedWCM is an adaptive plug-in for momentum-based federated learning. Each round it gathers lightweight global statistics, computes class-aware weights, and simultaneously adjusts the momentum coefficient and how client momenta are aggregated. The calibrated update steers every participant toward directions that respect minority as well as majority classes, avoiding the unstable swings that standard momentum induces under distribution skew—all while leaving model architectures untouched and adding virtually no runtime cost.

2. Why the method was developed

Conventional momentum accelerates convergence on balanced data yet amplifies bias when global distributions are long-tailed, causing some algorithms to diverge and wasting bandwidth. The authors sought a minimal intervention that preserves momentum’s speed benefits while neutralising its class-imbalance side effects, so federated learning remains viable when rare events vastly outnumber common ones.

3. Who should care

4. How the method works

Clients train locally with classical SGD while tracking a compact signature of their gradient distribution. The server merges these signatures to estimate global head-versus-tail imbalance, then outputs two scalars: a momentum weight matrix that down-scales dominant-class directions and an adaptive decay factor that tempers historical gradients during severe bias. In the next round clients blend fresh gradients with this rebalanced global momentum, continuously correcting drift without exchanging raw data.

5. How it was evaluated

Experiments on CIFAR-10, CIFAR-100 and Tiny-ImageNet simulated 40–100 clients, varying Dirichlet heterogeneity (β = 0.1–1.0) and imbalance factors down to 0.01. Baselines included FedAvg, FedProx, SCAFFOLD, FedCM, FedDyn, FedGraB and feature-level re-balancers such as CReFF. Metrics tracked top-1 accuracy, rounds to 70 % accuracy and absolute accuracy drop under extreme skew; ablations removed either aggregation weighting or decay adaptation to gauge their individual impact.

6. How it performed

FedWCM converged where FedCM stalled, cutting required rounds by up to 45 % on moderate skew and delivering 7–12 percentage-point accuracy gains on the most imbalanced settings. Against the best non-momentum baselines it still improved accuracy by 3–5 points while matching their communication cost. Removing either adaptive component halved these gains, confirming both re-weighting and decay are essential. (Source: arXiv 2507.14980, 2025)

← Back to dossier index