A Federated Distributionally Robust Support Vector Machine with Mixture of Wasserstein Balls Ambiguity Set for Distributed Fault Diagnosis
The training of classification models for fault diagnosis tasks using geographically dispersed data is a crucial task for original equipment manufacturers (OEMs) seeking to provide long-term service contracts (LTSCs) to their customers. Due to privacy and bandwidth constraints, such models must be t...
Saved in:
Main Authors | , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
04.10.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The training of classification models for fault diagnosis tasks using
geographically dispersed data is a crucial task for original equipment
manufacturers (OEMs) seeking to provide long-term service contracts (LTSCs) to
their customers. Due to privacy and bandwidth constraints, such models must be
trained in a federated fashion. Moreover, due to harsh industrial settings the
data often suffers from feature and label uncertainty. Therefore, we study the
problem of training a distributionally robust (DR) support vector machine (SVM)
in a federated fashion over a network comprised of a central server and $G$
clients without sharing data. We consider the setting where the local data of
each client $g$ is sampled from a unique true distribution $\mathbb{P}_g$, and
the clients can only communicate with the central server. We propose a novel
Mixture of Wasserstein Balls (MoWB) ambiguity set that relies on local
Wasserstein balls centered at the empirical distribution of the data at each
client. We study theoretical aspects of the proposed ambiguity set, deriving
its out-of-sample performance guarantees and demonstrating that it naturally
allows for the separability of the DR problem. Subsequently, we propose two
distributed optimization algorithms for training the global FDR-SVM: i) a
subgradient method-based algorithm, and ii) an alternating direction method of
multipliers (ADMM)-based algorithm. We derive the optimization problems to be
solved by each client and provide closed-form expressions for the computations
performed by the central server during each iteration for both algorithms.
Finally, we thoroughly examine the performance of the proposed algorithms in a
series of numerical experiments utilizing both simulation data and popular
real-world datasets. |
---|---|
DOI: | 10.48550/arxiv.2410.03877 |