A Bayesian Incentive Mechanism for Poison-Resilient Federated Learning
By: Daniel Commey , Rebecca A. Sarpong , Griffith S. Klogo and more
Potential Business Impact:
Stops bad guys from messing up shared computer learning.
Federated learning (FL) enables collaborative model training across decentralized clients while preserving data privacy. However, its open-participation nature exposes it to data-poisoning attacks, in which malicious actors submit corrupted model updates to degrade the global model. Existing defenses are often reactive, relying on statistical aggregation rules that can be computationally expensive and that typically assume an honest majority. This paper introduces a proactive, economic defense: a lightweight Bayesian incentive mechanism that makes malicious behavior economically irrational. Each training round is modeled as a Bayesian game of incomplete information in which the server, acting as the principal, uses a small, private validation dataset to verify update quality before issuing payments. The design satisfies Individual Rationality (IR) for benevolent clients, ensuring their participation is profitable, and Incentive Compatibility (IC), making poisoning an economically dominated strategy. Extensive experiments on non-IID partitions of MNIST and FashionMNIST demonstrate robustness: with 50% label-flipping adversaries on MNIST, the mechanism maintains 96.7% accuracy, only 0.3 percentage points lower than in a scenario with 30% label-flipping adversaries. This outcome is 51.7 percentage points better than standard FedAvg, which collapses under the same 50% attack. The mechanism is computationally light, budget-bounded, and readily integrates into existing FL frameworks, offering a practical route to economically robust and sustainable FL ecosystems.
Similar Papers
Find a Scapegoat: Poisoning Membership Inference Attack and Defense to Federated Learning
Cryptography and Security
Protects private data from sneaky model hackers.
Maximizing Uncertainty for Federated learning via Bayesian Optimisation-based Model Poisoning
Machine Learning (CS)
Makes AI models less trustworthy by causing confusion.
Fairness-Constrained Optimization Attack in Federated Learning
Machine Learning (CS)
Makes AI unfairly biased, even when it seems accurate.