Cds
A Contribution-based Device Selection Scheme in Federated Learning. We propose a shapley-based calulation method to measure contribution from FL clients.
Install / Use
/learn @lamnd09/CdsREADME
A Contribution-based Device Selection Scheme in Federated Learning
This respository introduces a client selection based on the individual contribition in Federated Learning.
Abstract:
In a Federated Learning (FL) setup, a number of devices contribute to the training of a common model. We present a method for selecting the devices that provide updates in order to achieve improved generalization, fast convergence, and better device-level performance. We formulate a min-max optimization problem and decompose it into a primal-dual setup, where the duality gap is used to quantify the device-level performance. Our strategy combines exploration of data freshness through a random device selection with exploitation through simplified estimates of device contributions. This improves the performance of the trained model both in terms of generalization and personalization. A modified Truncated Monte-Carlo (TMC) method is applied during the exploitation phase to estimate the device's contribution and lower the communication overhead. The experimental results show that the proposed approach has a competitive performance, with lower communication overhead and competitive personalization performance against the baseline schemes.
Results:

We have presented a simplified solution of the device selection problem in Federated Learning (FL), aiming to jointly improve model performance and lower the communication costs. In doing so, we have first formulated a min-max optimization problem. to solve it, we have developed a strategy that constitutes a mixture of exploration phase, where random selection of devices is made, similar to the plain FL approach but under a primal-dual setting of the learning problem, and an additional exploitation scheme that quantifies the contribution of selected devices in improving the model performance via efficient model aggregation. Extensive simulations on realworld dataset have demonstrated the efficacy of the proposed approach against the baselines in improving model performance, i.e., better generalization and personalization, lowering communication costs, and achieving fast convergence rate.
How to use
The full code will be published and guided to use here after revision.
System Requirements:
- Pytorch
- Python ^3.7.0
- numpy, pickle, tqdm
Citation:
Pandey, Shashi Raj, Lam D. Nguyen, and Petar Popovski. "A Contribution-based Device Selection Scheme in Federated Learning." arXiv preprint arXiv:2203.05369 (2022). Link: https://arxiv.org/pdf/2203.05369.pdf
#References:
- Federated Learning Pytorch: https://github.com/AshwinRJ/Federated-Learning-PyTorch
- FedFomo: https://github.com/NVlabs/FedFomo
Related Skills
claude-opus-4-5-migration
110.6kMigrate prompts and code from Claude Sonnet 4.0, Sonnet 4.5, or Opus 4.1 to Opus 4.5
model-usage
351.2kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
TrendRadar
51.1k⭐AI-driven public opinion & trend monitor with multi-platform aggregation, RSS, and smart alerts.🎯 告别信息过载,你的 AI 舆情监控助手与热点筛选工具!聚合多平台热点 + RSS 订阅,支持关键词精准筛选。AI 智能筛选新闻 + AI 翻译 + AI 分析简报直推手机,也支持接入 MCP 架构,赋能 AI 自然语言对话分析、情感洞察与趋势预测等。支持 Docker ,数据本地/云端自持。集成微信/飞书/钉钉/Telegram/邮件/ntfy/bark/slack 等渠道智能推送。
mcp-for-beginners
15.8kThis open-source curriculum introduces the fundamentals of Model Context Protocol (MCP) through real-world, cross-language examples in .NET, Java, TypeScript, JavaScript, Rust and Python. Designed for developers, it focuses on practical techniques for building modular, scalable, and secure AI workflows from session setup to service orchestration.
