https://www.mdu.se/

mdu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Enhancing Global Model Performance in Federated Learning With Non-IID Data Using a Data-Free Generative Diffusion Model
Chosun Univ, Dept Comp Engn, South Korea..
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems. Tallinn Univ Technol, Dept Comp Syst, Tallinn, Estonia..
Chosun Univ, Dept Comp Engn, South Korea..
Chosun Univ, Dept Comp Engn, Gwangju, South Korea..
Show others and affiliations
2024 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 12, p. 148230-148239Article in journal (Refereed) Published
Abstract [en]

Federated Learning (FL) presents a decentralized approach to machine learning, allowing multiple clients to jointly train neural networks while maintaining the privacy of their local data. However, FL faces challenges due to data heterogeneity, leading to slow convergence and reduced performance. While sharing client information can mitigate data heterogeneity, it poses a dilemma between privacy preservation and model performance improvement. This study aims to tackle the challenge of data heterogeneity, particularly for Non-Identical and Independent Distributions (Non-IID) clients, by enhancing the global model. We propose a data-free knowledge distillation method (FedDiff) to fine-tune the global model on the server. FedDiff leverages a Diffusion model as a generator to explore the input space of local models and transfer knowledge from local models to the global one. Additionally, we customize the diffusion model's data generative scheme to reduce training time. Extensive experiments demonstrate that FedDiff reduces communication rounds between clients and the server by up to 57% for CIFAR-10 classification and up to 71% for CIFAR-100 on average, compared to other state-of-the-art FL methods, while maintaining the same level of accuracy. This makes it particularly suitable for low-power devices with limitations on data transmission and reception, such as satellites and medical care devices.Furthermore, it preserves better average accuracy for all clients at the end of the training phase.

Place, publisher, year, edition, pages
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC , 2024. Vol. 12, p. 148230-148239
Keywords [en]
Federated learning, non-identical and independent distributions (Non-IID), diffusion model, knowledge distillation
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:mdh:diva-68763DOI: 10.1109/ACCESS.2024.3474056ISI: 001339094500001Scopus ID: 2-s2.0-85207114760OAI: oai:DiVA.org:mdh-68763DiVA, id: diva2:1909349
Available from: 2024-10-30 Created: 2024-10-30 Last updated: 2025-10-10Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Daneshtalab, Masoud

Search in DiVA

By author/editor
Daneshtalab, Masoud
By organisation
Embedded Systems
In the same journal
IEEE Access
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 99 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf