AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (10.8 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Open Access | Just Accepted

Optimizing Federated Incremental Learning: Efficient Malicious Data Removal for Big Data Analytics

Kongyang Chen1,2,3Wengao Li1Jiannong Cao4,5Bing Mi6( )Jiaxing Shen7

1 School of Artificial Intelligence, Guangzhou University, Guangzhou 510006, China

2 Pazhou Lab, Guangzhou 510330, China

3 Yunnan Key Laboratory of Service Computing, Yunnan University of Finance and Economics, Kunming, 650221, China

Department of Computing, The Hong Kong Polytechnic University, Hong Kong, China

5 Research Institute for Artificial Intelligence of Things (RIAIoT), The Hong Kong Polytechnic University, Hong Kong, China

6 Guangdong University of Finance and Economics, Guangzhou 510320, China

7 Department of Computing and Decision Sciences, Lingnan University, Hong Kong, China

Show Author Information

Abstract

Federated incremental learning facilitates decentralized and continuous model updates across multiple clients, presenting a promising framework for big data analytics in distributed environments. However, the presence of poisoned or malicious data introduces significant challenges, including compromised model performance and system reliability. To tackle these issues, this paper proposes an efficient and resource-aware machine unlearning method tailored for federated incremental learning. The approach utilizes a membership inference attack mechanism to accurately identify poisoned data based on prediction confidence levels. Once detected, a targeted forgetting mechanism is applied, leveraging fine-tuning techniques to erase the influence of the poisoned data while preserving the model’s incremental learning capabilities. By aligning the distributions of poisoned data and third-party datasets, the method achieves reliable unlearning without introducing excessive computational overhead. Extensive experiments conducted on diverse datasets validate the method’s effectiveness, demonstrating a significant reduction in forgetting time (up to 21.05× speedup compared to baseline approaches) while maintaining robust model performance in incremental learning tasks. This work offers a scalable and efficient solution to the data forgetting problem, advancing the reliability and practicality of federated incremental learning in distributed and resource-constrained scenarios.

Tsinghua Science and Technology
Cite this article:
Chen K, Li W, Cao J, et al. Optimizing Federated Incremental Learning: Efficient Malicious Data Removal for Big Data Analytics. Tsinghua Science and Technology, 2025, https://doi.org/10.26599/TST.2025.901002

66

Views

11

Downloads

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Altmetrics

Received: 01 January 2025
Revised: 16 February 2025
Accepted: 04 March 2025
Available online: 03 July 2025

© The author(s) 2025

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).

Return