UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
Progressive Poisoned Data Isolation for Training-Time Backdoor Defense
Chen, Yiming; Wu, Haiwei; Zhou, Jiantao
2024-03-25
Conference Name38th AAAI Conference on Artificial Intelligence, AAAI 2024
Source PublicationProceedings of the AAAI Conference on Artificial Intelligence
Volume38
Issue10
Pages11425-11433
Conference Date20 February 2024through 27 February 2024
Conference PlaceVancouver
CountryCanada
Abstract

Deep Neural Networks (DNN) are susceptible to backdoor attacks where malicious attackers manipulate the model’s predictions via data poisoning. It is hence imperative to develop a strategy for training a clean model using a potentially poisoned dataset. Previous training-time defense mechanisms typically employ an one-time isolation process, often leading to suboptimal isolation outcomes. In this study, we present a novel and efficacious defense method, termed Progressive Isolation of Poisoned Data (PIPD), that progressively isolates poisoned data to enhance the isolation accuracy and mitigate the risk of benign samples being misclassified as poisoned ones. Once the poisoned portion of the dataset has been identified, we introduce a selective training process to train a clean model. Through the implementation of these techniques, we ensure that the trained model manifests a significantly diminished attack success rate against the poisoned data. Extensive experiments on multiple benchmark datasets and DNN models, assessed against nine state-of-the-art backdoor attacks, demonstrate the superior performance of our PIPD method for backdoor defense. For instance, our PIPD achieves an average True Positive Rate (TPR) of 99.95% and an average False Positive Rate (FPR) of 0.06% for diverse attacks over CIFAR-10 dataset, markedly surpassing the performance of state-of-the-art methods. The code is available at https://github.com/RorschachChen/PIPD.git.

KeywordMl: Privacy
DOI10.1609/aaai.v38i10.29023
URLView the original
Language英語English
Scopus ID2-s2.0-85189745115
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorZhou, Jiantao
AffiliationState Key Laboratory of Internet of Things for Smart City Department of Computer and Information Science, University of Macau, Macao
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Chen, Yiming,Wu, Haiwei,Zhou, Jiantao. Progressive Poisoned Data Isolation for Training-Time Backdoor Defense[C], 2024, 11425-11433.
APA Chen, Yiming., Wu, Haiwei., & Zhou, Jiantao (2024). Progressive Poisoned Data Isolation for Training-Time Backdoor Defense. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 11425-11433.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Chen, Yiming]'s Articles
[Wu, Haiwei]'s Articles
[Zhou, Jiantao]'s Articles
Baidu academic
Similar articles in Baidu academic
[Chen, Yiming]'s Articles
[Wu, Haiwei]'s Articles
[Zhou, Jiantao]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Chen, Yiming]'s Articles
[Wu, Haiwei]'s Articles
[Zhou, Jiantao]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.