Residential Collegefalse
Status已發表Published
Impartial Adversarial Distillation: Addressing Biased Data-Free Knowledge Distillation via Adaptive Constrained Optimization
Liao, Donping1; Gao, Xitong2; Xu, Chengzhong1
2024-03-25
Conference Name38th AAAI Conference on Artificial Intelligence, AAAI 2024
Source PublicationProceedings of the AAAI Conference on Artificial Intelligence
Volume38
Issue4
Pages3342-3350
Conference Date20 February 2024through 27 February 2024
Conference PlaceVancouver
CountryCanada
Abstract

Data-Free Knowledge Distillation (DFKD) enables knowledge transfer from a pretrained teacher to a light-weighted student without original training data. Existing works are limited by a strong assumption that samples used to pretrain the teacher model are balanced, which is, however, unrealistic for many real-world tasks. In this work, we investigated a pragmatic yet under-explored problem: how to perform DFKD from a teacher model pretrained from imbalanced data. We observe a seemingly counter-intuitive phenomenon, i.e., adversarial DFKD algorithms favour minority classes, while causing a disastrous impact on majority classes. We theoretically prove that a biased teacher could cause severe disparity on different groups of synthetic data in adversarial distillation, which further exacerbates the mode collapse of a generator and consequently degenerates the overall accuracy of a distilled student model. To tackle this problem, we propose a class-adaptive regularization method, aiming to encourage impartial representation learning of a generator among different classes under a constrained learning formulation. We devise a primal-dual algorithm to solve the target optimization problem. Through extensive experiments, we show that our method mitigates the biased learning of majority classes in DFKD and improves the overall performance compared with baselines. Code will be available at https://github.com/ldpbuaa/ipad.

KeywordCv: Bias Fairness & Privacy Cv: Applications
DOI10.1609/aaai.v38i4.28120
URLView the original
Language英語English
Scopus ID2-s2.0-85189566648
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Faculty of Science and Technology
INSTITUTE OF COLLABORATIVE INNOVATION
Corresponding AuthorXu, Chengzhong
Affiliation1.State Key Lab of IoTSC, Department of Computer and Information Science, University of Macau, Macau, Macao
2.Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Liao, Donping,Gao, Xitong,Xu, Chengzhong. Impartial Adversarial Distillation: Addressing Biased Data-Free Knowledge Distillation via Adaptive Constrained Optimization[C], 2024, 3342-3350.
APA Liao, Donping., Gao, Xitong., & Xu, Chengzhong (2024). Impartial Adversarial Distillation: Addressing Biased Data-Free Knowledge Distillation via Adaptive Constrained Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 38(4), 3342-3350.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Liao, Donping]'s Articles
[Gao, Xitong]'s Articles
[Xu, Chengzhong]'s Articles
Baidu academic
Similar articles in Baidu academic
[Liao, Donping]'s Articles
[Gao, Xitong]'s Articles
[Xu, Chengzhong]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Liao, Donping]'s Articles
[Gao, Xitong]'s Articles
[Xu, Chengzhong]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.