Residential College | false |
Status | 已發表Published |
Boosting Active Learning via Improving Test Performance | |
Wang, Tianyang1; Li, Xingjian2,6; Yang, Pengkun3; Hu, Guosheng4; Zeng, Xiangrui7; Huang, Siyu5; Xu, Cheng Zhong6![]() ![]() | |
2022-06-30 | |
Conference Name | 36th AAAI Conference on Artificial Intelligence / 34th Conference on Innovative Applications of Artificial Intelligence / 12th Symposium on Educational Advances in Artificial Intelligence |
Source Publication | Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022
![]() |
Volume | 36 |
Pages | 8566-8574 |
Conference Date | FEB 22-MAR 01, 2022 |
Conference Place | ELECTR NETWORK |
Abstract | Central to active learning (AL) is what data should be selected for annotation. Existing works attempt to select highly uncertain or informative data for annotation. Nevertheless, it remains unclear how selected data impacts the test performance of the task model used in AL. In this work, we explore such an impact by theoretically proving that selecting unlabeled data of higher gradient norm leads to a lower upper-bound of test loss, resulting in a better test performance. However, due to the lack of label information, directly computing gradient norm for unlabeled data is infeasible. To address this challenge, we propose two schemes, namely expected-gradnorm and entropy-gradnorm. The former computes the gradient norm by constructing an expected empirical loss while the latter constructs an unsupervised loss with entropy. Furthermore, we integrate the two schemes in a universal AL framework. We evaluate our method on classical image classification and semantic segmentation tasks. To demonstrate its competency in domain applications and its robustness to noise, we also validate our method on a cellular imaging analysis task, namely cryo-Electron Tomography subtomogram classification. Results demonstrate that our method achieves superior performance against the state of the art. We refer readers to https://arxiv.org/pdf/2112.05683.pdf for the full version of this paper which includes the appendix and source code link. |
URL | View the original |
Indexed By | CPCI-S |
Language | 英語English |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Artificial Intelligence |
WOS ID | WOS:000893639101065 |
Scopus ID | 2-s2.0-85140371727 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology |
Corresponding Author | Xu, Min |
Affiliation | 1.Austin Peay State University, United States 2.Baidu Inc, China 3.Tsinghua University, China 4.Oosto, Harvard University, United States 5.Harvard University, United States 6.University of Macau, Macao 7.Carnegie Mellon University, United States |
Recommended Citation GB/T 7714 | Wang, Tianyang,Li, Xingjian,Yang, Pengkun,et al. Boosting Active Learning via Improving Test Performance[C], 2022, 8566-8574. |
APA | Wang, Tianyang., Li, Xingjian., Yang, Pengkun., Hu, Guosheng., Zeng, Xiangrui., Huang, Siyu., Xu, Cheng Zhong., & Xu, Min (2022). Boosting Active Learning via Improving Test Performance. Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022, 36, 8566-8574. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment