Residential Collegefalse
Status即將出版Forthcoming
From Cluster Assumption to Graph Convolution: Graph-Based Semi-Supervised Learning Revisited
Wang, Zheng1; Ding, Hongming2; Pan, Li1; Li, Jianhua1; Gong, Zhiguo3; Yu, Philip S.4
2024-10-07
Source PublicationIEEE Transactions on Neural Networks and Learning Systems
ISSN2162-237X
Abstract

Graph-based semi-supervised learning (GSSL) has long been a research focus. Traditional methods are generally shallow learners, based on the cluster assumption. Recently, graph convolutional networks (GCNs) have become the predominant techniques for their promising performance. However, a critical question remains largely unanswered: why do deep GCNs encounter the oversmoothing problem, while traditional shallow GSSL methods do not, despite both progressing through the graph in a similar iterative manner? In this article, we theoretically discuss the relationship between these two types of methods in a unified optimization framework. One of the most intriguing findings is that, unlike traditional ones, typical GCNs may not effectively incorporate both graph structure and label information at each layer. Motivated by this, we propose three simple but powerful graph convolution methods. The first, optimized simple graph convolution (), is a supervised method, which guides the graph convolution process with labels. The others are two 'no-learning' unsupervised methods: graph structure preserving graph convolution () and its multiscale version GGCM, both aiming to preserve the graph structure information during the convolution process. Finally, we conduct extensive experiments to show the effectiveness of our methods.

KeywordData Mining Graph Convolutional Neural Networks Graph-based Semi-supervised Learning (Gssl)
DOI10.1109/TNNLS.2024.3454710
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering
WOS SubjectComputer Science, Artificial intelligenceComputer Science, Hardware & Architecture ; Computer Science, Theory & Methods ; Engineering, Electrical & Electronic
WOS IDWOS:001336025100001
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC445 HOES LANE, PISCATAWAY, NJ 08855-4141
Scopus ID2-s2.0-85207105454
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorPan, Li
Affiliation1.Institute of Cyber Science and Technology, Shanghai Jiao Tong University, Shanghai 200240, China
2.NIO Technology, Shanghai, 230051, China
3.State Key Laboratory of Internet of Things for Smart City and the Department of Computer and Information Science, University of Macau, Macao 999078, China
4.Department of Computer Science, University of Illinois Chicago, Chicago, IL 60607 USA
Recommended Citation
GB/T 7714
Wang, Zheng,Ding, Hongming,Pan, Li,et al. From Cluster Assumption to Graph Convolution: Graph-Based Semi-Supervised Learning Revisited[J]. IEEE Transactions on Neural Networks and Learning Systems, 2024.
APA Wang, Zheng., Ding, Hongming., Pan, Li., Li, Jianhua., Gong, Zhiguo., & Yu, Philip S. (2024). From Cluster Assumption to Graph Convolution: Graph-Based Semi-Supervised Learning Revisited. IEEE Transactions on Neural Networks and Learning Systems.
MLA Wang, Zheng,et al."From Cluster Assumption to Graph Convolution: Graph-Based Semi-Supervised Learning Revisited".IEEE Transactions on Neural Networks and Learning Systems (2024).
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Wang, Zheng]'s Articles
[Ding, Hongming]'s Articles
[Pan, Li]'s Articles
Baidu academic
Similar articles in Baidu academic
[Wang, Zheng]'s Articles
[Ding, Hongming]'s Articles
[Pan, Li]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Wang, Zheng]'s Articles
[Ding, Hongming]'s Articles
[Pan, Li]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.