TY - JOUR
T1 - You Can Have Better Graph Neural Networks by Not Training Weights at All
T2 - Finding Untrained GNNs Tickets
AU - Huang, Tianjin
AU - Chen, Tianlong
AU - Fang, Meng
AU - Menkovski, Vlado
AU - Zhao, Jiaxu
AU - Yin, Lu
AU - Pei, Yulong
AU - Mocanu, Decebal Constantin
AU - Wang, Zhangyang
AU - Pechenizkiy, Mykola
AU - Liu, Shiwei
N1 - Accepted by the LoG conference 2022 as a spotlight
PY - 2022/11/28
Y1 - 2022/11/28
N2 - Recent works have impressively demonstrated that there exists a subnetwork in randomly initialized convolutional neural networks (CNNs) that can match the performance of the fully trained dense networks at initialization, without any optimization of the weights of the network (i.e., untrained networks). However, the presence of such untrained subnetworks in graph neural networks (GNNs) still remains mysterious. In this paper we carry out the first-of-its-kind exploration of discovering matching untrained GNNs. With sparsity as the core tool, we can find \textit{untrained sparse subnetworks} at the initialization, that can match the performance of \textit{fully trained dense} GNNs. Besides this already encouraging finding of comparable performance, we show that the found untrained subnetworks can substantially mitigate the GNN over-smoothing problem, hence becoming a powerful tool to enable deeper GNNs without bells and whistles. We also observe that such sparse untrained subnetworks have appealing performance in out-of-distribution detection and robustness of input perturbations. We evaluate our method across widely-used GNN architectures on various popular datasets including the Open Graph Benchmark (OGB).
AB - Recent works have impressively demonstrated that there exists a subnetwork in randomly initialized convolutional neural networks (CNNs) that can match the performance of the fully trained dense networks at initialization, without any optimization of the weights of the network (i.e., untrained networks). However, the presence of such untrained subnetworks in graph neural networks (GNNs) still remains mysterious. In this paper we carry out the first-of-its-kind exploration of discovering matching untrained GNNs. With sparsity as the core tool, we can find \textit{untrained sparse subnetworks} at the initialization, that can match the performance of \textit{fully trained dense} GNNs. Besides this already encouraging finding of comparable performance, we show that the found untrained subnetworks can substantially mitigate the GNN over-smoothing problem, hence becoming a powerful tool to enable deeper GNNs without bells and whistles. We also observe that such sparse untrained subnetworks have appealing performance in out-of-distribution detection and robustness of input perturbations. We evaluate our method across widely-used GNN architectures on various popular datasets including the Open Graph Benchmark (OGB).
KW - cs.LG
U2 - 10.48550/arXiv.2211.15335
DO - 10.48550/arXiv.2211.15335
M3 - Article
SN - 2331-8422
VL - 2022
JO - arXiv
JF - arXiv
M1 - 2211.15335
ER -