Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/90145
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorZhou, Ken_US
dc.creatorHuang, Xen_US
dc.creatorLi, Yen_US
dc.creatorZha, Den_US
dc.creatorChen, Ren_US
dc.creatorHu, Xen_US
dc.date.accessioned2021-05-21T07:23:49Z-
dc.date.available2021-05-21T07:23:49Z-
dc.identifier.urihttp://hdl.handle.net/10397/90145-
dc.description34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, Canada, 2020en_US
dc.language.isoenen_US
dc.publisherNeurIPSen_US
dc.rightsPosted with permission of the publisher.en_US
dc.rightsThe following publication In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan & H. Lin (Eds.), Advances in Neural Information Processing Systems 33 (NeurIPS 2020), p. 1-12. NeurIPS, 2020 is published at NeurIPS 2020, https://proceedings.neurips.cc/paper/2020.en_US
dc.titleTowards deeper graph neural networks with differentiable group normalizationen_US
dc.typeConference Paperen_US
dc.identifier.spage1en_US
dc.identifier.epage12en_US
dcterms.abstractGraph neural networks (GNNs), which learn the representation of a node by aggregating its neighbors, have become an effective computational tool in downstream applications. Over-smoothing is one of the key issues which limit the performance of GNNs as the number of layers increases. It is because the stacked aggregators would make node representations converge to indistinguishable vectors. Several attempts have been made to tackle the issue by bringing linked node pairs close and unlinked pairs distinct. However, they often ignore the intrinsic community structures and would result in sub-optimal performance. The representations of nodes within the same community/class need be similar to facilitate the classification, while different classes are expected to be separated in embedding space. To bridge the gap, we introduce two over-smoothing metrics and a novel technique, i.e., differentiable group normalization (DGN). It normalizes nodes within the same group independently to increase their smoothness, and separates node distributions among different groups to significantly alleviate the over-smoothing issue. Experiments on real-world datasets demonstrate that DGN makes GNN models more robust to over-smoothing and achieves better performance with deeper GNNs.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan & H. Lin (Eds.), Advances in Neural Information Processing Systems 33 (NeurIPS 2020), p. 1-12. NeurIPS, 2020en_US
dcterms.issued2020-
dc.relation.ispartofbookAdvances in Neural Information Processing Systems 33 (NeurIPS 2020)en_US
dc.relation.conferenceConference on Neural Information Processing Systems [NeurIPS]en_US
dc.description.validate202105 bcrcen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera0609-n01-
dc.identifier.SubFormID580-
dc.description.pubStatusPublisheden_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
a0609-n01_580.pdf559.29 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

108
Last Week
0
Last month
Citations as of Apr 21, 2024

Downloads

34
Citations as of Apr 21, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.