Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/106820
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Applied Mathematicsen_US
dc.creatorShen, Gen_US
dc.date.accessioned2024-06-04T07:39:57Z-
dc.date.available2024-06-04T07:39:57Z-
dc.identifier.issn2640-3498en_US
dc.identifier.urihttp://hdl.handle.net/10397/106820-
dc.description41st International Conference on Machine Learning (ICML 2024), Vienna, Austria, July 21st - 27th 2024en_US
dc.language.isoenen_US
dc.publisherMLResearchPressen_US
dc.rightsCopyright 2024 by the author(s)en_US
dc.rightsPosted with permission of the author.en_US
dc.rightsThe following publication Shen, G. (2024). Exploring the Complexity of Deep Neural Networks through Functional Equivalence Proceedings of the 41st International Conference on Machine Learning, Proceedings of Machine Learning Research is available at https://proceedings.mlr.press/v235/shen24a.html.en_US
dc.titleExploring the Complexity of Deep Neural Networks through Functional Equivalenceen_US
dc.typeConference Paperen_US
dc.identifier.spage44643en_US
dc.identifier.epage44664en_US
dc.identifier.volume235en_US
dcterms.abstractWe investigate the complexity of deep neural networks through the lens of functional equivalence, which posits that different parameterizations can yield the same network function. Leveraging the equivalence property, we present a novel bound on the covering number for deep neural networks, which reveals that the complexity of neural networks can be reduced. Additionally, we demonstrate that functional equivalence benefits optimization, as overparameterized networks tend to be easier to train since increasing network width leads to a diminishing volume of the effective parameter space. These findings can offer valuable insights into the phenomenon of overparameterization and have implications for understanding generalization and optimization in deep learning.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of ICML 2024, v. 235, https://proceedings.mlr.press/v235/en_US
dcterms.issued2024-
dc.relation.ispartofbookPMLR Proceedings of machine learning research, Volume 235: International Conference on Machine Learning, 21-27 July 2024, Vienna, Austriaen_US
dc.relation.conferenceInternational Conference on Machine Learning [ICML]en_US
dc.description.validate202406 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera2752, a2964-
dc.identifier.SubFormID48240, 48943-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextGuohao Shen’s research is (partially) supported by the Hong Kong Research Grants Council (Grant No. 15305523) and research grants from The Hong Kong Polytechnic University.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCopyright retained by authoren_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
shen24a.pdf444.64 kBUnknownView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

40
Citations as of Apr 14, 2025

Downloads

32
Citations as of Apr 14, 2025

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.