Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/106820
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Applied Mathematics | en_US |
| dc.creator | Shen, G | en_US |
| dc.date.accessioned | 2024-06-04T07:39:57Z | - |
| dc.date.available | 2024-06-04T07:39:57Z | - |
| dc.identifier.issn | 2640-3498 | en_US |
| dc.identifier.uri | http://hdl.handle.net/10397/106820 | - |
| dc.description | 41st International Conference on Machine Learning (ICML 2024), Vienna, Austria, July 21st - 27th 2024 | en_US |
| dc.language.iso | en | en_US |
| dc.publisher | MLResearchPress | en_US |
| dc.rights | Copyright 2024 by the author(s) | en_US |
| dc.rights | Posted with permission of the author. | en_US |
| dc.rights | The following publication Shen, G. (2024). Exploring the Complexity of Deep Neural Networks through Functional Equivalence Proceedings of the 41st International Conference on Machine Learning, Proceedings of Machine Learning Research is available at https://proceedings.mlr.press/v235/shen24a.html. | en_US |
| dc.title | Exploring the Complexity of Deep Neural Networks through Functional Equivalence | en_US |
| dc.type | Conference Paper | en_US |
| dc.identifier.spage | 44643 | en_US |
| dc.identifier.epage | 44664 | en_US |
| dc.identifier.volume | 235 | en_US |
| dcterms.abstract | We investigate the complexity of deep neural networks through the lens of functional equivalence, which posits that different parameterizations can yield the same network function. Leveraging the equivalence property, we present a novel bound on the covering number for deep neural networks, which reveals that the complexity of neural networks can be reduced. Additionally, we demonstrate that functional equivalence benefits optimization, as overparameterized networks tend to be easier to train since increasing network width leads to a diminishing volume of the effective parameter space. These findings can offer valuable insights into the phenomenon of overparameterization and have implications for understanding generalization and optimization in deep learning. | en_US |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | Proceedings of ICML 2024, v. 235, https://proceedings.mlr.press/v235/ | en_US |
| dcterms.issued | 2024 | - |
| dc.relation.ispartofbook | PMLR Proceedings of machine learning research, Volume 235: International Conference on Machine Learning, 21-27 July 2024, Vienna, Austria | en_US |
| dc.relation.conference | International Conference on Machine Learning [ICML] | en_US |
| dc.description.validate | 202406 bcch | en_US |
| dc.description.oa | Version of Record | en_US |
| dc.identifier.FolderNumber | a2752, a2964 | - |
| dc.identifier.SubFormID | 48240, 48943 | - |
| dc.description.fundingSource | RGC | en_US |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | Guohao Shen’s research is (partially) supported by the Hong Kong Research Grants Council (Grant No. 15305523) and research grants from The Hong Kong Polytechnic University. | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Copyright retained by author | en_US |
| Appears in Collections: | Conference Paper | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| shen24a.pdf | 444.64 kB | Unknown | View/Open |
Page views
40
Citations as of Apr 14, 2025
Downloads
32
Citations as of Apr 14, 2025
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



