Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/116816
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Computing | - |
| dc.creator | Wang, Y | - |
| dc.creator | Guo, J | - |
| dc.creator | Guo, S | - |
| dc.creator | Liu, Y | - |
| dc.creator | Zhang, J | - |
| dc.creator | Zhang, W | - |
| dc.date.accessioned | 2026-01-21T03:52:53Z | - |
| dc.date.available | 2026-01-21T03:52:53Z | - |
| dc.identifier.isbn | 979-8-4007-0686-8 | - |
| dc.identifier.uri | http://hdl.handle.net/10397/116816 | - |
| dc.description | 32nd ACM International Conference on Multimedia, Melbourne VIC, Australia, 28 October 2024 - 1 November 2024 | en_US |
| dc.language.iso | en | en_US |
| dc.publisher | The Association for Computing Machinery | en_US |
| dc.rights | This work is licensed under a Creative Commons Attribution International 4.0 License (https://creativecommons.org/licenses/by/4.0/). | en_US |
| dc.rights | ©2024 Copyright held by the owner/author(s). | en_US |
| dc.rights | The following publication Wang, Y., Guo, J., Guo, S., Liu, Y., Zhang, J., & Zhang, W. (2024). SFP: Spurious Feature-Targeted Pruning for Out-of-Distribution Generalization Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne VIC, Australia is available at https://doi.org/10.1145/3664647.3680969. | en_US |
| dc.subject | Deep neural network | en_US |
| dc.subject | Model pruning | en_US |
| dc.subject | Module detection | en_US |
| dc.subject | Out-of-distribution generalization | en_US |
| dc.title | SFP : spurious feature-targeted pruning for out-of-distribution generalization | en_US |
| dc.type | Conference Paper | en_US |
| dc.identifier.spage | 8412 | - |
| dc.identifier.epage | 8420 | - |
| dc.identifier.doi | 10.1145/3664647.3680969 | - |
| dcterms.abstract | Recent studies reveal that even highly biased dense networks can contain an invariant substructure with superior out-of-distribution (OOD) generalization. While existing works commonly seek these substructures using global sparsity constraints, the uniform imposition of sparse penalties across samples with diverse levels of spurious contents renders such methods suboptimal. The precise adaptation of model sparsity, specifically tailored for spurious features, remains a significant challenge. Motivated by the insight that in-distribution (ID) data containing spurious features may exhibit lower experiential risk, we propose a novel Spurious Feature-targeted Pruning framework, dubbed SFP, to induce the authentic invariant substructures without referring to the above concerns. Specifically, SFP distinguishes spurious features within ID instances during training by a theoretically validated threshold. It then penalizes the corresponding feature projections onto the model space, steering the optimization towards subspaces spanned by those invariant factors. Moreover, we also conduct detailed theoretical analysis to provide a rationality guarantee and a proof framework for OOD structures based on model sparsity. Experiments on various OOD datasets show that SFP can significantly outperform both structure-based and non-structure-based OOD generalization state-of-the-art (SOTA) methods by large margins. | - |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | In MM ’24: Proceedings of the 32nd ACM International Conference on Multimedia, p. 8412-8420. New York, NY: The Association for Computing Machinery, 2024 | - |
| dcterms.issued | 2024 | - |
| dc.identifier.scopus | 2-s2.0-85209796463 | - |
| dc.relation.ispartofbook | MM ’24: Proceedings of the 32nd ACM International Conference on Multimedia | - |
| dc.relation.conference | ACM International Conference on Multimedia [MM] | - |
| dc.publisher.place | New York, NY | en_US |
| dc.description.validate | 202601 bcch | - |
| dc.description.oa | Version of Record | en_US |
| dc.identifier.FolderNumber | OA_Scopus/WOS | en_US |
| dc.description.fundingSource | RGC | en_US |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | This research was supported by funding from the Hong Kong RGC General Research Fund (No. 152211/23E, 152164/24E, 152203/20E, 152244/21E, 152169/22E, 152228/23E), Research Impact Fund (No. R5060-19, No. R5034-18), Areas of Excellence Scheme (AoE/E-601/22-R), the National Natural Science Foundation of China (No. 62102327, 62192781, 62172326, and 62137002), the PolyU Internal Fund (No. P0043932), the Key-Area Research and Development Program of Guangdong Province (No. 2021B0101400003), and the Project of China Knowledge Centre for Engineering Science and Technology.This research was also supported by NVIDIA AI Technology Center. | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | CC | en_US |
| Appears in Collections: | Conference Paper | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| 3664647.3680969.pdf | 2.15 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



