Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/115758
| Title: | Unlocking high-fidelity learning : towards neuron-grained model extraction | Authors: | Xiao, Y Hu, H Ye, Q Tang, L Liang, Z Zheng, H |
Issue Date: | Nov-2025 | Source: | IEEE transactions on dependable and secure computing, Nov.-Dec. 2025, v. 22, no. 6, p. 6622-6635 | Abstract: | Model extraction (ME) attacks replicate valuable black-box machine learning (ML) models via malicious query interactions. Cutting-edge attacks focus on actively designing query samples to enhance model fidelity and imprudently adhere to the standard ML training approach. This causes a deviation from the true objective of learning a model over a task. In this paper, we innovatively shift our focus from query selection to training process optimization, aiming to boost the similarity of the copy model with the victim model from neuron to model level. We leverage neuron matching theory to attain this objective and develop a general training booster framework, MEBooster, to fully exploit this theory. MEBooster comprises an initial bootstrapping phase that furnishes initial parameters and an optimal model architecture, followed by a post-processing phase that employs fine-tuning for enhanced neuron matching. Notably, MEBooster can seamlessly integrate with all existing model extraction attacks, enhancing their overall performance. Performance evaluation shows up to 58.10% fidelity gain in image classification. From a defender's perspective, we introduce a novel defensive strategy called Stochastic Norm Enlargement (SNE) to mitigate the risk of such attacks by enlarging the model parameters' norm property in training. Performance evaluation shows up to 58.81% extractability (i.e., fidelity) reduction. | Keywords: | Defense against model extraction Machine learning privacy Model extraction attack |
Publisher: | Institute of Electrical and Electronics Engineers | Journal: | IEEE transactions on dependable and secure computing | ISSN: | 1545-5971 | EISSN: | 1941-0018 | DOI: | 10.1109/TDSC.2025.3588857 | Rights: | © 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The following publication Y. Xiao, H. Hu, Q. Ye, L. Tang, Z. Liang and H. Zheng, "Unlocking High-Fidelity Learning: Towards Neuron-Grained Model Extraction," in IEEE Transactions on Dependable and Secure Computing, vol. 22, no. 6, pp. 6622-6635, Nov.-Dec. 2025 is available at https://doi.org/10.1109/TDSC.2025.3588857. |
| Appears in Collections: | Journal/Magazine Article |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Xiao_Unlocking_High_Fidelity.pdf | Pre-Published version | 4.1 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



