Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118131
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Industrial and Systems Engineeringen_US
dc.creatorZhang, Xen_US
dc.creatorWang, Ten_US
dc.creatorWang, XLen_US
dc.creatorFan, FLen_US
dc.creatorCheung, YMen_US
dc.creatorBose, Ien_US
dc.date.accessioned2026-03-18T04:03:03Z-
dc.date.available2026-03-18T04:03:03Z-
dc.identifier.issn2168-2216en_US
dc.identifier.urihttp://hdl.handle.net/10397/118131-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2026 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication X. Zhang, T. Wang, X. -L. Wang, F. -L. Fan, Y. -M. Cheung and I. Bose, 'Causality-Informed Neural Networks for Regularized Learning in Regression Problems,' in IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 56, no. 3, pp. 1895-1910, March 2026 is available at https://doi.org/10.1109/TSMC.2025.3646993.en_US
dc.subjectCausal inferenceen_US
dc.subjectCausality-informed neural network (CINN)en_US
dc.subjectDeep learningen_US
dc.subjectInformed learningen_US
dc.titleCausality-informed neural networks for regularized learning in regression problemsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage1895en_US
dc.identifier.epage1910en_US
dc.identifier.volume56en_US
dc.identifier.issue3en_US
dc.identifier.doi10.1109/TSMC.2025.3646993en_US
dcterms.abstractNeural networks that overlook the underlying causal relationships among observed variables pose significant risks in high-stakes decision-making contexts due to concerns about the robustness and stability of model performance. To tackle this issue, we present a general approach for embedding hierarchical causal structure among observed variables into a neural network to inform its learning. The proposed methodology, termed causality-informed neural network (CINN), exploits hierarchical causal structure learned from observational data as a structurally informed prior to guide the layer-to-layer architectural design of the neural network while maintaining the orientation of causal relationships in the discovered causal graph. The proposed method involves three steps. First, CINN mines causal relationships from observational data via directed acyclic graph (DAG) learning, where causal discovery is recast as a continuous optimization problem to circumvent the combinatorial nature of DAG learning. Second, we encode the discovered hierarchical causal graph among observed variables into a neural network via a dedicated architecture and loss function. By classifying observed variables in the DAG as root, intermediate, and leaf nodes, we translate the hierarchical causal DAG into CINN by creating a one-to-one correspondence between DAG nodes and certain CINN neurons. For the loss function, both intermediate and leaf nodes in the DAG are treated as target outputs during CINN training, facilitating the co-learning of causal relationships among the observed variables. Finally, as multiple loss components emerge in CINN, we leverage the projection of conflicting gradients (PCGrads) to mitigate the gradient interference among the multiple learning tasks. Computational studies indicate that CINN outperforms several state-of-the-art methods across a broad range of datasets. In addition, an ablation study that incrementally incorporates structural and quantitative causal knowledge into the neural network is conducted to highlight the pivotal role of causal knowledge in enhancing neural network’s prediction performance.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on systems, man, and cybernetics. Systems, Mar. 2026, v. 56, no. 3, p. 1895-1910en_US
dcterms.isPartOfIEEE transactions on systems, man, and cybernetics. Systemsen_US
dcterms.issued2026-03-
dc.identifier.scopus2-s2.0-105028030320-
dc.identifier.eissn2168-2232en_US
dc.description.validate202603 bcjzen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.SubFormIDG001288/2026-02-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported in part by the Research Grants Council of Hong Kong Special Administrative Region, China, under Project PolyU 25206422; in part by the National Natural Science Foundation of China under Grant 62406269; in part by the Research Committee of The Hong Kong Polytechnic University under Project RKB0 and Project G-UARJ; in part by the NSFC/Research Grants Council (RGC) Joint Research Scheme under Project N_HKBU214/21; in part by the Seed Funding for Collaborative Research Grants of Hong Kong Baptist University (HKBU) under Grant RC-SFCRG/23-24/R2/SCI/10; and in part by Guangdong and Hong Kong Universities “1 + 1 + 1” Cross-Campus Research Collaboration Scheme under Grant 2025A0505000004.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Zhang_Causality-informed_Neural_Networks.pdfPre-Published version8.38 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.