Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118131
PIRA download icon_1.1View/Download Full Text
Title: Causality-informed neural networks for regularized learning in regression problems
Authors: Zhang, X 
Wang, T 
Wang, XL
Fan, FL
Cheung, YM
Bose, I
Issue Date: Mar-2026
Source: IEEE transactions on systems, man, and cybernetics. Systems, Mar. 2026, v. 56, no. 3, p. 1895-1910
Abstract: Neural networks that overlook the underlying causal relationships among observed variables pose significant risks in high-stakes decision-making contexts due to concerns about the robustness and stability of model performance. To tackle this issue, we present a general approach for embedding hierarchical causal structure among observed variables into a neural network to inform its learning. The proposed methodology, termed causality-informed neural network (CINN), exploits hierarchical causal structure learned from observational data as a structurally informed prior to guide the layer-to-layer architectural design of the neural network while maintaining the orientation of causal relationships in the discovered causal graph. The proposed method involves three steps. First, CINN mines causal relationships from observational data via directed acyclic graph (DAG) learning, where causal discovery is recast as a continuous optimization problem to circumvent the combinatorial nature of DAG learning. Second, we encode the discovered hierarchical causal graph among observed variables into a neural network via a dedicated architecture and loss function. By classifying observed variables in the DAG as root, intermediate, and leaf nodes, we translate the hierarchical causal DAG into CINN by creating a one-to-one correspondence between DAG nodes and certain CINN neurons. For the loss function, both intermediate and leaf nodes in the DAG are treated as target outputs during CINN training, facilitating the co-learning of causal relationships among the observed variables. Finally, as multiple loss components emerge in CINN, we leverage the projection of conflicting gradients (PCGrads) to mitigate the gradient interference among the multiple learning tasks. Computational studies indicate that CINN outperforms several state-of-the-art methods across a broad range of datasets. In addition, an ablation study that incrementally incorporates structural and quantitative causal knowledge into the neural network is conducted to highlight the pivotal role of causal knowledge in enhancing neural network’s prediction performance.
Keywords: Causal inference
Causality-informed neural network (CINN)
Deep learning
Informed learning
Publisher: Institute of Electrical and Electronics Engineers
Journal: IEEE transactions on systems, man, and cybernetics. Systems 
ISSN: 2168-2216
EISSN: 2168-2232
DOI: 10.1109/TSMC.2025.3646993
Rights: © 2026 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
The following publication X. Zhang, T. Wang, X. -L. Wang, F. -L. Fan, Y. -M. Cheung and I. Bose, 'Causality-Informed Neural Networks for Regularized Learning in Regression Problems,' in IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 56, no. 3, pp. 1895-1910, March 2026 is available at https://doi.org/10.1109/TSMC.2025.3646993.
Appears in Collections:Journal/Magazine Article

Files in This Item:
File Description SizeFormat 
Zhang_Causality-informed_Neural_Networks.pdfPre-Published version8.38 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.