Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/117588
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Industrial and Systems Engineering-
dc.creatorQiu, J-
dc.creatorFang, Q-
dc.creatorKang, W-
dc.date.accessioned2026-02-26T03:47:12Z-
dc.date.available2026-02-26T03:47:12Z-
dc.identifier.urihttp://hdl.handle.net/10397/117588-
dc.language.isoenen_US
dc.publisherMDPI AGen_US
dc.rightsCopyright: © 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Qiu, J., Fang, Q., & Kang, W. (2025). Towards Controllable and Explainable Text Generation via Causal Intervention in LLMs. Electronics, 14(16), 3279 is available at https://doi.org/10.3390/electronics14163279.en_US
dc.subjectCounterfactual trainingen_US
dc.subjectHidden-state interventionen_US
dc.subjectMulti-attribute disentanglementen_US
dc.subjectResource-efficient generationen_US
dc.subjectStructural causal model (SCM)en_US
dc.titleTowards controllable and explainable text generation via causal intervention in LLMsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume14-
dc.identifier.issue16-
dc.identifier.doi10.3390/electronics14163279-
dcterms.abstractLarge Language Models (LLMs) excel in diverse text generation tasks but still face limited controllability, opaque decision processes, and frequent hallucinations. This paper presents a structural causal intervention framework that models input–hidden–output dependencies through a structural causal model and performs targeted interventions on hidden representations. By combining counterfactual sample construction with contrastive training, our method enables precise control of style, sentiment, and factual consistency while providing explicit causal explanations for output changes. Experiments on three representative tasks demonstrate consistent and substantial improvements: style transfer accuracy reaches 92.3% (+7–14 percentage points over strong baselines), sentiment-controlled generation achieves 90.1% accuracy (+1.3–10.9 points), and multi-attribute conflict rates drop to 3.7% (a 40–60% relative reduction). Our method also improves causal attribution scores to 0.83–0.85 and human agreement rates to 87–88%, while reducing training and inference latency by 25–30% through sparse masking that modifies ≤10% of hidden units per attribute. These results confirm that integrating structural causal intervention with counterfactual training advances controllability, interpretability, and efficiency in LLM-based generation, offering a robust foundation for deployment in reliability-critical and resource-constrained applications.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationElectronics (Switzerland), Aug. 2025, v. 14, no. 16, 3279-
dcterms.isPartOfElectronics (Switzerland)-
dcterms.issued2025-08-
dc.identifier.scopus2-s2.0-105014405949-
dc.identifier.eissn2079-9292-
dc.identifier.artn3279-
dc.description.validate202602 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
electronics-14-03279-v2.pdf1.47 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.