Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118272
DC FieldValueLanguage
dc.contributorDepartment of Building Environment and Energy Engineeringen_US
dc.contributorDepartment of Land Surveying and Geo-Informaticsen_US
dc.contributorResearch Institute for Sustainable Urban Developmenten_US
dc.creatorLan, Hen_US
dc.creatorHou, HCen_US
dc.creatorWong, MSen_US
dc.date.accessioned2026-03-30T01:59:47Z-
dc.date.available2026-03-30T01:59:47Z-
dc.identifier.issn0360-1323en_US
dc.identifier.urihttp://hdl.handle.net/10397/118272-
dc.language.isoenen_US
dc.publisherPergamon Pressen_US
dc.subjectIndoor thermal environment managementen_US
dc.subjectMachine learningen_US
dc.subjectMultimodal modelen_US
dc.subjectOccupant-centric controlen_US
dc.subjectSelf-attention mechanismen_US
dc.subjectThermal comfort predictionen_US
dc.titleIntegrating infrared facial thermal imaging and tabular data for multimodal prediction of occupants' thermal sensationen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume275en_US
dc.identifier.doi10.1016/j.buildenv.2025.112814en_US
dcterms.abstractDeveloping robust thermal comfort models is essential for occupant-centric control (OCC) to optimize the indoor thermal environment while minimizing energy consumption. Conventional single-modal machine learning models, relying solely on either tabular or image data, often suffer from limited prediction accuracy and versatility. To address these challenges, this study proposes a multimodal framework that integrates both data types. A dataset of 610 paired records, encompassing environmental data, individual attributes, thermal sensation votes (TSV), and occupants’ facial thermal images, was collected. Separate single-modal models were trained on tabular and image data to identify the best-performing model for each modality. These were subsequently integrated using a self-attention mechanism to develop a unified multimodal predictive model. Results demonstrate that the artificial neural network (ANN), utilizing only tabular data, achieved an accuracy of 69.67% without incorporating temperature variables from facial regions of interest (ROIs), increasing to 72.46% when these variables were included. Conversely, the Inception-V3 model, trained solely on facial thermal images, achieved 63.44% accuracy. By integrating these approaches, the ANN+Inception-V3 multimodal model achieved a significantly improved accuracy of 81.48%, effectively capturing interaction effects from both data types. This study presents a robust framework and methodological reference for advancing multimodal thermal comfort prediction models, enabling scalable, personalized, and energy-efficient management strategies for indoor environments.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationBuilding and environment, 1 May 2025, v. 275, 112814en_US
dcterms.isPartOfBuilding and environmenten_US
dcterms.issued2025-05-01-
dc.identifier.scopus2-s2.0-86000724700-
dc.identifier.eissn1873-684Xen_US
dc.identifier.artn112814en_US
dc.description.validate202603 bchyen_US
dc.description.oaNot applicableen_US
dc.identifier.SubFormIDG001359/2025-12-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextFunding text 1: This project received ethical approval from The Hong Kong Polytechnic University under the reference number HSEARS20230906001. We sincerely acknowledge the support provided by The Hong Kong Polytechnic University in facilitating in this project. We are also deeply grateful to the participants for their time, effort, and valuable contributions, which made this study possible. Cynthia Hou thanks the funding support from the Hong Kong Polytechnic University under project ID P0052446. M.S. Wong thanks the funding support from the General Research Fund (grant no. 15603920 and 15609421), and the Collaborative Research Fund (grant no. C5062-21GF) from the Research Grants Council, Hong Kong, China; and the funding support from the Research Institute for Sustainable Urban Development, The Hong Kong Polytechnic University, Hong Kong, China (grant no. 1-BBG2).; Funding text 2: Cynthia Hou thanks the funding support from the Hong Kong Polytechnic University under Project ID P0052446. M.S. Wong thanks the funding support from the General Research Fund (Grant No. 15603920 and 15609421), and the Collaborative Research Fund (Grant No. C5062-21GF) from the Research Grants Council, Hong Kong, China; and the funding support from the Research Institute for Sustainable Urban Development, The Hong Kong Polytechnic University, Hong Kong, China (Grant No. 1-BBG2).en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2027-05-01en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2027-05-01
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.