Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/112654
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Land Surveying and Geo-Informatics-
dc.creatorMahmoud, Men_US
dc.creatorZhao, Zen_US
dc.creatorChen, Wen_US
dc.creatorAdham, Men_US
dc.creatorLi, Yen_US
dc.date.accessioned2025-04-25T02:48:19Z-
dc.date.available2025-04-25T02:48:19Z-
dc.identifier.urihttp://hdl.handle.net/10397/112654-
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.rights© 2025 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Mahmoud, M., Zhao, Z., Chen, W., Adham, M., & Li, Y. (2025). Automated Scan-to-BIM: A Deep Learning-Based Framework for Indoor Environments with Complex Furniture Elements. Journal of Building Engineering, 112596, 106 is available at https://doi.org/10.1016/j.jobe.2025.112596.en_US
dc.subject3D modelsen_US
dc.subjectBuilding information modeling (BIM)en_US
dc.subjectDeep learningen_US
dc.subjectFurnitureen_US
dc.subjectPoint cloudsen_US
dc.titleAutomated Scan-to-BIM : a deep learning-based framework for indoor environments with complex furniture elementsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume106en_US
dc.identifier.doi10.1016/j.jobe.2025.112596en_US
dcterms.abstractExtensive 3D parametric datasets, such as Building Information Modeling (BIM) models, are crucial for reducing project costs, supporting planning, and enhancing operational efficiency in building management. However, conventional Scan-to-BIM methods rely heavily on manual or semi-automatic techniques, focusing on space-forming elements such as walls while often neglecting indoor space-occupying furniture. These methods struggle with incomplete point clouds, capturing shapes and orientations, and clustering inaccuracies. This paper presents an innovative and efficient deep learning-based framework to automatically reconstruct 3D models from point clouds. The framework accommodates diverse space-forming layouts and automatically generates parametric 3D BIM models for complex space-occupying elements like tables and chairs within the Revit platform. It also produces non-parametric 3D semantic representations of complete indoor scenes. Evaluation of publicly available and locally acquired datasets shows that the framework achieves over 98 % precision, recall, and F1-score, confirming its accuracy and effectiveness in generating complete 3D models. The reconstructed models preserve key real-world characteristics, including geometric fidelity, numerical attributes, spatial positioning, and various shapes and orientations of furniture. Seamless integration of deep learning and model-driven techniques overcomes the limitations of traditional Scan-to-BIM methods, providing an accurate and efficient solution for complex indoor space reconstruction.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationJournal of building engineering, 15 July 2025, v. 106, 112596en_US
dcterms.isPartOfJournal of building engineeringen_US
dcterms.issued2025-07-15-
dc.identifier.scopus2-s2.0-105001929424-
dc.identifier.eissn2352-7102en_US
dc.identifier.artn112596en_US
dc.description.validate202504 bchy-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_TA-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextResearch Institute of Land and Space at the Hong Kong Polytechnic Universityen_US
dc.description.pubStatusPublisheden_US
dc.description.TAElsevier (2025)en_US
dc.description.oaCategoryTAen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
1-s2.0-S2352710225008332-main.pdf11.69 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.