Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/108014
DC FieldValueLanguage
dc.contributorDepartment of Building Environment and Energy Engineering-
dc.creatorZheng, Hen_US
dc.creatorWang, Men_US
dc.creatorWang, Zen_US
dc.creatorHuang, Xen_US
dc.date.accessioned2024-07-23T01:36:19Z-
dc.date.available2024-07-23T01:36:19Z-
dc.identifier.issn0950-7051en_US
dc.identifier.urihttp://hdl.handle.net/10397/108014-
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.subjectChatGPT-Fireen_US
dc.subjectFire segmentationen_US
dc.subjectFireDMen_US
dc.subjectMulti-scaleen_US
dc.subjectStable Diffusion 2.1en_US
dc.subjectStable Diffusion XL 1.0en_US
dc.titleFireDM : a weakly-supervised approach for massive generation of multi-scale and multi-scene fire segmentation datasetsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume290en_US
dc.identifier.doi10.1016/j.knosys.2024.111547en_US
dcterms.abstractData availability and quality are crucial for the development of semantic segmentation techniques. However, creating high-quality fire scene datasets in a safe and efficient manner remains an unsolved challenge. To fill this gap, we introduce FireDM, the first method to generate unlimited fire segmentation datasets at virtually no cost. FireDM takes full advantage of the combined strengths of a combination of pre-trained diffusion models (Stable Diffusion XL 1.0 and Stable Diffusion 2.1) and text-guided diffusion using ChatGPT4-Fire to generate multi-scale and detail-rich fire images. The innovative fire-decoder module in FireDM then efficiently converts the cross-attention and multi-scale feature maps obtained during diffusion into accurate segmentation masks. This process requires only about 100 images and their corresponding segmentation masks for training. In our experiments, we trained the segmentation algorithms using the large-scale segmentation dataset generated by FireDM and all publicly available fire segmentation datasets respectively, and found that the segmentation algorithms trained with the former dataset outperformed the latter by at least 5% or more in terms of IoU, accuracy, F1-score and AP. This demonstrates the capability of FireDM in expanding a limited fire segmentation dataset. Additionally, the datasets generated by FireDM, with their multiple image resolutions, can adapt to the input sizes of different segmentation algorithms, significantly reducing information loss caused by resizing the image (e.g., cropping and scaling). Finally, we have created the world's first high-quality fire segmentation dataset benchmark using FireDM. The complete code and dataset of FireDM are publicly available at https://github.com/ZhengHongtao2001/FireDM.-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationKnowledge-based systems, 22 Apr. 2024, v. 290, 111547en_US
dcterms.isPartOfKnowledge-based systemsen_US
dcterms.issued2024-04-22-
dc.identifier.scopus2-s2.0-85186629492-
dc.identifier.artn111547en_US
dc.description.validate202407 bcwh-
dc.identifier.FolderNumbera3084b-
dc.identifier.SubFormID49430-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2026-04-22en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2026-04-22
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

79
Citations as of Nov 10, 2025

SCOPUSTM   
Citations

15
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

10
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.