Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/108002
DC FieldValueLanguage
dc.contributorDepartment of Building Environment and Energy Engineering-
dc.creatorZheng, H-
dc.creatorDing, Y-
dc.creatorWang, Z-
dc.creatorHuang, X-
dc.date.accessioned2024-07-23T01:36:13Z-
dc.date.available2024-07-23T01:36:13Z-
dc.identifier.urihttp://hdl.handle.net/10397/108002-
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.subjectContrastive lossen_US
dc.subjectLatent diffusion processen_US
dc.subjectMultimodel fusionen_US
dc.subjectOpen-vocabularyen_US
dc.subjectUniversalen_US
dc.titleSegLD : achieving universal, zero-shot and open-vocabulary segmentation through multimodal fusion via latent diffusion processesen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume111-
dc.identifier.doi10.1016/j.inffus.2024.102509-
dcterms.abstractOpen-vocabulary learning can identify categories marked during training (seen categories) and generalize to categories not annotated in the training set (unseen categories). It could theoretically extend segmentation systems to more universal applications. However, current open-vocabulary segmentation frameworks are primarily suited for specific tasks or require retraining according to the task, and they significantly underperform in inferring seen categories compared to fully supervised frameworks. Therefore, we introduce a universal open-vocabulary segmentation framework based on the latent diffusion process (SegLD), which requires only a single training session on a panoptic dataset to achieve inference across all open-vocabulary segmentation tasks, and reaches SOTA segmentation performance for both seen and unseen categories in every task. Specifically, SegLD comprises two stages: in the first stage, we deploy two parallel latent diffusion processes to deeply fuse the text (image caption or category labels) and image information, further aggregating the multi-scale features output from both latent diffusion processes on a scale basis. In the second stage, we introduce text queries, text list queries, and task queries, facilitating the learning of inter-category and inter-task differences through the computation of contrastive losses between them. Text queries are then further fed into a Transformer Decoder to obtain category-agnostic segmentation masks. Then we establish classification loss functions for the type of text input during training, whether image captions or category labels, to help assign a category label from the open vocabulary to each predicted binary mask. Experimental results show that, with just a single training session, SegLD significantly outperforms other contemporary SOTA fully supervised segmentation frameworks and open-vocabulary segmentation frameworks across almost all evaluation metrics for both known and unknown categories on the ADE20K, Cityscapes, and COCO datasets. This highlights SegLD's capability as a universal segmentation framework, with the potential to replace other segmentation frameworks and adapt to various segmentation domains. The project link for SegLD is https://zht-segld.github.io/.-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationInformation fusion, Nov. 2024, v. 111, 102509-
dcterms.isPartOfInformation fusion-
dcterms.issued2024-11-
dc.identifier.scopus2-s2.0-85196418989-
dc.identifier.eissn1566-2535-
dc.identifier.artn102509-
dc.description.validate202407 bcwh-
dc.identifier.FolderNumbera3082ben_US
dc.identifier.SubFormID49416en_US
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2026-11-30en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2026-11-30
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

80
Citations as of Nov 10, 2025

SCOPUSTM   
Citations

5
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

4
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.