Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/113081
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorSchool of Professional Education and Executive Development-
dc.creatorRong, PZ-
dc.date.accessioned2025-05-19T00:53:02Z-
dc.date.available2025-05-19T00:53:02Z-
dc.identifier.urihttp://hdl.handle.net/10397/113081-
dc.language.isoenen_US
dc.publisherMolecular Diversity Preservation International (MDPI)en_US
dc.rights© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Rong, P. DDNet: Depth Dominant Network for Semantic Segmentation of RGB-D Images. Sensors 2024, 24, 6914 is available at https://dx.doi.org/10.3390/s24216914.en_US
dc.subjectIndoor semantic segmentationen_US
dc.subjectConvolutional neural networken_US
dc.subjectRGB-D imagesen_US
dc.subjectInformation fusionen_US
dc.titleDDNet : depth dominant network for semantic segmentation of RGB-D imagesen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume24-
dc.identifier.issue21-
dc.identifier.doi10.3390/s24216914-
dcterms.abstractConvolutional neural networks (CNNs) have been widely applied to parse indoor scenes and segment objects represented by color images. Nonetheless, the lack of geometric and context information is a problem for most RGB-based methods, with which depth features are only used as an auxiliary module in RGB-D semantic segmentation. In this study, a novel depth dominant network (DDNet) is proposed to fully utilize the rich context information in the depth map. The critical insight is that obvious geometric information from the depth image is more conducive to segmentation than RGB data. Compared with other methods, DDNet is a depth-based network with two branches of CNNs to extract color and depth features. As the core of the encoder network, the depth branch is given a larger fusion weight to extract geometric information, while semantic information and complementary geometric information are provided by the color branch for the depth feature maps. The effectiveness of our proposed depth-based architecture has been demonstrated by comprehensive experimental evaluations and ablation studies on challenging RGB-D semantic segmentation benchmarks, including NYUv2 and a subset of ScanNetv2.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationSensors, Nov. 2025, v. 24, no. 21, 6914-
dcterms.isPartOfSensors-
dcterms.issued2024-11-
dc.identifier.isiWOS:001351025000001-
dc.identifier.eissn1424-8220-
dc.identifier.artn6914-
dc.description.validate202505 bcrc-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
sensors-24-06914.pdf2.18 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.