Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/113081
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | School of Professional Education and Executive Development | - |
| dc.creator | Rong, PZ | - |
| dc.date.accessioned | 2025-05-19T00:53:02Z | - |
| dc.date.available | 2025-05-19T00:53:02Z | - |
| dc.identifier.uri | http://hdl.handle.net/10397/113081 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Molecular Diversity Preservation International (MDPI) | en_US |
| dc.rights | © 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). | en_US |
| dc.rights | The following publication Rong, P. DDNet: Depth Dominant Network for Semantic Segmentation of RGB-D Images. Sensors 2024, 24, 6914 is available at https://dx.doi.org/10.3390/s24216914. | en_US |
| dc.subject | Indoor semantic segmentation | en_US |
| dc.subject | Convolutional neural network | en_US |
| dc.subject | RGB-D images | en_US |
| dc.subject | Information fusion | en_US |
| dc.title | DDNet : depth dominant network for semantic segmentation of RGB-D images | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.volume | 24 | - |
| dc.identifier.issue | 21 | - |
| dc.identifier.doi | 10.3390/s24216914 | - |
| dcterms.abstract | Convolutional neural networks (CNNs) have been widely applied to parse indoor scenes and segment objects represented by color images. Nonetheless, the lack of geometric and context information is a problem for most RGB-based methods, with which depth features are only used as an auxiliary module in RGB-D semantic segmentation. In this study, a novel depth dominant network (DDNet) is proposed to fully utilize the rich context information in the depth map. The critical insight is that obvious geometric information from the depth image is more conducive to segmentation than RGB data. Compared with other methods, DDNet is a depth-based network with two branches of CNNs to extract color and depth features. As the core of the encoder network, the depth branch is given a larger fusion weight to extract geometric information, while semantic information and complementary geometric information are provided by the color branch for the depth feature maps. The effectiveness of our proposed depth-based architecture has been demonstrated by comprehensive experimental evaluations and ablation studies on challenging RGB-D semantic segmentation benchmarks, including NYUv2 and a subset of ScanNetv2. | - |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | Sensors, Nov. 2025, v. 24, no. 21, 6914 | - |
| dcterms.isPartOf | Sensors | - |
| dcterms.issued | 2024-11 | - |
| dc.identifier.isi | WOS:001351025000001 | - |
| dc.identifier.eissn | 1424-8220 | - |
| dc.identifier.artn | 6914 | - |
| dc.description.validate | 202505 bcrc | - |
| dc.description.oa | Version of Record | en_US |
| dc.identifier.FolderNumber | OA_Scopus/WOS | en_US |
| dc.description.fundingSource | RGC | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | CC | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| sensors-24-06914.pdf | 2.18 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



