Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/94803
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electronic and Information Engineering-
dc.creatorWang, LW-
dc.creatorSiu, WC-
dc.creatorLiu, ZS-
dc.creatorLi, CT-
dc.creatorLun, DPK-
dc.date.accessioned2022-08-30T07:30:58Z-
dc.date.available2022-08-30T07:30:58Z-
dc.identifier.isbn978-1-7281-8808-9 (Electronic)-
dc.identifier.isbn978-1-7281-8809-6 (Print on Demand(PoD))-
dc.identifier.urihttp://hdl.handle.net/10397/94803-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication L. -W. Wang, W. -C. Siu, Z. -S. Liu, C. -T. Li and D. P. -K. Lun, "Video Lightening with Dedicated CNN Architecture," 2020 25th International Conference on Pattern Recognition (ICPR), 2021, pp. 6447-6454 is available at https://dx.doi.org/10.1109/ICPR48806.2021.9413235.en_US
dc.subjectDeep learningen_US
dc.subjectLow-light video enhancementen_US
dc.titleVideo lightening with dedicated CNN architectureen_US
dc.typeConference Paperen_US
dc.identifier.spage6447-
dc.identifier.epage6454-
dc.identifier.doi10.1109/ICPR48806.2021.9413235-
dcterms.abstractDarkness brings us uncertainty, worry and low confidence. This is a problem not only applicable to us walking in a dark evening but also for drivers driving a car on the road with very dim or even without lighting condition. To address this problem, we propose a new CNN structure named as Video Lightening Network (VLN) that regards the low-light enhancement as a residual learning task, which is useful as reference to indirectly lightening the environment, or for vision-based application systems, such as driving assistant systems. The VLN consists of several Lightening Back-Projection (LBP) and Temporal Aggregation (TA) blocks. Each LBP block enhances the low-light frame by domain transfer learning that iteratively maps the frame between the low- and normal-light domains. A TA block handles the motion among neighboring frames by investigating the spatial and temporal relationships. Several TAs work in a multi-scale way, which compensates the motions at different levels. The proposed architecture has a consistent enhancement for different levels of illuminations, which significantly increases the visual quality even in the extremely dark environment. Extensive experimental results show that the proposed approach outperforms other methods under both objective and subjective metrics.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of ICPR 2020 : 25th International Conference on Pattern Recognition : Milan, 10-15 January 2021, p. 6447-6454-
dcterms.issued2021-
dc.identifier.scopus2-s2.0-85110456716-
dc.relation.conferenceInternational Conference on Pattern Recognition [ICPR]-
dc.identifier.artn9413235-
dc.description.validate202208 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera1422en_US
dc.identifier.SubFormID44926en_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThe Hong Kong Polytechnic University under consultancy project P19-0319en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Wang_Video_Lightening_Dedicated.pdfPre-Published version1.75 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

108
Last Week
0
Last month
Citations as of Apr 14, 2025

Downloads

96
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

5
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

4
Citations as of Oct 10, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.