Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/104374
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Industrial and Systems Engineering-
dc.creatorZhang, Yen_US
dc.creatorHuang, Yen_US
dc.creatorZhao, Den_US
dc.creatorWu, CHen_US
dc.creatorIp, WHen_US
dc.creatorYung, KLen_US
dc.date.accessioned2024-02-05T08:49:14Z-
dc.date.available2024-02-05T08:49:14Z-
dc.identifier.issn1380-7501en_US
dc.identifier.urihttp://hdl.handle.net/10397/104374-
dc.language.isoenen_US
dc.publisherSpringer New York LLCen_US
dc.rights© Springer Science+Business Media, LLC, part of Springer Nature 2021en_US
dc.rightsThis version of the article has been accepted for publication, after peer review (when applicable) and is subject to Springer Nature’s AM terms of use(https://www.springernature.com/gp/open-research/policies/accepted-manuscript-terms), but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: http://dx.doi.org/10.1007/s11042-021-11101-w.en_US
dc.subjectConvolutional neural networken_US
dc.subjectDeep feature mergingen_US
dc.subjectDenseNet121en_US
dc.subjectScene text detectoren_US
dc.titleA scene text detector based on deep feature mergingen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage29005en_US
dc.identifier.epage29016en_US
dc.identifier.volume80en_US
dc.identifier.issue19en_US
dc.identifier.doi10.1007/s11042-021-11101-wen_US
dcterms.abstractScene text detection has become an important research topic. It can be broadly applied to much industrial equipment, such as smart phones, intelligent scanners, and IoT devices. Many existing scene text detection methods have achieved advanced performance. However, text in scene images is presented with differing orientations and varying shapes, rendering scene text detection a challenging task. This paper proposes a method for detecting texts in scene images. First, four stages of low-level features is extracted using DenseNet121. Low-level features are then merged by transposed convolution and skip connection. Second, the merged feature map is used to generate a score map, box map, and angle map. Finally, the Locality-Aware Non-Maximum Suppression (LANMS) is applied as post-processing to generate the final bounding box. The proposed method achieves an F-measure of 0.826 on ICDAR 2015 and 0.761 on MSRA-TD500, respectively.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationMultimedia tools and applications, Aug. 2021, v. 80, no. 19, p. 29005-29016en_US
dcterms.isPartOfMultimedia tools and applicationsen_US
dcterms.issued2021-08-
dc.identifier.scopus2-s2.0-85108016779-
dc.identifier.eissn1573-7721en_US
dc.description.validate202402 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberISE-0125-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextScience and Technology Plan Projects of Shenzhen; Graduate Education Reform Project of Shenzhen University; The Hong Kong Polytechnic Universityen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS56389511-
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Ip_Scene_Text_Detector.pdfPre-Published version1.35 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

111
Last Week
2
Last month
Citations as of Nov 30, 2025

Downloads

94
Citations as of Nov 30, 2025

SCOPUSTM   
Citations

7
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

2
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.