Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105694
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorWang, Ken_US
dc.creatorLin, Len_US
dc.creatorZuo, Wen_US
dc.creatorGu, Sen_US
dc.creatorZhang, Len_US
dc.date.accessioned2024-04-15T07:35:56Z-
dc.date.available2024-04-15T07:35:56Z-
dc.identifier.isbn978-1-4673-8851-1 (Electronic)en_US
dc.identifier.isbn978-1-4673-8852-8 (Print on Demand(PoD))en_US
dc.identifier.urihttp://hdl.handle.net/10397/105694-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2016 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication K. Wang, L. Lin, W. Zuo, S. Gu and L. Zhang, "Dictionary Pair Classifier Driven Convolutional Neural Networks for Object Detection," 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 2016, pp. 2138-2146 is available at https://doi.org/10.1109/CVPR.2016.235.en_US
dc.titleDictionary pair classifier driven convolutional neural networks for object detectionen_US
dc.typeConference Paperen_US
dc.identifier.spage2138en_US
dc.identifier.epage2146en_US
dc.identifier.doi10.1109/CVPR.2016.235en_US
dcterms.abstractFeature representation and object category classification are two key components of most object detection methods. While significant improvements have been achieved for deep feature representation learning, traditional SVM/softmax classifiers remain the dominant methods for the final object category classification. However, SVM/softmax classifiers lack the capacity of explicitly exploiting the complex structure of deep features, as they are purely discriminative methods. The recently proposed discriminative dictionary pair learning (DPL) model involves a fidelity term to minimize the reconstruction loss and a discrimination term to enhance the discriminative capability of the learned dictionary pair, and thus is appropriate for balancing the representation and discrimination to boost object detection performance. In this paper, we propose a novel object detection system by unifying DPL with the convolutional feature learning. Specifically, we incorporate DPL as a Dictionary Pair Classifier Layer (DPCL) into the deep architecture, and develop an end-to-end learning algorithm for optimizing the dictionary pairs and the neural networks simultaneously. Moreover, we design a multi-task loss for guiding our model to accomplish the three correlated tasks: objectness estimation, categoryness computation, and bounding box regression. From the extensive experiments on PASCAL VOC 2007/2012 benchmarks, our approach demonstrates the effectiveness to substantially improve the performances over the popular existing object detection frameworks (e.g., R-CNN [13] and FRCN [12]), and achieves new state-of-the-arts.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitation2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 26 June - 1 July 2016, Las Vegas, Nevada, p. 2138-2146en_US
dcterms.issued2016-
dc.identifier.scopus2-s2.0-84986327418-
dc.relation.conferenceIEEE Conference on Computer Vision and Pattern Recognition [CVPR]-
dc.description.validate202402 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberCOMP-1383-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHong Kong Polyutechnic University’s Joint Supervision Scheme with the Chinese Mainland, Taiwan and Macao Universities; Guangdong Natural Science Foundation; Special Program for Applied Research on Super Computation of the NSFC-Guangdong Joint Fund; Fundamental Research Funds for the Central Universitiesen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS13932353-
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Gu_Dictionary_Pair_Classifier.pdfPre-Published version1.76 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

7
Citations as of May 12, 2024

Downloads

1
Citations as of May 12, 2024

SCOPUSTM   
Citations

44
Citations as of May 17, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.