Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/189
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electronic and Information Engineering-
dc.creatorCho, SY-
dc.creatorChi, ZG-
dc.creatorSiu, WC-
dc.creatorTsoi, AC-
dc.date.accessioned2014-12-11T08:23:09Z-
dc.date.available2014-12-11T08:23:09Z-
dc.identifier.issn1045-9227-
dc.identifier.urihttp://hdl.handle.net/10397/189-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2003 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.en_US
dc.rightsThis material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.en_US
dc.subjectAdaptive processing of data structuresen_US
dc.subjectBack-propagation through structure (BPTS)en_US
dc.subjectLeast-squares methoden_US
dc.subjectLong-term dependencyen_US
dc.titleAn improved algorithm for learning long-term dependency problems in adaptive processing of data structuresen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage781-
dc.identifier.epage793-
dc.identifier.volume14-
dc.identifier.issue4-
dc.identifier.doi10.1109/TNN.2003.813831-
dcterms.abstractFor the past decade, many researchers have explored the use of neural-network representations for the adaptive processing of data structures. One of the most popular learning formulations of data structure processing is backpropagation through structure (BPTS). The BPTS algorithm has been successful applied to a number of learning tasks that involve structural patterns such as logo and natural scene classification. The main limitations of the BPTS algorithm are attributed to slow convergence speed and the long-term dependency problem for the adaptive processing of data structures. In this paper, an improved algorithm is proposed to solve these problems. The idea of this algorithm is to optimize the free learning parameters of the neural network in the node representation by using least-squares-based optimization methods in a layer-by-layer fashion. Not only can fast convergence speed be achieved, but the long-term dependency problem can also be overcome since the vanishing of gradient information is avoided when our approach is applied to very deep tree structures.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on neural networks, July 2003, v. 14, no. 4, p. 781-793-
dcterms.isPartOfIEEE transactions on neural networks-
dcterms.issued2003-07-
dc.identifier.isiWOS:000184371900005-
dc.identifier.scopus2-s2.0-0042525850-
dc.identifier.rosgroupidr17674-
dc.description.ros2003-2004 > Academic research: refereed > Publication in refereed journal-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_IR/PIRAen_US
dc.description.pubStatusPublisheden_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
adaptive-processing_03.pdf891.03 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

86
Last Week
0
Last month
Citations as of Mar 24, 2024

Downloads

109
Citations as of Mar 24, 2024

SCOPUSTM   
Citations

37
Last Week
0
Last month
1
Citations as of Mar 28, 2024

WEB OF SCIENCETM
Citations

29
Last Week
0
Last month
0
Citations as of Mar 28, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.