Back to results list
Please use this identifier to cite or link to this item:
|Title:||Localized generalization error model and its applications to supervised pattern classification problems||Authors:||Ng, Wing-yin||Keywords:||Hong Kong Polytechnic University -- Dissertations
Neural networks (Computer science)
|Issue Date:||2006||Publisher:||The Hong Kong Polytechnic University||Abstract:||The objective of this thesis is to investigate the localized generalization error of a classifier trained for supervised pattern classification problems. This is motivated by the straightforward idea that one should not expect a classifier to recognize correctly unseen samples which are totally different from the training samples. Therefore, a localized generalization error model (L-GEM) is proposed to give an upper bound on the generalization error for the unseen samples located within neighborhoods of the training samples. The L-GEM is applied to address three fundamental issues in supervised pattern classification problems: architecture selection for a neural network, feature selection and active learning. For architecture selection problem, one can use the L-GEM to select the largest neighborhoods around the training samples, subject to a predefined generalization error bound (Maximal Coverage Classification problem with Selected Generalization error bound (MC²SG)). A number of application problems in civil engineering, computer network security and image classification were solved by using Radial Basis Function Neural Network (RBFNN) trained by MC²SG L-GEM can also be used as a feature selection/reduction criterion. This is accomplished iteratively by measuring the feature which affects the generalization error the least. The problem of active learning is resolved by doing the opposite, i.e., each time selecting the training sample which yields the largest L-GEM value to the trained classifier. Since its derivation is based on the stochastic sensitivity measure of a classifier, the L-GEM is applicable to any classifier for which the stochastic sensitivity measure could be defined, e.g. RBFNN, multilayer perception neural networks and support vector machines. This thesis presents the L-GEM for a RBFNN, and a pilot study on the extension of the L-GEM to other classifiers including the multiple classifier system is discussed briefly.||Description:||195 p. : ill. ; 30 cm.
PolyU Library Call No.: [THS] LG51 .H577P COMP 2006 Ng
|URI:||http://hdl.handle.net/10397/3044||Rights:||All rights reserved.|
|Appears in Collections:||Thesis|
Show full item record
Files in This Item:
|b20696875_link.htm||For PolyU Users||162 B||HTML||View/Open|
|b20696875_ir.pdf||For All Users (Non-printable)||2.47 MB||Adobe PDF||View/Open|
Citations as of Feb 18, 2019
Citations as of Feb 18, 2019
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.