Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/110261
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Electrical and Electronic Engineering | - |
| dc.creator | Yeung, Ching Chi | - |
| dc.identifier.uri | https://theses.lib.polyu.edu.hk/handle/200/13272 | - |
| dc.language.iso | English | - |
| dc.title | Deep learning for vision-based defect inspection | - |
| dc.type | Thesis | - |
| dcterms.abstract | Vision-based defect inspection is an essential quality control task in various industries. With the development of deep learning, deep learning-based visual defect inspection methods have achieved remarkable performance. However, existing deep learning-based visual defect inspection models face three main challenges according to their specific application requirements, including inspection efficiency, precise localization and classification, and generalization ability. Therefore, this thesis aims to investigate deep learning-based models to address these challenges, which are particularly relevant to three specific applications of vision-based defect inspection. These applications include steel surface defect detection, defect semantic segmentation, and pavement crack detection. | - |
| dcterms.abstract | In this thesis, we study the efficient design of steel surface defect detection models. We propose a fused-attention network (FANet) to balance the trade-off between accuracy and speed. This model applies an attention mechanism to a single balanced feature map to improve accuracy while maintaining detection speed. Moreover, it introduces a feature fusion and an attention module to handle defects with multiple scales and shape variations. | - |
| dcterms.abstract | Furthermore, we investigate the model design to boost the localization and classification performance for defect semantic segmentation. We propose an attentive boundary-aware transformer framework, namely ABFormer, to precisely segment different types of defects. This framework introduces a feature fusion scheme to split and fuse the boundary and context features with two different attention modules. This facilitates the different learning aspects of the attention modules. In addition, the two attention modules capture the spatial and channel interdependencies of the features, respectively, to address the intraclass difference and interclass indiscrimination problems. | - |
| dcterms.abstract | Finally, we focus on improving the generalization ability of pavement crack detection models. We propose a contrastive decoupling network (CDNet) to effectively detect cracks in seen and unseen domains. This framework separately extracts global and local features with contrastive learning to produce generalized and discriminative representations. Besides, it introduces a semantic enhancement module, detail refinement module, and feature aggregation scheme to tackle diverse cracks with complex backgrounds in input images. | - |
| dcterms.abstract | The vision-based defect inspection models proposed in this thesis are evaluated by comparing them with other state-of-the-art methods on different defect inspection datasets. Experimental results validate that our models can achieve promising performance. These models have great potential to advance deep learning-based methods for various applications of vision-based defect inspection. | - |
| dcterms.accessRights | open access | - |
| dcterms.educationLevel | Ph.D. | - |
| dcterms.extent | xxvi, 124 pages : color illustrations | - |
| dcterms.issued | 2024 | - |
| dcterms.LCSH | Deep learning (Machine learning) | - |
| dcterms.LCSH | Engineering inspection -- Automation | - |
| dcterms.LCSH | Computer vision -- Industrial applications | - |
| dcterms.LCSH | Steel -- Surfaces --Defects | - |
| dcterms.LCSH | Image segmentation | - |
| dcterms.LCSH | Pavements --Cracking | - |
| dcterms.LCSH | Hong Kong Polytechnic University -- Dissertations | - |
| Appears in Collections: | Thesis | |
Access
View full-text via https://theses.lib.polyu.edu.hk/handle/200/13272
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.


