Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/111613
| Title: | Navigating the unseen : out-of-distribution generalization and detection in open environments | Authors: | Zhang, Yabin | Degree: | Ph.D. | Issue Date: | 2024 | Abstract: | In open environments, artificial intelligence (AI) models face two main types of out-of-distribution (OOD) samples that deviate from their training data (i.e., in-distribution data): covariate-shifted OOD samples, which are consistent in semantics but differ in covariate shifts, and semantic-shifted OOD samples, which have different semantic labels. Such OOD samples can severely challenge the safety and reliability of AI systems by inducing high-confidence errors. Comprising four studies, this thesis targets enhancing generalization to covariate shifts through methods like style augmentation and memory networks, and improving detection of semantic-shifted samples using strategies such as prompt tuning and adaptive negative proxies. These efforts are crucial for the reliable performance of AI models in open environments. In Chapter 1, we introduce the concepts of covariate-shifted and semantic-shifted OOD samples and review existing methods and challenges associated with OOD generalization and detection. We detail the objectives, contributions, and the structure of the thesis. In Chapter 2, we introduce Exact Feature Distribution Matching (EFDM), a novel technique that advances style augmentation by integrating higher-order statistics for enhanced generalization to covariate-shifted OOD samples. EFDM employs empirical Cumulative Distribution Functions and a Sort-Matching technique, demonstrating superior performance over traditional methods in extensive experiments. In Chapter 3, we develop the dual memory networks to extend the generalization capabilities of vision-language models (VLMs) like CLIP. This strategy significantly improves performance on both in-distribution and covariate-shifted OOD samples, validated through rigorous testing across a variety of datasets. Moving forward, Chapters 4 and 5 focus on detecting semantic-shifted OOD samples. In Chapter 4, we introduce Label-driven Automated Prompt Tuning (LAPT) to address the limitations of manual prompt engineering in VLMs-based OOD detection. Using distribution-aware prompts and automatically collected negative training data, LAPT reduces manual effort and improves detection performance across various tasks. In Chapter 5, we focus on constructing adaptive negative proxy with test images in a test-time adaption manner. This approach facilitates online mining of negative test samples, enhancing the model’s ability to distinguish between in-distribution and OOD instances, as proven on standard benchmarks. In summary, this thesis contributes to the field of OOD generalization and detection by introducing innovative methods that enhance performance and reduce manual intervention. By addressing specific challenges associated with covariate and semantic shifts in OOD samples, these studies significantly improve the reliability and safety of AI systems in open environments. |
Subjects: | Artificial intelligence Machine learning Hong Kong Polytechnic University -- Dissertations |
Pages: | xix, 157 pages : color illustrations |
| Appears in Collections: | Thesis |
Access
View full-text via https://theses.lib.polyu.edu.hk/handle/200/13423
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.


