Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/79034
Title: | Fast-PADMA : rapidly adapting facial affect mode from similar individuals | Authors: | Huang, MX Li, JJ Ngai, G Leong, HV Hua, KA |
Issue Date: | Jul-2018 | Source: | IEEE transactions on multimedia, July 2018, v. 20, no. 7, p. 1901-1915 | Abstract: | A user-specific model generally performs better in facial affect recognition. Existing solutions, however, have usability issues since the annotation can be long and tedious for the end users (e.g., consumers). We address this critical issue by presenting a more user-friendly user-adaptive model to make the personalized approach more practical. This paper proposes a novel user-adaptive model, which we have called fast-Personal Affect Detection with Minimal Annotation (Fast-PADMA). Fast-PADMA integrates data from multiple source subjects with a small amount of data from the target subject. Collecting this target subject data is feasible since fast-PADMA requires only one self-reported affect annotation per facial video segment. To alleviate overfitting in this context of limited individual training data, we propose an efficient bootstrapping technique, which strengthens the contribution of multiple similar source subjects. Specifically, we employ an ensemble classifier to construct pretrained weak generic classifiers from data of multiple source subjects, which is weighted according to the available data from the target user. The result is a model that does not require expensive computation, such as distribution dissimilarity calculation or model retraining. We evaluate our method with in-depth experimental evaluations on five publicly available facial datasets, with results that compare favorably with the state-of-the-art performance on classifying pain, arousal, and valence. Our findings show that fast-PADMA is effective at rapidly constructing a user-adaptive model that outperforms both its generic and user-specific counterparts. This efficient technique has the potential to significantly improve user-adaptive facial affect recognition for personal use and, therefore, enable comprehensive affect-aware applications. | Keywords: | Affective computing Facial affect Rapid modeling User-adaptive model |
Publisher: | Institute of Electrical and Electronics Engineers | Journal: | IEEE transactions on multimedia | ISSN: | 1520-9210 | EISSN: | 1941-0077 | DOI: | 10.1109/TMM.2017.2775206 | Rights: | © 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The following publication M. X. Huang, J. Li, G. Ngai, H. V. Leong and K. A. Hua, "Fast-PADMA: Rapidly Adapting Facial Affect Model From Similar Individuals," in IEEE Transactions on Multimedia, vol. 20, no. 7, pp. 1901-1915, July 2018 is available at https://doi.org/10.1109/TMM.2017.2775206. |
Appears in Collections: | Journal/Magazine Article |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Li_Fast-Padma_Rapidly_Adapting.pdf | Pre-Published version | 1.63 MB | Adobe PDF | View/Open |
Page views
64
Last Week
0
0
Last month
Citations as of Mar 24, 2024
Downloads
71
Citations as of Mar 24, 2024
SCOPUSTM
Citations
2
Citations as of Mar 29, 2024
WEB OF SCIENCETM
Citations
2
Last Week
0
0
Last month
Citations as of Mar 28, 2024
Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.