Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/79034
PIRA download icon_1.1View/Download Full Text
Title: Fast-PADMA : rapidly adapting facial affect mode from similar individuals
Authors: Huang, MX
Li, JJ 
Ngai, G 
Leong, HV 
Hua, KA
Issue Date: Jul-2018
Source: IEEE transactions on multimedia, July 2018, v. 20, no. 7, p. 1901-1915
Abstract: A user-specific model generally performs better in facial affect recognition. Existing solutions, however, have usability issues since the annotation can be long and tedious for the end users (e.g., consumers). We address this critical issue by presenting a more user-friendly user-adaptive model to make the personalized approach more practical. This paper proposes a novel user-adaptive model, which we have called fast-Personal Affect Detection with Minimal Annotation (Fast-PADMA). Fast-PADMA integrates data from multiple source subjects with a small amount of data from the target subject. Collecting this target subject data is feasible since fast-PADMA requires only one self-reported affect annotation per facial video segment. To alleviate overfitting in this context of limited individual training data, we propose an efficient bootstrapping technique, which strengthens the contribution of multiple similar source subjects. Specifically, we employ an ensemble classifier to construct pretrained weak generic classifiers from data of multiple source subjects, which is weighted according to the available data from the target user. The result is a model that does not require expensive computation, such as distribution dissimilarity calculation or model retraining. We evaluate our method with in-depth experimental evaluations on five publicly available facial datasets, with results that compare favorably with the state-of-the-art performance on classifying pain, arousal, and valence. Our findings show that fast-PADMA is effective at rapidly constructing a user-adaptive model that outperforms both its generic and user-specific counterparts. This efficient technique has the potential to significantly improve user-adaptive facial affect recognition for personal use and, therefore, enable comprehensive affect-aware applications.
Keywords: Affective computing
Facial affect
Rapid modeling
User-adaptive model
Publisher: Institute of Electrical and Electronics Engineers
Journal: IEEE transactions on multimedia 
ISSN: 1520-9210
EISSN: 1941-0077
DOI: 10.1109/TMM.2017.2775206
Rights: © 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
The following publication M. X. Huang, J. Li, G. Ngai, H. V. Leong and K. A. Hua, "Fast-PADMA: Rapidly Adapting Facial Affect Model From Similar Individuals," in IEEE Transactions on Multimedia, vol. 20, no. 7, pp. 1901-1915, July 2018 is available at https://doi.org/10.1109/TMM.2017.2775206.
Appears in Collections:Journal/Magazine Article

Files in This Item:
File Description SizeFormat 
Li_Fast-Padma_Rapidly_Adapting.pdfPre-Published version1.63 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Page views

64
Last Week
0
Last month
Citations as of Mar 24, 2024

Downloads

71
Citations as of Mar 24, 2024

SCOPUSTM   
Citations

2
Citations as of Mar 29, 2024

WEB OF SCIENCETM
Citations

2
Last Week
0
Last month
Citations as of Mar 28, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.