Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/79034
DC Field | Value | Language |
---|---|---|
dc.contributor | Department of Computing | en_US |
dc.creator | Huang, MX | en_US |
dc.creator | Li, JJ | en_US |
dc.creator | Ngai, G | en_US |
dc.creator | Leong, HV | en_US |
dc.creator | Hua, KA | en_US |
dc.date.accessioned | 2018-10-26T01:22:11Z | - |
dc.date.available | 2018-10-26T01:22:11Z | - |
dc.identifier.issn | 1520-9210 | en_US |
dc.identifier.uri | http://hdl.handle.net/10397/79034 | - |
dc.language.iso | en | en_US |
dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
dc.rights | © 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en_US |
dc.rights | The following publication M. X. Huang, J. Li, G. Ngai, H. V. Leong and K. A. Hua, "Fast-PADMA: Rapidly Adapting Facial Affect Model From Similar Individuals," in IEEE Transactions on Multimedia, vol. 20, no. 7, pp. 1901-1915, July 2018 is available at https://doi.org/10.1109/TMM.2017.2775206. | en_US |
dc.subject | Affective computing | en_US |
dc.subject | Facial affect | en_US |
dc.subject | Rapid modeling | en_US |
dc.subject | User-adaptive model | en_US |
dc.title | Fast-PADMA : rapidly adapting facial affect mode from similar individuals | en_US |
dc.type | Journal/Magazine Article | en_US |
dc.identifier.spage | 1901 | en_US |
dc.identifier.epage | 1915 | en_US |
dc.identifier.volume | 20 | en_US |
dc.identifier.issue | 7 | en_US |
dc.identifier.doi | 10.1109/TMM.2017.2775206 | en_US |
dcterms.abstract | A user-specific model generally performs better in facial affect recognition. Existing solutions, however, have usability issues since the annotation can be long and tedious for the end users (e.g., consumers). We address this critical issue by presenting a more user-friendly user-adaptive model to make the personalized approach more practical. This paper proposes a novel user-adaptive model, which we have called fast-Personal Affect Detection with Minimal Annotation (Fast-PADMA). Fast-PADMA integrates data from multiple source subjects with a small amount of data from the target subject. Collecting this target subject data is feasible since fast-PADMA requires only one self-reported affect annotation per facial video segment. To alleviate overfitting in this context of limited individual training data, we propose an efficient bootstrapping technique, which strengthens the contribution of multiple similar source subjects. Specifically, we employ an ensemble classifier to construct pretrained weak generic classifiers from data of multiple source subjects, which is weighted according to the available data from the target user. The result is a model that does not require expensive computation, such as distribution dissimilarity calculation or model retraining. We evaluate our method with in-depth experimental evaluations on five publicly available facial datasets, with results that compare favorably with the state-of-the-art performance on classifying pain, arousal, and valence. Our findings show that fast-PADMA is effective at rapidly constructing a user-adaptive model that outperforms both its generic and user-specific counterparts. This efficient technique has the potential to significantly improve user-adaptive facial affect recognition for personal use and, therefore, enable comprehensive affect-aware applications. | en_US |
dcterms.accessRights | open access | en_US |
dcterms.bibliographicCitation | IEEE transactions on multimedia, July 2018, v. 20, no. 7, p. 1901-1915 | en_US |
dcterms.isPartOf | IEEE transactions on multimedia | en_US |
dcterms.issued | 2018-07 | - |
dc.identifier.isi | WOS:000435570100024 | - |
dc.identifier.eissn | 1941-0077 | en_US |
dc.identifier.rosgroupid | 2017005355 | - |
dc.description.ros | 2017-2018 > Academic research: refereed > Publication in refereed journal | en_US |
dc.description.validate | 201810 bcrc | en_US |
dc.description.oa | Accepted Manuscript | en_US |
dc.identifier.FolderNumber | COMP-0891 | - |
dc.description.fundingSource | RGC | en_US |
dc.description.pubStatus | Published | en_US |
dc.identifier.OPUS | 6802391 | - |
Appears in Collections: | Journal/Magazine Article |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Li_Fast-Padma_Rapidly_Adapting.pdf | Pre-Published version | 1.63 MB | Adobe PDF | View/Open |
Page views
69
Last Week
0
0
Last month
Citations as of Apr 21, 2024
Downloads
75
Citations as of Apr 21, 2024
SCOPUSTM
Citations
2
Citations as of Apr 19, 2024
WEB OF SCIENCETM
Citations
2
Last Week
0
0
Last month
Citations as of Apr 18, 2024
Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.