Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116814
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorZhang, X-
dc.creatorZhang, W-
dc.creatorWei, X-
dc.creatorWu, J-
dc.creatorZhang, Z-
dc.creatorLei, Z-
dc.creatorLi, Q-
dc.date.accessioned2026-01-21T03:52:52Z-
dc.date.available2026-01-21T03:52:52Z-
dc.identifier.isbn979-8-4007-0686-8-
dc.identifier.urihttp://hdl.handle.net/10397/116814-
dc.description32nd ACM International Conference on Multimedia, Melbourne VIC, Australia, 28 October 2024 - 1 November 2024en_US
dc.language.isoenen_US
dc.publisherThe Association for Computing Machineryen_US
dc.rightsThis work is licensed under a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0).en_US
dc.rights©2024 Copyright held by the owner/author(s).en_US
dc.rightsThe following publication Zhang, X., Zhang, W., Wei, X., Wu, J., Zhang, Z., Lei, Z., & Li, Q. (2024). Generative Active Learning for Image Synthesis Personalization Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne VIC, Australia is available at https://doi.org/10.1145/3664647.3680773.en_US
dc.subjectGenerative active learningen_US
dc.subjectImage synthesisen_US
dc.subjectPersonalizationen_US
dc.titleGenerative active learning for image synthesis personalizationen_US
dc.typeConference Paperen_US
dc.identifier.spage10669-
dc.identifier.epage10677-
dc.identifier.doi10.1145/3664647.3680773-
dcterms.abstractThis paper presents a pilot study that explores the application of active learning, traditionally studied in the context of discriminative models, to generative models. We specifically focus on image synthesis personalization tasks. The primary challenge in conducting active learning on generative models lies in the open-ended nature of querying, which differs from the closed form of querying in discriminative models that typically target a single concept. We introduce the concept of anchor directions to transform the querying process into a semi-open problem. We propose a direction-based uncertainty sampling strategy to enable generative active learning and tackle the exploitation-exploration dilemma. Extensive experiments are conducted to validate the effectiveness of our approach, demonstrating that an open-source model can achieve superior performance compared to closed-source models developed by large companies, such as Google's StyleDrop. The source code is available at https://github.com/zhangxulu1996/GAL4Personalization.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn MM ’24: Proceedings of the 32nd ACM International Conference on Multimedia, p. 10669-10677. New York, NY: The Association for Computing Machinery, 2024-
dcterms.issued2024-
dc.identifier.scopus2-s2.0-85209775546-
dc.relation.ispartofbookMM ’24: Proceedings of the 32nd ACM International Conference on Multimedia-
dc.relation.conferenceACM International Conference on Multimedia [MM]-
dc.publisher.placeNew York, NYen_US
dc.description.validate202601 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis research was supported by the HK RGC Theme-based Research Scheme (PolyU No.: T43-513/23-N), the National Natural Science Foundation of China (Grant No.: 62372314 and 62276254), and InnoHK program.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
3664647.3680773.pdf12.44 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.