Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/96875
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.creatorSaxena, Den_US
dc.creatorKulshrestha, Ten_US
dc.creatorCao, Jen_US
dc.creatorCheung, SCen_US
dc.date.accessioned2022-12-22T06:38:39Z-
dc.date.available2022-12-22T06:38:39Z-
dc.identifier.issn1057-7149en_US
dc.identifier.urihttp://hdl.handle.net/10397/96875-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication D. Saxena, T. Kulshrestha, J. Cao and S. -C. Cheung, "Multi-Constraint Adversarial Networks for Unsupervised Image-to-Image Translation," in IEEE Transactions on Image Processing, vol. 31, pp. 1601-1612, 2022 is available at https://dx.doi.org/10.1109/TIP.2022.3144886.en_US
dc.subjectGenerative adversarial networksen_US
dc.subjectUnsupervised image-to-image translationen_US
dc.subjectGenerative modelingen_US
dc.subjectGANsen_US
dc.subjectImage synthesisen_US
dc.titleMulti-constraint adversarial networks for unsupervised image-to-image translationen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage1601en_US
dc.identifier.epage1612en_US
dc.identifier.volume31en_US
dc.identifier.doi10.1109/TIP.2022.3144886en_US
dcterms.abstractUnsupervised image-to-image translation aims to learn the mapping from an input image in a source domain to an output image in a target domain without paired training dataset. Recently, remarkable progress has been made in translation due to the development of generative adversarial networks (GANs). However, existing methods suffer from the training instability as gradients passing from discriminator to generator become less informative when the source and target domains exhibit sufficiently large discrepancies in appearance or shape. To handle this challenging problem, in this paper, we propose a novel multi-constraint adversarial model (MCGAN) for image translation in which multiple adversarial constraints are applied at generator’s multi-scale outputs by a single discriminator to pass gradients to all the scales simultaneously and assist generator training for capturing large discrepancies in appearance between two domains. We further notice that the solution to regularize generator is helpful in stabilizing adversarial training, but results may have unreasonable structure or blurriness due to less context information flow from discriminator to generator. Therefore, we adopt dense combinations of the dilated convolutions at discriminator for supporting more information flow to generator. With extensive experiments on three public datasets, cat-to-dog, horse-to-zebra, and apple-to-orange, our method significantly improves state-of-the-arts on all datasets.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on image processing, 2022, v. 31, p. 1601-1612en_US
dcterms.isPartOfIEEE transactions on image processingen_US
dcterms.issued2022-
dc.identifier.isiWOS:000750373700006-
dc.identifier.pmid35081027-
dc.identifier.eissn1941-0042en_US
dc.description.validate202212 bcrcen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera1533-
dc.identifier.SubFormID45362-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextOthers: The Hong Kong Polytechnic University (PolyU) Internal Start-Up Fund under Grant P0038876en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Saxena_Multi-Constraint_Adversarial_Networks.pdfPre-Published version13.15 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

114
Citations as of Apr 14, 2025

Downloads

245
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

15
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

10
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.