Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/106892
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineeringen_US
dc.creatorYi, Len_US
dc.creatorMak, MWen_US
dc.date.accessioned2024-06-07T00:58:41Z-
dc.date.available2024-06-07T00:58:41Z-
dc.identifier.isbn978-1-7138-2069-7en_US
dc.identifier.urihttp://hdl.handle.net/10397/106892-
dc.description21st Annual Conference of the International Speech Communication Association, INTERSPEECH 2020, 25-29 October 2020, Shanghai, Chinaen_US
dc.language.isoenen_US
dc.publisherInternational Speech Communication Association (ISCA)en_US
dc.rightsCopyright © 2020 ISCAen_US
dc.rightsThe following publication Yi, L., Mak, M.-W. (2020) Adversarial Separation and Adaptation Network for Far-Field Speaker Verification. Proc. Interspeech 2020, 4298-4302 is available at https://doi.org/10.21437/Interspeech.2020-2372.en_US
dc.titleAdversarial separation and adaptation network for far-field speaker verificationen_US
dc.typeConference Paperen_US
dc.identifier.spage4298en_US
dc.identifier.epage4302en_US
dc.identifier.doi10.21437/Interspeech.2020-2372en_US
dcterms.abstractTypically, speaker verification systems are highly optimized on the speech collected by close-talking microphones. However, these systems will perform poorly when the users use far-field microphones during verification. In this paper, we propose an adversarial separation and adaptation network (ADSAN) to extract speaker discriminative and domain-invariant features through adversarial learning. The idea is based on the notion that speaker embedding comprises domain-specific components and domain-shared components, and that the two components can be disentangled by the interplay of the separation network and the adaptation network in the ADSAN. We also propose to incorporate a mutual information neural estimator into the domain adaptation network to retain speaker discriminative information. Experiments on the VOiCES Challenge 2019 demonstrate that the proposed approaches can produce more domain-invariant and speaker discriminative representations, which could help to reduce the domain shift caused by different types of microphones and reverberant environments.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH 2020, 25-29 October 2020, Shanghai, China, p. 4298-4302en_US
dcterms.issued2020-
dc.identifier.scopus2-s2.0-85098122245-
dc.relation.conferenceInternational Speech Communication Association [Interspeech]en_US
dc.description.validate202405 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberEIE-0143-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextNSFCen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS55968715-
dc.description.oaCategoryVoR alloweden_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
yi20b_interspeech.pdf381.14 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

60
Citations as of May 11, 2025

Downloads

25
Citations as of May 11, 2025

SCOPUSTM   
Citations

8
Citations as of Jun 12, 2025

WEB OF SCIENCETM
Citations

8
Citations as of Jun 5, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.