Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/115064
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineering-
dc.creatorJing, M-
dc.creatorSethu, V-
dc.creatorAhmed, B-
dc.creatorLee, KA-
dc.date.accessioned2025-09-09T07:40:27Z-
dc.date.available2025-09-09T07:40:27Z-
dc.identifier.issn0885-2308-
dc.identifier.urihttp://hdl.handle.net/10397/115064-
dc.language.isoenen_US
dc.publisherAcademic Pressen_US
dc.rights© 2025 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Jing, M., Sethu, V., Ahmed, B., & Lee, K. A. (2025). Quantifying prediction uncertainties in automatic speaker verification systems. Computer Speech & Language, 94, 101806 is available at https://doi.org/10.1016/j.csl.2025.101806.en_US
dc.subjectBayes-by-backpropen_US
dc.subjectHamiltonian Monte-Carloen_US
dc.subjectPLDAen_US
dc.subjectSpeaker verificationen_US
dc.subjectStochastic gradient Langevin dynamicsen_US
dc.subjectUncertaintyen_US
dc.titleQuantifying prediction uncertainties in automatic speaker verification systemsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume94-
dc.identifier.doi10.1016/j.csl.2025.101806-
dcterms.abstractFor modern automatic speaker verification (ASV) systems, explicitly quantifying the confidence for each prediction strengthens the system’s reliability by indicating in which case the system is with trust. However, current paradigms do not take this into consideration. We thus propose to express confidence in the prediction by quantifying the uncertainty in ASV predictions. This is achieved by developing a novel Bayesian framework to obtain a score distribution for each input. The mean of the distribution is used to derive the decision while the spread of the distribution represents the uncertainty arising from the plausible choices of the model parameters. To capture the plausible choices, we sample the probabilistic linear discriminant analysis (PLDA) back-end model posterior through Hamiltonian Monte-Carlo (HMC) and approximate the embedding model posterior through stochastic Langevin dynamics (SGLD) and Bayes-by-backprop. Given the resulting score distribution, a further quantification and decomposition of the prediction uncertainty are achieved by calculating the score variance, entropy, and mutual information. The quantified uncertainties include the aleatoric uncertainty and epistemic uncertainty (model uncertainty). We evaluate them by observing how they change while varying the amount of training speech, the duration, and the noise level of testing speech. The experiments indicate that the behaviour of those quantified uncertainties reflects the changes we made to the training and testing data, demonstrating the validity of the proposed method as a measure of uncertainty.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationComputer speech and language, Nov. 2025, v. 94, 101806-
dcterms.isPartOfComputer speech and language-
dcterms.issued2025-11-
dc.identifier.scopus2-s2.0-105004203368-
dc.identifier.eissn1095-8363-
dc.identifier.artn101806-
dc.description.validate202509 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
1-s2.0-S0885230825000312-main.pdf5.42 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.