Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118012
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of English and Communication-
dc.creatorLu, J-
dc.creatorRogers, J-
dc.date.accessioned2026-03-12T01:02:48Z-
dc.date.available2026-03-12T01:02:48Z-
dc.identifier.urihttp://hdl.handle.net/10397/118012-
dc.language.isoenen_US
dc.publisherElsevier Ltd.en_US
dc.rights© 2026 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license ( http://creativecommons.org/licenses/by-nc-nd/4.0/ ).en_US
dc.rightsThe following publication Lu, J., & Rogers, J. (2026). Evaluating and enhancing the accuracy of automated fluency annotation tools in L2 research. Research Methods in Applied Linguistics, 5(1), 100302 is available at https://doi.org/10.1016/j.rmal.2026.100302.en_US
dc.subjectAutomatic fluency assessmenten_US
dc.subjectHybrid automated-manual pipelineen_US
dc.subjectSecond language speechen_US
dc.subjectTemporal fluency featuresen_US
dc.subjectTool comparisonen_US
dc.titleEvaluating and enhancing the accuracy of automated fluency annotation tools in L2 researchen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume5-
dc.identifier.issue1-
dc.identifier.doi10.1016/j.rmal.2026.100302-
dcterms.abstractFluency is a central dimension of L2 oral proficiency. Further, fluency assessment is important for many applied contexts, including pedagogical and assessment purposes. Yet, the measurement of fluency using manual annotation is labor-intensive, which limits its broad application and scalability. We evaluate two automated tools — an acoustic-based tool (de Jong et al., 2021) and a machine-learning tool (Matsuura et al., 2025) — using data from L1-Chinese learners of English. Accuracy was assessed for three metrics, articulation rate (AR), pause ratio (PR), and mean pause duration (MPD), via Pearson correlations with manual annotation. We compared two automated tools and tested whether targeted manual post-processing (TextGrid checks and transcript adjustments) improves metric extraction using Steiger’s test. Results from our sample indicated that de Jong et al. (2021) yielded higher accuracy for silence-based metrics (PR, MPD). However, text-dependent metrics (syllable number after removing disfluency words in AR) benefited from corrected TextGrids (for the acoustic tool) or corrected transcripts (for the machine-learning tool). These findings suggest a scalable division of labor: use an acoustic-based tool for silence-driven metrics, and apply corrected transcripts with a machine-learning tool when extracting text-sensitive metrics.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationResearch methods in applied linguistics, Apr. 2026, v. 5, no. 1, 100302-
dcterms.isPartOfResearch methods in applied linguistics-
dcterms.issued2026-04-
dc.identifier.scopus2-s2.0-105028960517-
dc.identifier.eissn2772-7661-
dc.identifier.artn100302-
dc.description.validate202603 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_TAen_US
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.TAElsevier (2026)en_US
dc.description.oaCategoryTAen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
1-s2.0-S277276612600008X-main.pdf1.63 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.