Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/98870
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorDang, EKFen_US
dc.creatorLuk, RWPen_US
dc.creatorAllan, Jen_US
dc.date.accessioned2023-06-01T06:05:18Z-
dc.date.available2023-06-01T06:05:18Z-
dc.identifier.issn1046-8188en_US
dc.identifier.urihttp://hdl.handle.net/10397/98870-
dc.language.isoenen_US
dc.publisherAssociation for Computing Machinaryen_US
dc.rights© 2021 Association for Computing Machinery. This is the author's version of the work. It is posted here for your personal use. Not for redistribution. The definitive Version of Record was published in ACM Transactions on Information Systems, http://dx.doi.org/10.1145/3483612.en_US
dc.subjectComparisonen_US
dc.subjectEvaluationen_US
dc.subjectInformation retrievalen_US
dc.subjectMultiple hypotheses testingen_US
dc.subjectRetrieval modelen_US
dc.titleA comparison between term-independence retrieval models for ad hoc retrievalen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume40en_US
dc.identifier.issue3en_US
dc.identifier.doi10.1145/3483612en_US
dcterms.abstractIn Information Retrieval, numerous retrieval models or document ranking functions have been developed in the quest for better retrieval effectiveness. Apart from some formal retrieval models formulated on a theoretical basis, various recent works have applied heuristic constraints to guide the derivation of document ranking functions. While many recent methods are shown to improve over established and successful models, comparison among these new methods under a common environment is often missing. To address this issue, we perform an extensive and up-To-date comparison of leading term-independence retrieval models implemented in our own retrieval system. Our study focuses on the following questions: (RQ1) Is there a retrieval model that consistently outperforms all other models across multiple collections; (RQ2) What are the important features of an effective document ranking function? Our retrieval experiments performed on several TREC test collections of a wide range of sizes (up to the terabyte-sized Clueweb09 Category B) enable us to answer these research questions. This work also serves as a reproducibility study for leading retrieval models. While our experiments show that no single retrieval model outperforms all others across all tested collections, some recent retrieval models, such as MATF and MVD, consistently perform better than the common baselines.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationACM transactions on information systems, July 2022, v. 40, no. 3, 62en_US
dcterms.isPartOfACM transactions on information systemsen_US
dcterms.issued2022-07-
dc.identifier.scopus2-s2.0-85127622824-
dc.identifier.eissn1558-2868en_US
dc.identifier.artn62en_US
dc.description.validate202305 bcwwen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera2050-
dc.identifier.SubFormID46378-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHK PolyU project P0030932en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Dang_Comparison_Term-independence_Retrieval .pdfPre-Published version753.57 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

82
Citations as of Apr 14, 2025

Downloads

97
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

3
Citations as of Sep 12, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.