Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/113739
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineering-
dc.creatorKwong, NW-
dc.creatorChan, YL-
dc.creatorTsang, SH-
dc.creatorHuang, Z-
dc.creatorLam, KM-
dc.date.accessioned2025-06-19T06:25:02Z-
dc.date.available2025-06-19T06:25:02Z-
dc.identifier.issn0018-9316-
dc.identifier.urihttp://hdl.handle.net/10397/113739-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication N. -W. Kwong, Y. -L. Chan, S. -H. Tsang, Z. Huang and K. -M. Lam, "Deep Learning Approach for No-Reference Screen Content Video Quality Assessment," in IEEE Transactions on Broadcasting, vol. 70, no. 2, pp. 555-569, June 2024 is available at https://doi.org/10.1109/TBC.2024.3374042.en_US
dc.subjectHuman visual experienceen_US
dc.subjectMulti-channel convolutional neural networken_US
dc.subjectMulti-task learningen_US
dc.subjectNo reference video quality assessmenten_US
dc.subjectScreen content video quality assessmenten_US
dc.subjectSelfsupervised learningen_US
dc.subjectSpatiotemporal featuresen_US
dc.titleDeep learning approach for no-reference screen content video quality assessmenten_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage555-
dc.identifier.epage569-
dc.identifier.volume70-
dc.identifier.issue2-
dc.identifier.doi10.1109/TBC.2024.3374042-
dcterms.abstractScreen content video (SCV) has drawn much more attention than ever during the COVID-19 period and has evolved from a niche to a mainstream due to the recent proliferation of remote offices, online meetings, shared-screen collaboration, and gaming live streaming. Therefore, quality assessments for screen content media are highly demanded to maintain service quality recently. Although many practical natural scene video quality assessment methods have been proposed and achieved promising results, these methods cannot be applied to the screen content video quality assessment (SCVQA) task directly since the content characteristics of SCV are substantially different from natural scene video. Besides, only one no-reference SCVQA (NR-SCVQA) method, which requires handcrafted features, has been proposed in the literature. Therefore, we propose the first deep learning approach explicitly designed for NR-SCVQA. First, a multi-channel convolutional neural network (CNN) model is used to extract spatial quality features of pictorial and textual regions separately. Since there is no human annotated quality for each screen content frame (SCF), the CNN model is pre-trained in a multi-task self-supervised fashion to extract spatial quality feature representation of SCF. Second, we propose a time-distributed CNN transformer model (TCNNT) to further process all SCF spatial quality feature representations of an SCV and learn spatial and temporal features simultaneously so that high-level spatiotemporal features of SCV can be extracted and used to assess the whole SCV quality. Experimental results demonstrate the robustness and validity of our model, which is closely related to human perception.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on broadcasting, June 2024, v. 70, no. 2, p. 555-569-
dcterms.isPartOfIEEE transactions on broadcasting-
dcterms.issued2024-06-
dc.identifier.scopus2-s2.0-85189167228-
dc.identifier.eissn1557-9611-
dc.description.validate202506 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera3728ben_US
dc.identifier.SubFormID50879en_US
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextInnovation and Technology Fund - Partnership Research Programme (ITF-PRP) under PRP/036/21FXen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Kwong_Deep_Learning_Approach.pdfPre-Published version1.52 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.