Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105635
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorCai, S-
dc.creatorZuo, W-
dc.creatorDavis, LS-
dc.creatorZhang, L-
dc.date.accessioned2024-04-15T07:35:34Z-
dc.date.available2024-04-15T07:35:34Z-
dc.identifier.isbn978-3-030-01263-2-
dc.identifier.isbn978-3-030-01264-9 (eBook)-
dc.identifier.issn0302-9743-
dc.identifier.urihttp://hdl.handle.net/10397/105635-
dc.description15th European Conference, Munich, Germany, September 8-14, 2018en_US
dc.language.isoenen_US
dc.publisherSpringeren_US
dc.rights© Springer Nature Switzerland AG 2018en_US
dc.rightsThis version of the article has been accepted for publication, after peer review (when applicable) and is subject to Springer Nature’s AM terms of use(https://www.springernature.com/gp/open-research/policies/accepted-manuscript-terms), but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: http://dx.doi.org/10.1007/978-3-030-01264-9_12.en_US
dc.subjectVariational autoencoderen_US
dc.subjectVideo summarizationen_US
dc.titleWeakly-supervised video summarization using variational encoder-decoder and web prioren_US
dc.typeConference Paperen_US
dc.identifier.spage193-
dc.identifier.epage210-
dc.identifier.volume11218-
dc.identifier.doi10.1007/978-3-030-01264-9_12-
dcterms.abstractVideo summarization is a challenging under-constrained problem because the underlying summary of a single video strongly depends on users’ subjective understandings. Data-driven approaches, such as deep neural networks, can deal with the ambiguity inherent in this task to some extent, but it is extremely expensive to acquire the temporal annotations of a large-scale video dataset. To leverage the plentiful web-crawled videos to improve the performance of video summarization, we present a generative modelling framework to learn the latent semantic video representations to bridge the benchmark data and web data. Specifically, our framework couples two important components: a variational autoencoder for learning the latent semantics from web videos, and an encoder-attention-decoder for saliency estimation of raw video and summary generation. A loss term to learn the semantic matching between the generated summaries and web videos is presented, and the overall framework is further formulated into a unified conditional variational encoder-decoder, called variational encoder-summarizer-decoder (VESD). Experiments conducted on the challenging datasets CoSum and TVSum demonstrate the superior performance of the proposed VESD to existing state-of-the-art methods. The source code of this work can be found at https://github.com/cssjcai/vesd.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationLecture notes in computer science (including subseries Lecture notes in artificial intelligence and lecture notes in bioinformatics), 2018, v. 11218, p. 193-210-
dcterms.isPartOfLecture notes in computer science (including subseries Lecture notes in artificial intelligence and lecture notes in bioinformatics)-
dcterms.issued2018-
dc.identifier.scopus2-s2.0-85055675665-
dc.relation.conferenceEuropean Conference on Computer Vision [ECCV]-
dc.identifier.eissn1611-3349-
dc.description.validate202402 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberCOMP-1043en_US
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS13568017en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Cai_Weakly-Supervised_Video_Summarization.pdfPre-Published version1.73 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

167
Last Week
4
Last month
Citations as of Nov 9, 2025

Downloads

101
Citations as of Nov 9, 2025

SCOPUSTM   
Citations

12
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

51
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.