Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105557
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorLu, Pen_US
dc.creatorZhu, Fen_US
dc.creatorLi, Pen_US
dc.creatorKim, Jen_US
dc.creatorSheng, Ben_US
dc.creatorMao, Len_US
dc.date.accessioned2024-04-15T07:35:01Z-
dc.date.available2024-04-15T07:35:01Z-
dc.identifier.isbn978-3-030-61863-6en_US
dc.identifier.isbn978-3-030-61864-3 (eBook)en_US
dc.identifier.issn0302-9743en_US
dc.identifier.urihttp://hdl.handle.net/10397/105557-
dc.description37th Computer Graphics International Conference, CGI 2020, Geneva, Switzerland, October 20–23, 2020en_US
dc.language.isoenen_US
dc.publisherSpringeren_US
dc.rights© Springer Nature Switzerland AG 2020en_US
dc.rightsThis version of the proceeding paper has been accepted for publication, after peer review (when applicable) and is subject to Springer Nature’s AM terms of use(https://www.springernature.com/gp/open-research/policies/accepted-manuscript-terms), but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: http://dx.doi.org/10.1007/978-3-030-61864-3_3.en_US
dc.subjectEye-head coordinationen_US
dc.subjectHand tracken_US
dc.subjectHierarchical renderingen_US
dc.subjectLODen_US
dc.subjectVRen_US
dc.titleHierarchical rendering system based on viewpoint prediction in virtual realityen_US
dc.typeConference Paperen_US
dc.identifier.spage24en_US
dc.identifier.epage32en_US
dc.identifier.volume12221en_US
dc.identifier.doi10.1007/978-3-030-61864-3_3en_US
dcterms.abstractVirtual reality (VR) systems use multi-modal interfaces to explore three-dimensional virtual worlds. During exploration, the user may look at different objects of interest or in different directions. The field of view of human vision is 135∘× 160∘, but the one requiring the highest resolution is only in 1.5∘× 2∘. It is estimated that in modern VR, only 4% of the pixel resources of the head-mounted display are mapped to the visual center. Therefore, allocating more computing resources to the visual center and allocating fewer viewpoint prediction rendering techniques elsewhere can greatly speed up the rendering of the scene, especially for VR devices equipped with eye trackers. However, eye trackers as additional equipment may be relatively expensive and be harder to use, at the same time, there is considerable work to be done in the development of eye trackers and their integration with commercial head-mounted equipment. Therefore, this article uses an eye-head coordination model combined with the saliencey of the scene to predict the gaze position, and then uses a hybrid method of Level of Detail (LOD) and grid degeneration to reduce rendering time as much as possible without losing the perceived details and required calculations.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationLecture notes in computer science (including subseries Lecture notes in artificial intelligence and lecture notes in bioinformatics), 2020, v. 12221, p. 24-32en_US
dcterms.isPartOfLecture notes in computer science (including subseries Lecture notes in artificial intelligence and lecture notes in bioinformatics)en_US
dcterms.issued2020-
dc.identifier.scopus2-s2.0-85096514776-
dc.relation.conferenceComputer Graphics International Conference [CGI]en_US
dc.identifier.eissn1611-3349en_US
dc.description.validate202402 bcchen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberCOMP-0441-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextNational Key Research and Development Program of China; National Natural Science Foundation of China; Science and Technology Commission of Shanghai Municipalityen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS43142416-
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Li_Hierarchical_Rendering_System.pdfPre-Published version1.61 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

109
Last Week
3
Last month
Citations as of Nov 30, 2025

Downloads

37
Citations as of Nov 30, 2025

SCOPUSTM   
Citations

1
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

3
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.