Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/109489
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorHe, Cen_US
dc.creatorLi, Ren_US
dc.creatorLi, Sen_US
dc.creatorZhang, Len_US
dc.date.accessioned2024-11-01T08:04:36Z-
dc.date.available2024-11-01T08:04:36Z-
dc.identifier.isbn978-1-6654-6946-3en_US
dc.identifier.urihttp://hdl.handle.net/10397/109489-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication C. He, R. Li, S. Li and L. Zhang, "Voxel Set Transformer: A Set-to-Set Approach to 3D Object Detection from Point Clouds," 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 2022, pp. 8407-8417 is available at https://doi.org/10.1109/CVPR52688.2022.00823.en_US
dc.titleVoxel set transformer : a set-to-set approach to 3D object detection from point cloudsen_US
dc.typeConference Paperen_US
dc.identifier.spage8407en_US
dc.identifier.epage8417en_US
dc.identifier.doi10.1109/CVPR52688.2022.00823en_US
dcterms.abstractTransformer has demonstrated promising performance in many 2D vision tasks. However, it is cumbersome to compute the self-attention on large-scale point cloud data because point cloud is a long sequence and unevenly distributed in 3D space. To solve this issue, existing methods usually compute self-attention locally by grouping the points into clusters of the same size, or perform convolutional self-attention on a discretized representation. However, the former results in stochastic point dropout, while the latter typically has narrow attention fields. In this paper, we propose a novel voxel-based architecture, namely Voxel Set Transformer (VoxSeT), to detect 3D objects from point clouds by means of set-to-set translation. VoxSeT is built upon a voxel-based set attention (VSA) module, which reduces the self-attention in each voxel by two cross-attentions and models features in a hidden space induced by a group of latent codes. With the VSA module, VoxSeT can manage voxelized point clusters with arbitrary size in a wide range, and process them in parallel with linear complexity. The proposed VoxSeT integrates the high performance of transformer with the efficiency of voxel-based model, which can be used as a good alternative to the convolutional and point-based backbones. VoxSeT reports competitive results on the KITTI and Waymo detection benchmarks. The source codes can be found at https://github.com/skyhehe123/VoxSeT.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitation2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition : New Orleans, Louisiana, 19 - 24 June 2022, p. 8407-8417en_US
dcterms.issued2022-
dc.identifier.scopus2-s2.0-85137827073-
dc.relation.ispartofbook2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition : New Orleans, Louisiana, 19 - 24 June 2022en_US
dc.relation.conferenceConference on Computer Vision and Pattern Recognition [CVPR]-
dc.description.validate202411 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberOA_Others-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
He_Voxel_Set_Transformer.pdfPre-Published version2.33 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

45
Citations as of Apr 14, 2025

Downloads

31
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

168
Citations as of May 29, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.