scispace - formally typeset
Open AccessJournal ArticleDOI

Multi Point-Voxel Convolution (MPVConv) for deep learning on point clouds

Jiang-hua Ding
- 01 May 2023 - 
- Vol. 112, pp 72-80
Reads0
Chats0
TLDR
Wu et al. as mentioned in this paper proposed a new convolutional neural network, called Multi Point-Voxel Convolution (MPVConv), for deep learning on point clouds.
Abstract
The existing 3D deep learning methods adopt either individual point-based features or local-neighboring voxel-based features, and demonstrate great potential for processing 3D data. However, the point-based models are inefficient due to the unordered nature of point clouds and the voxel-based models suffer from large information loss. Motivated by the success of recent point-voxel representation, such as PVCNN and DRINet, we propose a new convolutional neural network, called Multi Point-Voxel Convolution (MPVConv), for deep learning on point clouds. Integrating both the advantages of voxel and point-based methods, MPVConv can effectively increase the neighboring collection between point-based features and also promote independence among voxel-based features. Extensive experiments on benchmark datasets such as ShapeNet Part, S3DIS and KITTI for various tasks show that MPVConv improves the accuracy of the backbone (PointNet) by up to 36%, and achieves higher accuracy than the voxel-based model with up to 34× speedups. In addition, MPVConv outperforms the state-of-the-art point-based models with up to 8× speedups. Also, our MPVConv only needs 65% of the GPU memory required by the latest point-voxel-based model (DRINet). The source code of our method is attached in https://github.com/NWUzhouwei/MPVConv.

read more

Content maybe subject to copyright    Report

Citations
More filters
References
More filters
Journal ArticleDOI

Vision meets robotics: The KITTI dataset

TL;DR: A novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research, using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras and a high-precision GPS/IMU inertial navigation system.
Journal ArticleDOI

Learning of 3D Graph Convolution Networks for Point Cloud Analysis.

TL;DR: Wang et al. as mentioned in this paper proposed 3D Graph Convolution Networks (3D-GCN), which uniquely learns 3D kernels with graph max-pooling mechanisms for extracting geometric features from point cloud data across different scales.
Posted Content

DiffusionNet: Discretization Agnostic Learning on Surfaces

TL;DR: In this paper, a simple diffusion layer is proposed for spatial communication on 3D mesh surfaces, and the spatial support of diffusion is optimized as a continuous network parameter ranging from purely local to totally global, removing the burden of manually choosing neighborhood sizes.
Journal ArticleDOI

DiffusionNet: Discretization Agnostic Learning on Surfaces

TL;DR: In this paper , a simple diffusion layer is proposed for spatial communication and the spatial support of diffusion is optimized as a continuous network parameter ranging from purely local to totally global, removing the burden of manually choosing neighborhood sizes.
Journal ArticleDOI

PVT: Point‐voxel transformer for point cloud learning

TL;DR: Wang et al. as mentioned in this paper proposed a sparse window attention module to gather coarse-grained local features from nonempty voxels, which not only bypasses the expensive irregular data structuring and invalid empty voxel computation, but also obtains linear computational complexity with respect to voxell resolution.