Corpus ID: 233219495

Back-tracing Representative Points for Voting-based 3D Object Detection in Point Clouds

  title={Back-tracing Representative Points for Voting-based 3D Object Detection in Point Clouds},
  author={Bowen Cheng and Lu Sheng and Shaoshuai Shi and Ming Yang and Dong Xu},
  • Bowen Cheng, Lu Sheng, +2 authors Dong Xu
  • Published in CVPR 2021
  • Computer Science
3D object detection in point clouds is a challenging vision task that benefits various applications for understanding the 3D visual world. Lots of recent research focuses on how to exploit end-to-end trainable Hough voting for generating object proposals. However, the current voting strategy can only receive partial votes from the surfaces of potential objects together with severe outlier votes from the cluttered backgrounds, which hampers full utilization of the information from the input… Expand

Figures and Tables from this paper

Multi-Modality Task Cascade for 3D Object Detection
A novel MultiModality Task Cascade network (MTC-RCNN) that leverages 3D box proposals to improve 2D segmentation predictions, which are then used to further refine the 3D boxes and shows that including a 2D network between two stages of 3D modules significantly improves both 2D and 3D task performance. Expand
Suppress-and-Refine Framework for End-to-End 3D Object Detection
This work presents the first fully end-to-end 3D detector, SRDet, on the basis of VoteNet, which achieves state-of-the-art performance on the challenging ScanNetV2 and SUN RGB-D datasets with the fastest speed ever. Expand
VoteHMR: Occlusion-Aware Voting Network for Robust 3D Human Mesh Recovery from Partial Point Clouds
  • Guanze Liu, Yu Rong, Lu Sheng
  • Computer Science
  • ACM Multimedia
  • 2021
The first attempt to reconstruct reliable 3D human shapes from single-frame partial point clouds is made, using an end-to-end learnable method, named VoteHMR, which achieves state-of-the-art performances on two large-scale datasets, namely SURREAL and DFAUST. Expand


Deep Hough Voting for 3D Object Detection in Point Clouds
This work proposes VoteNet, an end-to-end 3D object detection network based on a synergy of deep point set networks and Hough voting that achieves state-of-the-art 3D detection on two large datasets of real 3D scans, ScanNet and SUN RGB-D with a simple design, compact model size and high efficiency. Expand
A Hierarchical Graph Network for 3D Object Detection on Point Clouds
A new graph convolution (GConv) based hierarchical graph network (HGNet) for 3D object detection, which processes raw point clouds directly to predict 3D bounding boxes and outperforms state-of-the-art methods on two large-scale point cloud datasets. Expand
Density-Based Clustering for 3D Object Detection in Point Clouds
This work introduces a novel approach for 3D object detection that is significant in two main aspects: a cascaded modular approach that focuses the receptive field of each module on specific points in the point cloud, for improved feature learning and a class agnostic instance segmentation module that is initiated using unsupervised clustering. Expand
PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud
Extensive experiments on the 3D detection benchmark of KITTI dataset show that the proposed architecture outperforms state-of-the-art methods with remarkable margins by using only point cloud as input. Expand
3D-MPA: Multi-Proposal Aggregation for 3D Semantic Instance Segmentation
It is shown that grouping proposals improves over NMS and outperforms previous state-of-the-art methods on the tasks of 3D object detection and semantic instance segmentation on the ScanNetV2 benchmark and the S3DIS dataset. Expand
Frustum PointNets for 3D Object Detection from RGB-D Data
This work directly operates on raw point clouds by popping up RGBD scans and leverages both mature 2D object detectors and advanced 3D deep learning for object localization, achieving efficiency as well as high recall for even small objects. Expand
VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection
  • Yin Zhou, Oncel Tuzel
  • Computer Science
  • 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
  • 2018
VoxelNet is proposed, a generic 3D detection network that unifies feature extraction and bounding box prediction into a single stage, end-to-end trainable deep network and learns an effective discriminative representation of objects with various geometries, leading to encouraging results in3D detection of pedestrians and cyclists. Expand
2D-Driven 3D Object Detection in RGB-D Images
The approach makes best use of the 2D information to quickly reduce the search space in 3D, benefiting from state-of-the-art 2D object detection techniques. Expand
Multi-view 3D Object Detection Network for Autonomous Driving
This paper proposes Multi-View 3D networks (MV3D), a sensory-fusion framework that takes both LIDAR point cloud and RGB images as input and predicts oriented 3D bounding boxes and designs a deep fusion scheme to combine region-wise features from multiple views and enable interactions between intermediate layers of different paths. Expand
CenterNet: Keypoint Triplets for Object Detection
This paper presents an efficient solution that explores the visual patterns within individual cropped regions with minimal costs, and builds the framework upon a representative one-stage keypoint-based detector named CornerNet, which improves both precision and recall. Expand