scispace - formally typeset
B

Benjamin Caine

Researcher at Google

Publications -  21
Citations -  2249

Benjamin Caine is an academic researcher from Google. The author has contributed to research in topics: Object detection & Computer science. The author has an hindex of 7, co-authored 15 publications receiving 737 citations.

Papers
More filters
Posted Content

Scalability in Perception for Autonomous Driving: Waymo Open Dataset

TL;DR: This work introduces a new large scale, high quality, diverse dataset, consisting of well synchronized and calibrated high quality LiDAR and camera data captured across a range of urban and suburban geographies, and studies the effects of dataset size and generalization across geographies on 3D detection methods.
Proceedings ArticleDOI

Scalability in Perception for Autonomous Driving: Waymo Open Dataset

TL;DR: In this paper, a large scale, high quality, and diverse dataset for self-driving data is presented, consisting of LiDAR and camera data captured across a range of urban and suburban geographies.
Posted Content

StarNet: Targeted Computation for Object Detection in Point Clouds.

TL;DR: This work presents an object detection system called StarNet designed specifically to take advantage of the sparse and 3D nature of point cloud data, and shows how this design leads to competitive or superior performance on the large Waymo Open Dataset and the KITTI detection dataset, as compared to convolutional baselines.
Proceedings ArticleDOI

DeepFusion: Lidar-Camera Deep Fusion for Multi-Modal 3D Object Detection

TL;DR: This paper proposes two novel techniques: InverseAug that inverses geometric-related augmentations, e.g., rotation, to enable accurate geometric alignment between lidar points and image pixels, and LearnableAlign that leverages cross-attention to dynamically capture the correlations between image and lidar features during fusion.
Posted Content

Large Scale Interactive Motion Forecasting for Autonomous Driving : The Waymo Open Motion Dataset

TL;DR: In this article, the authors introduce a large-scale interactive motion dataset with over 100,000 scenes, each 20 seconds long at 10 Hz, collected by mining for interesting interactions between vehicles, pedestrians, and cyclists across six cities within the United States.