Vanets Meet Autonomous Vehicles: Multimodal Surrounding Recognition Using Manifold Alignment
Reads0
Chats0
TLDR
The vision to create a beneficial link between the two worlds of vehicular ad-hoc networks and autonomous vehicles is presented by designing a multimodal scheme for object detection, recognition, and mapping based on the fusion of stereo camera frames, point cloud Velodyne LIDAR scans, and vehicle-to-vehicle (V2V) basic safety messages (BSMs) exchanges using VANET protocols.Abstract:
In the past two years, calls for developing synergistic links between the two worlds of vehicular ad-hoc networks (VANETs) and autonomous vehicles have significantly gone up to achieve further on-road safety and benefits for end-users. In this paper, we present our vision to create such a beneficial link by designing a multimodal scheme for object detection, recognition, and mapping based on the fusion of stereo camera frames, point cloud Velodyne LIDAR scans, and vehicle-to-vehicle (V2V) basic safety messages (BSMs) exchanges using VANET protocols. Exploiting the high similarities in the underlying manifold properties of the three data sets, and their high neighborhood correlation, the proposed scheme employs semi-supervised manifold alignment to merge the key features of rich texture descriptions of objects from 2-D images, depth and distance between objects provided by 3-D point cloud, and the awareness of self-declared vehicles from BSMs’ 3-D information including the ones not seen by camera and LIDAR. The proposed scheme is applied to create joint pixel-to-point-cloud and pixel-to-V2V correspondences of objects in frames from the KITTI Vision Benchmark Suite, using a semi-supervised manifold alignment, to achieve camera-LIDAR and camera-V2V mapping of their recognized objects. We present the alignment accuracy results over two different driving sequences and show the additional acquired knowledge of objects from the various input modalities. We also study the effect of the number of neighbors employed in the alignment process on the alignment accuracy. With proper choice of parameters, the testing of our proposed scheme over two entire driving sequences exhibits 100% accuracy in the majority of cases, 74%–92% and 50%–72% average alignment accuracy for vehicles and pedestrians and up to 150% additional object recognition of the testing vehicle’s surrounding.read more
Citations
More filters
Journal ArticleDOI
A Point Cloud-Based Robust Road Curb Detection and Tracking Method
TL;DR: A robust 3D-LiDAR-based method for road curb detection and tracking in a structured environment and can deal with straight and curved road without being influenced by surrounding obstacles is proposed.
Journal ArticleDOI
Ground Surface Filtering of 3D Point Clouds Based on Hybrid Regression Technique
TL;DR: The results show that the proposed method performs well in most real scenarios, even in the cases of ground undulation, occlusion, and sparse point clouds.
Journal ArticleDOI
Parallel Vehicular Networks: A CPSS-Based Approach via Multimodal Big Data in IoV
TL;DR: The proposed PVN offers a competitive solution for achieving a smooth, safe, and efficient cooperation among connected vehicles in future ITSs and is expected to achieve the descriptive intelligence, predictive intelligence, and prescription intelligence for VNs.
Journal ArticleDOI
A Progressive Review: Emerging Technologies for ADAS Driven Solutions
TL;DR: In this paper , the state-of-the-art of ADAS and its levels of autonomy are reviewed and a detailed description of vision intelligence and computational intelligence for ADAS is provided.
Journal ArticleDOI
Benchmarking Particle Filter Algorithms for Efficient Velodyne-Based Vehicle Localization.
José Luis Blanco-Claraco,Francisco Mañas-Alvarez,José Luis Torres-Moreno,Francisco Rodríguez,Antonio Giménez-Fernández +4 more
TL;DR: This work shows that both standard SIR sampling and rejection-based optimal sampling are suitable for efficient (10 to 20 ms) real-time pose tracking without feature detection that is using raw point clouds from a 3D LiDAR.
References
More filters
Proceedings ArticleDOI
You Only Look Once: Unified, Real-Time Object Detection
TL;DR: Compared to state-of-the-art detection systems, YOLO makes more localization errors but is less likely to predict false positives on background, and outperforms other detection methods, including DPM and R-CNN, when generalizing from natural images to other domains like artwork.
Journal ArticleDOI
Vision meets robotics: The KITTI dataset
TL;DR: A novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research, using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras and a high-precision GPS/IMU inertial navigation system.
Proceedings ArticleDOI
Object scene flow for autonomous vehicles
Moritz Menze,Andreas Geiger +1 more
TL;DR: A novel model and dataset for 3D scene flow estimation with an application to autonomous driving by representing each element in the scene by its rigid motion parameters and each superpixel by a 3D plane as well as an index to the corresponding object.
Book ChapterDOI
A Unified Multi-scale Deep Convolutional Neural Network for Fast Object Detection
TL;DR: A unified deep neural network, denoted the multi-scale CNN (MS-CNN), is proposed for fast multi- scale object detection, which is learned end-to-end, by optimizing a multi-task loss.
Proceedings ArticleDOI
Decomposing a scene into geometric and semantically consistent regions
TL;DR: A region-based model which combines appearance and scene geometry to automatically decompose a scene into semantically meaningful regions and which achieves state-of-the-art performance on the tasks of both multi-class image segmentation and geometric reasoning.