scispace - formally typeset
Open AccessJournal ArticleDOI

Vanets Meet Autonomous Vehicles: Multimodal Surrounding Recognition Using Manifold Alignment

Reads0
Chats0
TLDR
The vision to create a beneficial link between the two worlds of vehicular ad-hoc networks and autonomous vehicles is presented by designing a multimodal scheme for object detection, recognition, and mapping based on the fusion of stereo camera frames, point cloud Velodyne LIDAR scans, and vehicle-to-vehicle (V2V) basic safety messages (BSMs) exchanges using VANET protocols.
Abstract
In the past two years, calls for developing synergistic links between the two worlds of vehicular ad-hoc networks (VANETs) and autonomous vehicles have significantly gone up to achieve further on-road safety and benefits for end-users. In this paper, we present our vision to create such a beneficial link by designing a multimodal scheme for object detection, recognition, and mapping based on the fusion of stereo camera frames, point cloud Velodyne LIDAR scans, and vehicle-to-vehicle (V2V) basic safety messages (BSMs) exchanges using VANET protocols. Exploiting the high similarities in the underlying manifold properties of the three data sets, and their high neighborhood correlation, the proposed scheme employs semi-supervised manifold alignment to merge the key features of rich texture descriptions of objects from 2-D images, depth and distance between objects provided by 3-D point cloud, and the awareness of self-declared vehicles from BSMs’ 3-D information including the ones not seen by camera and LIDAR. The proposed scheme is applied to create joint pixel-to-point-cloud and pixel-to-V2V correspondences of objects in frames from the KITTI Vision Benchmark Suite, using a semi-supervised manifold alignment, to achieve camera-LIDAR and camera-V2V mapping of their recognized objects. We present the alignment accuracy results over two different driving sequences and show the additional acquired knowledge of objects from the various input modalities. We also study the effect of the number of neighbors employed in the alignment process on the alignment accuracy. With proper choice of parameters, the testing of our proposed scheme over two entire driving sequences exhibits 100% accuracy in the majority of cases, 74%–92% and 50%–72% average alignment accuracy for vehicles and pedestrians and up to 150% additional object recognition of the testing vehicle’s surrounding.

read more

Citations
More filters
Journal ArticleDOI

A Point Cloud-Based Robust Road Curb Detection and Tracking Method

TL;DR: A robust 3D-LiDAR-based method for road curb detection and tracking in a structured environment and can deal with straight and curved road without being influenced by surrounding obstacles is proposed.
Journal ArticleDOI

Ground Surface Filtering of 3D Point Clouds Based on Hybrid Regression Technique

TL;DR: The results show that the proposed method performs well in most real scenarios, even in the cases of ground undulation, occlusion, and sparse point clouds.
Journal ArticleDOI

Parallel Vehicular Networks: A CPSS-Based Approach via Multimodal Big Data in IoV

TL;DR: The proposed PVN offers a competitive solution for achieving a smooth, safe, and efficient cooperation among connected vehicles in future ITSs and is expected to achieve the descriptive intelligence, predictive intelligence, and prescription intelligence for VNs.
Journal ArticleDOI

A Progressive Review: Emerging Technologies for ADAS Driven Solutions

TL;DR: In this paper , the state-of-the-art of ADAS and its levels of autonomy are reviewed and a detailed description of vision intelligence and computational intelligence for ADAS is provided.
Journal ArticleDOI

Benchmarking Particle Filter Algorithms for Efficient Velodyne-Based Vehicle Localization.

TL;DR: This work shows that both standard SIR sampling and rejection-based optimal sampling are suitable for efficient (10 to 20 ms) real-time pose tracking without feature detection that is using raw point clouds from a 3D LiDAR.
References
More filters
Proceedings ArticleDOI

You Only Look Once: Unified, Real-Time Object Detection

TL;DR: Compared to state-of-the-art detection systems, YOLO makes more localization errors but is less likely to predict false positives on background, and outperforms other detection methods, including DPM and R-CNN, when generalizing from natural images to other domains like artwork.
Journal ArticleDOI

Vision meets robotics: The KITTI dataset

TL;DR: A novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research, using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras and a high-precision GPS/IMU inertial navigation system.
Proceedings ArticleDOI

Object scene flow for autonomous vehicles

TL;DR: A novel model and dataset for 3D scene flow estimation with an application to autonomous driving by representing each element in the scene by its rigid motion parameters and each superpixel by a 3D plane as well as an index to the corresponding object.
Book ChapterDOI

A Unified Multi-scale Deep Convolutional Neural Network for Fast Object Detection

TL;DR: A unified deep neural network, denoted the multi-scale CNN (MS-CNN), is proposed for fast multi- scale object detection, which is learned end-to-end, by optimizing a multi-task loss.
Proceedings ArticleDOI

Decomposing a scene into geometric and semantically consistent regions

TL;DR: A region-based model which combines appearance and scene geometry to automatically decompose a scene into semantically meaningful regions and which achieves state-of-the-art performance on the tasks of both multi-class image segmentation and geometric reasoning.
Related Papers (5)