scispace - formally typeset
Open AccessProceedings ArticleDOI

VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition

TLDR
In this paper, a unified end-to-end trainable multi-task network that jointly handles lane and road marking detection and recognition that is guided by a vanishing point under adverse weather conditions is proposed.
Abstract
In this paper, we propose a unified end-to-end trainable multi-task network that jointly handles lane and road marking detection and recognition that is guided by a vanishing point under adverse weather conditions. We tackle rainy and low illumination conditions, which have not been extensively studied until now due to clear challenges. For example, images taken under rainy days are subject to low illumination, while wet roads cause light reflection and distort the appearance of lane and road markings. At night, color distortion occurs under limited illumination. As a result, no benchmark dataset exists and only a few developed algorithms work under poor weather conditions. To address this shortcoming, we build up a lane and road marking benchmark which consists of about 20,000 images with 17 lane and road marking classes under four different scenarios: no rain, rain, heavy rain, and night. We train and evaluate several versions of the proposed multi-task network and validate the importance of each task. The resulting approach, VPGNet, can detect and classify lanes and road markings, and predict a vanishing point with a single forward pass. Experimental results show that our approach achieves high accuracy and robustness under various conditions in realtime (20 fps). The benchmark and the VPGNet model will be publicly available

read more

Citations
More filters
Posted Content

CeyMo: See More on Roads - A Novel Benchmark Dataset for Road Marking Detection.

TL;DR: In this paper, the authors introduce a road marking benchmark dataset for road marking detection, addressing the limitations in the existing publicly available datasets such as lack of challenging scenarios, prominence given to lane markings, unavailability of an evaluation script, lack of annotation formats and lower resolutions.
Proceedings ArticleDOI

Generating All the Roads to Rome: Road Layout Randomization for Improved Road Marking Segmentation

TL;DR: The authors apply principles of domain randomization to the road layout and synthesize new images from altered semantic labels, and demonstrate that training on these synthetic pairs improves mIoU of rare road marking classes during realworld deployment in complex urban environments by more than 12 percentage points, while performance for other classes is retained.
Posted Content

Instance-wise Depth and Motion Learning from Monocular Videos

TL;DR: This work presents an end-to-end joint training framework that explicitly models 6-DoF motion of multiple dynamic objects, ego-motion and depth in a monocular camera setup without supervision and is shown to outperform the state-of-the-art depth and motion estimation methods.
Proceedings ArticleDOI

CeyMo: See More on Roads - A Novel Benchmark Dataset for Road Marking Detection

TL;DR: In this paper , a road marking bench-mark dataset for road marking detection is presented. But, the road marking annotations in polygons, bounding boxes and pixel-level segmentation masks are not included in the dataset.
Journal ArticleDOI

MTSAN: Multi-Task Semantic Attention Network for ADAS Applications

TL;DR: In this article, a lightweight Multi-task Semantic Attention Network (MTSAN) is proposed to collectively deal with object detection and semantic segmentation aiding real-time applications of Advanced Driver Assistance Systems (ADAS).
References
More filters
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Proceedings ArticleDOI

Histograms of oriented gradients for human detection

TL;DR: It is shown experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection, and the influence of each stage of the computation on performance is studied.
Book ChapterDOI

Microsoft COCO: Common Objects in Context

TL;DR: A new dataset with the goal of advancing the state-of-the-art in object recognition by placing the question of object recognition in the context of the broader question of scene understanding by gathering images of complex everyday scenes containing common objects in their natural context.
Proceedings ArticleDOI

Fully convolutional networks for semantic segmentation

TL;DR: The key insight is to build “fully convolutional” networks that take input of arbitrary size and produce correspondingly-sized output with efficient inference and learning.
Proceedings ArticleDOI

You Only Look Once: Unified, Real-Time Object Detection

TL;DR: Compared to state-of-the-art detection systems, YOLO makes more localization errors but is less likely to predict false positives on background, and outperforms other detection methods, including DPM and R-CNN, when generalizing from natural images to other domains like artwork.