scispace - formally typeset
Open AccessProceedings ArticleDOI

VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition

TLDR
In this paper, a unified end-to-end trainable multi-task network that jointly handles lane and road marking detection and recognition that is guided by a vanishing point under adverse weather conditions is proposed.
Abstract
In this paper, we propose a unified end-to-end trainable multi-task network that jointly handles lane and road marking detection and recognition that is guided by a vanishing point under adverse weather conditions. We tackle rainy and low illumination conditions, which have not been extensively studied until now due to clear challenges. For example, images taken under rainy days are subject to low illumination, while wet roads cause light reflection and distort the appearance of lane and road markings. At night, color distortion occurs under limited illumination. As a result, no benchmark dataset exists and only a few developed algorithms work under poor weather conditions. To address this shortcoming, we build up a lane and road marking benchmark which consists of about 20,000 images with 17 lane and road marking classes under four different scenarios: no rain, rain, heavy rain, and night. We train and evaluate several versions of the proposed multi-task network and validate the importance of each task. The resulting approach, VPGNet, can detect and classify lanes and road markings, and predict a vanishing point with a single forward pass. Experimental results show that our approach achieves high accuracy and robustness under various conditions in realtime (20 fps). The benchmark and the VPGNet model will be publicly available

read more

Citations
More filters
Proceedings ArticleDOI

Guided Dual Network Based Transfer with an Embedded Loss for Lane Detection in Nighttime Scene

TL;DR: Wang et al. as mentioned in this paper proposed a dual network which takes two images of the same scene in daytime and nighttime as input and detects lanes simultaneously, which adopts the transfer to generate nighttime images from the daytime images in an unsupervised manner.
Book ChapterDOI

3DLaneNAS: Neural Architecture Search for Accurate and Light-Weight 3D Lane Detection

TL;DR: 3DLaneNAS as mentioned in this paper utilizes a new multi-objective energy function to optimize the architecture of feature extraction and feature fusion modules simultaneously, and a transfer learning mechanism is used to improve the convergence of the search process.
Book ChapterDOI

An End-to-End Practical System for Road Marking Detection

TL;DR: Deep learning techniques, especially deep neural networks, have proven to be effective in coping with a variety of computer vision tasks, and usingDeep neural networks to construct road marking detection systems is a practical solution.
Journal ArticleDOI

Semi-Automatic Framework for Traffic Landmark Annotation

TL;DR: In this paper, a semi-automatic annotation method was proposed to build a large dataset for traffic landmark detection, where traffic landmarks include traffic signs, traffic lights as well as road markings.
References
More filters
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Proceedings ArticleDOI

Histograms of oriented gradients for human detection

TL;DR: It is shown experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection, and the influence of each stage of the computation on performance is studied.
Book ChapterDOI

Microsoft COCO: Common Objects in Context

TL;DR: A new dataset with the goal of advancing the state-of-the-art in object recognition by placing the question of object recognition in the context of the broader question of scene understanding by gathering images of complex everyday scenes containing common objects in their natural context.
Proceedings ArticleDOI

Fully convolutional networks for semantic segmentation

TL;DR: The key insight is to build “fully convolutional” networks that take input of arbitrary size and produce correspondingly-sized output with efficient inference and learning.
Proceedings ArticleDOI

You Only Look Once: Unified, Real-Time Object Detection

TL;DR: Compared to state-of-the-art detection systems, YOLO makes more localization errors but is less likely to predict false positives on background, and outperforms other detection methods, including DPM and R-CNN, when generalizing from natural images to other domains like artwork.