scispace - formally typeset
Search or ask a question
Author

Alain L. Kornhauser

Bio: Alain L. Kornhauser is an academic researcher from Princeton University. The author has contributed to research in topics: Personal rapid transit & Computer graphics. The author has an hindex of 20, co-authored 80 publications receiving 4051 citations.


Papers
More filters
Proceedings ArticleDOI
07 Dec 2015
TL;DR: This paper proposes to map an input image to a small number of key perception indicators that directly relate to the affordance of a road/traffic state for driving and argues that the direct perception representation provides the right level of abstraction.
Abstract: Today, there are two major paradigms for vision-based autonomous driving systems: mediated perception approaches that parse an entire scene to make a driving decision, and behavior reflex approaches that directly map an input image to a driving action by a regressor. In this paper, we propose a third paradigm: a direct perception approach to estimate the affordance for driving. We propose to map an input image to a small number of key perception indicators that directly relate to the affordance of a road/traffic state for driving. Our representation provides a set of compact yet complete descriptions of the scene to enable a simple controller to drive autonomously. Falling in between the two extremes of mediated perception and behavior reflex, we argue that our direct perception representation provides the right level of abstraction. To demonstrate this, we train a deep Convolutional Neural Network using recording from 12 hours of human driving in a video game and show that our model can work well to drive a car in a very diverse set of virtual environments. We also train a model for car distance estimation on the KITTI dataset. Results show that our direct perception approach can generalize well to real driving images. Source code and data are available on our project website.

1,420 citations

Posted Content
TL;DR: In this paper, the authors propose a direct perception approach to estimate the affordance for driving in a video game and train a deep Convolutional Neural Network using recording from 12 hours of human driving.
Abstract: Today, there are two major paradigms for vision-based autonomous driving systems: mediated perception approaches that parse an entire scene to make a driving decision, and behavior reflex approaches that directly map an input image to a driving action by a regressor. In this paper, we propose a third paradigm: a direct perception approach to estimate the affordance for driving. We propose to map an input image to a small number of key perception indicators that directly relate to the affordance of a road/traffic state for driving. Our representation provides a set of compact yet complete descriptions of the scene to enable a simple controller to drive autonomously. Falling in between the two extremes of mediated perception and behavior reflex, we argue that our direct perception representation provides the right level of abstraction. To demonstrate this, we train a deep Convolutional Neural Network using recording from 12 hours of human driving in a video game and show that our model can work well to drive a car in a very diverse set of virtual environments. We also train a model for car distance estimation on the KITTI dataset. Results show that our direct perception approach can generalize well to real driving images. Source code and data are available on our project website.

998 citations

Journal ArticleDOI
TL;DR: This paper considers map matching algorithms that can be used to reconcile inaccurate locational data with an inaccurate map/network.
Abstract: Third-generation personal navigation assistants (PNAs) (i.e., those that provide a map, the user's current location, and directions) must be able to reconcile the user's location with the underlying map. This process is known as map matching. Most existing research has focused on map matching when both the user's location and the map are known with a high degree of accuracy. However, there are many situations in which this is unlikely to be the case. Hence, this paper considers map matching algorithms that can be used to reconcile inaccurate locational data with an inaccurate map/network.

647 citations

01 Aug 1998
TL;DR: This paper examines Personal Navigation Assistants (PNAs) and explores map-matching algorithms that can be used to reconcile inaccurate locational data with an inaccurate map/network.
Abstract: The paper examines Personal Navigation Assistants (PNAs) and identifies three different types. The first one provides the user with a map and the ability to search the map in a variety of ways; the second provides both a map and the user's current location; and the third provides a map, the user's location, and directions of some kind. The paper then explores map-matching algorithms that can be used to reconcile inaccurate locational data with an inaccurate map/network. Point-to-point, point-to-curve, and curve-to-curve matching is discussed, and in all three cases algorithms that only use geometric information and that also use topological information are considered.

304 citations

Journal ArticleDOI
TL;DR: In this article, a system of incentives was designed for the receivers of deliveries the system combined Global Positioning System (GPS) remote sensing monitoring with GPS-enabled smart phones to induce a shift of deliveries to the off-hours from 7:00 p.m. to 6:00 a.m., the concept was pilot tested in Manhattan by 33 companies that switched delivery operations to offhours for a period of 1 month.
Abstract: This paper examines the chief findings of research conducted on policies to foster off-hour deliveries (OHDs) in the New York City metropolitan area. The goal was to estimate the overall impacts of eventual full implementation of an OHD program. As part of the research, a system of incentives was designed for the receivers of deliveries the system combined Global Positioning System (GPS) remote sensing monitoring with GPS-enabled smart phones to induce a shift of deliveries to the off-hours from 7:00 p.m. to 6:00 a.m. The concept was pilot tested in Manhattan by 33 companies that switched delivery operations to the off-hours for a period of 1 month. At the in-depth interviews conducted after the test, the participants reported being very satisfied with the experience. As an alternative to road pricing schemes that target freight carriers, this was the first real-life trial of the use of financial incentives to delivery receivers. The analyses indicate that the economic benefits of a full implementation of...

122 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this article, a review of deep learning-based object detection frameworks is provided, focusing on typical generic object detection architectures along with some modifications and useful tricks to improve detection performance further.
Abstract: Due to object detection’s close relationship with video analysis and image understanding, it has attracted much research attention in recent years. Traditional object detection methods are built on handcrafted features and shallow trainable architectures. Their performance easily stagnates by constructing complex ensembles that combine multiple low-level image features with high-level context from object detectors and scene classifiers. With the rapid development in deep learning, more powerful tools, which are able to learn semantic, high-level, deeper features, are introduced to address the problems existing in traditional architectures. These models behave differently in network architecture, training strategy, and optimization function. In this paper, we provide a review of deep learning-based object detection frameworks. Our review begins with a brief introduction on the history of deep learning and its representative tool, namely, the convolutional neural network. Then, we focus on typical generic object detection architectures along with some modifications and useful tricks to improve detection performance further. As distinct specific detection tasks exhibit different characteristics, we also briefly survey several specific tasks, including salient object detection, face detection, and pedestrian detection. Experimental analyses are also provided to compare various methods and draw some meaningful conclusions. Finally, several promising directions and tasks are provided to serve as guidelines for future work in both object detection and relevant neural network-based learning systems.

3,097 citations

Journal ArticleDOI
20 Nov 2017
TL;DR: In this paper, the authors provide a comprehensive tutorial and survey about the recent advances toward the goal of enabling efficient processing of DNNs, and discuss various hardware platforms and architectures that support DNN, and highlight key trends in reducing the computation cost of deep neural networks either solely via hardware design changes or via joint hardware and DNN algorithm changes.
Abstract: Deep neural networks (DNNs) are currently widely used for many artificial intelligence (AI) applications including computer vision, speech recognition, and robotics. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Accordingly, techniques that enable efficient processing of DNNs to improve energy efficiency and throughput without sacrificing application accuracy or increasing hardware cost are critical to the wide deployment of DNNs in AI systems. This article aims to provide a comprehensive tutorial and survey about the recent advances toward the goal of enabling efficient processing of DNNs. Specifically, it will provide an overview of DNNs, discuss various hardware platforms and architectures that support DNNs, and highlight key trends in reducing the computation cost of DNNs either solely via hardware design changes or via joint hardware design and DNN algorithm changes. It will also summarize various development resources that enable researchers and practitioners to quickly get started in this field, and highlight important benchmarking metrics and design considerations that should be used for evaluating the rapidly growing number of DNN hardware designs, optionally including algorithmic codesigns, being proposed in academia and industry. The reader will take away the following concepts from this article: understand the key design considerations for DNNs; be able to evaluate different DNN hardware implementations with benchmarks and comparison metrics; understand the tradeoffs between various hardware architectures and platforms; be able to evaluate the utility of various DNN design techniques for efficient processing; and understand recent implementation trends and opportunities.

2,391 citations

18 Oct 2017
TL;DR: This work introduces CARLA, an open-source simulator for autonomous driving research, and uses it to study the performance of three approaches to autonomous driving: a classic modular pipeline, an end-to-end model trained via imitation learning, and an end to-end models trained via reinforcement learning.
Abstract: We introduce CARLA, an open-source simulator for autonomous driving research. CARLA has been developed from the ground up to support development, training, and validation of autonomous urban driving systems. In addition to open-source code and protocols, CARLA provides open digital assets (urban layouts, buildings, vehicles) that were created for this purpose and can be used freely. The simulation platform supports flexible specification of sensor suites and environmental conditions. We use CARLA to study the performance of three approaches to autonomous driving: a classic modular pipeline, an end-to-end model trained via imitation learning, and an end-to-end model trained via reinforcement learning. The approaches are evaluated in controlled scenarios of increasing difficulty, and their performance is examined via metrics provided by CARLA, illustrating the platform's utility for autonomous driving research. The supplementary video can be viewed at this https URL

1,539 citations

Book ChapterDOI
08 Oct 2016
TL;DR: In this paper, the authors present an approach to rapidly create pixel-accurate semantic label maps for images extracted from modern computer games, which enables rapid propagation of semantic labels within and across images synthesized by the game, without access to the source code or the content.
Abstract: Recent progress in computer vision has been driven by high-capacity models trained on large datasets. Unfortunately, creating large datasets with pixel-level labels has been extremely costly due to the amount of human effort required. In this paper, we present an approach to rapidly creating pixel-accurate semantic label maps for images extracted from modern computer games. Although the source code and the internal operation of commercial games are inaccessible, we show that associations between image patches can be reconstructed from the communication between the game and the graphics hardware. This enables rapid propagation of semantic labels within and across images synthesized by the game, with no access to the source code or the content. We validate the presented approach by producing dense pixel-level semantic annotations for 25 thousand images synthesized by a photorealistic open-world computer game. Experiments on semantic segmentation datasets show that using the acquired data to supplement real-world images significantly increases accuracy and that the acquired data enables reducing the amount of hand-labeled real-world data: models trained with game data and just \(\tfrac{1}{3}\) of the CamVid training set outperform models trained on the complete CamVid training set.

1,319 citations

Patent
02 Aug 2001
TL;DR: In this article, an automotive auto-pilot mode is provided, which generates control signals which actuate a plurality of control systems of the one vehicle in a coordinated manner to maneuver it laterally and longitudinally to avoid each collision hazard, or, for motor vehicles, when a collision is unavoidable, to minimize injury or damage therefrom.
Abstract: GPS satellite (4) ranging signals (6) received (32) on comm1, and DGPS auxiliary range correction signals and pseudolite carrier phase ambiguity resolution signals (8) from a fixed known earth base station (10) received (34) on comm2, at one of a plurality of vehicles/aircraft/automobiles (2) are computer processed (36) to continuously determine the one's kinematic tracking position on a pathway (14) with centimeter accuracy. That GPS-based position is communicated with selected other status information to each other one of the plurality of vehicles (2), to the one station (10), and/or to one of a plurality of control centers (16), and the one vehicle receives therefrom each of the others' status information and kinematic tracking position. Objects (22) are detected from all directions (300) by multiple supplemental mechanisms, e.g., video (54), radar/lidar (56), laser and optical scanners. Data and information are computer processed and analyzed (50,52,200,452) in neural networks (132, FIGS. 6-8) in the one vehicle to identify, rank, and evaluate collision hazards/objects, an expert operating response to which is determined in a fuzzy logic associative memory (484) which generates control signals which actuate a plurality of control systems of the one vehicle in a coordinated manner to maneuver it laterally and longitudinally to avoid each collision hazard, or, for motor vehicles, when a collision is unavoidable, to minimize injury or damage therefrom. The operator is warned by a heads up display and other modes and may override. An automotive auto-pilot mode is provided.

1,134 citations