scispace - formally typeset
Journal ArticleDOI

HOTS: A Hierarchy of Event-Based Time-Surfaces for Pattern Recognition

TLDR
The central concept is to use the rich temporal information provided by events to create contexts in the form of time-surfaces which represent the recent temporal activity within a local spatial neighborhood and it is demonstrated that this concept can robustly be used at all stages of an event-based hierarchical model.
Abstract
This paper describes novel event-based spatio-temporal features called time-surfaces and how they can be used to create a hierarchical event-based pattern recognition architecture. Unlike existing hierarchical architectures for pattern recognition, the presented model relies on a time oriented approach to extract spatio-temporal features from the asynchronously acquired dynamics of a visual scene. These dynamics are acquired using biologically inspired frameless asynchronous event-driven vision sensors. Similarly to cortical structures, subsequent layers in our hierarchy extract increasingly abstract features using increasingly large spatio-temporal windows. The central concept is to use the rich temporal information provided by events to create contexts in the form of time-surfaces which represent the recent temporal activity within a local spatial neighborhood. We demonstrate that this concept can robustly be used at all stages of an event-based hierarchical model. First layer feature units operate on groups of pixels, while subsequent layer feature units operate on the output of lower level feature units. We report results on a previously published 36 class character recognition task and a four class canonical dynamic card pip task, achieving near 100 percent accuracy on each. We introduce a new seven class moving face recognition task, achieving 79 percent accuracy.

read more

Citations
More filters
Proceedings ArticleDOI

Learn to See by Events: Color Frame Synthesis from Event and RGB Cameras

TL;DR: In this paper, a deep learning-based frame synthesis method is proposed, consisting of an adversarial architecture combined with a recurrent module, which exploits the output stream of event cameras to synthesize RGB frames, relying on an initial or periodic set of color key-frames and the sequence of intermediate events.
Proceedings ArticleDOI

Learning from Event Cameras with Sparse Spiking Convolutional Neural Networks

TL;DR: In this article, a biologically inspired approach using event cameras and spiking neural networks (SNNs) is proposed to design more efficient computer vision algorithms, which enables the training of sparse spiking convolutional neural networks directly on event data using the popular deep learning framework PyTorch.
Journal ArticleDOI

Review on Vehicle Detection Technology for Unmanned Ground Vehicles

TL;DR: In this article, the authors introduce commonly used sensors for vehicle detection, lists their application scenarios and compares the strengths and weaknesses of different sensors, and several simulation platforms related to UGVs are presented for facilitating simulation testing of vehicle detection algorithms.
Journal ArticleDOI

Asynchronous Spatial Image Convolutions for Event Cameras

TL;DR: In this article, the authors propose a method to compute the convolution of a linear spatial kernel with the output of an event camera, which operates on the event stream output of the camera directly without synthesizing pseudo-image frames as is common in the literature.
Posted Content

Asynchronous Convolutional Networks for Object Detection in Neuromorphic Cameras

TL;DR: Two neural networks architectures for object detection are proposed: YOLE, which integrates the events into surfaces and uses a frame-based model to process them, and fcYOle, an asynchronous event-based fully convolutional network which uses a novel and general formalization of the convolutionAL and max pooling layers to exploit the sparsity of camera events.
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Journal ArticleDOI

Emergence of simple-cell receptive field properties by learning a sparse code for natural images

TL;DR: It is shown that a learning algorithm that attempts to find sparse linear codes for natural scenes will develop a complete family of localized, oriented, bandpass receptive fields, similar to those found in the primary visual cortex.
Proceedings Article

Large Scale Distributed Deep Networks

TL;DR: This paper considers the problem of training a deep network with billions of parameters using tens of thousands of CPU cores and develops two algorithms for large-scale distributed training, Downpour SGD and Sandblaster L-BFGS, which increase the scale and speed of deep network training.
Related Papers (5)